LLM Reasoning 相关度: 8/10

Leveraging LLM Parametric Knowledge for Fact Checking without Retrieval

Artem Vazhentsev, Maria Marina, Daniil Moskovskiy, Sergey Pletenev, Mikhail Seleznyov, Mikhail Salnikov, Elena Tutubalina, Vasily Konovalov, Irina Nikishina, Alexander Panchenko, Viktor Moskvoretskii
arXiv: 2603.05471v1 发布: 2026-03-05 更新: 2026-03-05

AI 摘要

论文提出了一种不依赖检索的LLM事实核查方法,并通过实验验证了其有效性和泛化能力。

主要贡献

  • 提出了不依赖检索的事实核查任务
  • 设计了一个全面的评估框架,关注泛化性
  • 提出了INTRA方法,利用内部表示实现SOTA

方法论

通过分析LLM内部表示,提出INTRA方法,无需外部知识检索即可进行事实核查,并在多个数据集上进行评估。

原文摘要

Trustworthiness is a core research challenge for agentic AI systems built on Large Language Models (LLMs). To enhance trust, natural language claims from diverse sources, including human-written text, web content, and model outputs, are commonly checked for factuality by retrieving external knowledge and using an LLM to verify the faithfulness of claims to the retrieved evidence. As a result, such methods are constrained by retrieval errors and external data availability, while leaving the models intrinsic fact-verification capabilities largely unused. We propose the task of fact-checking without retrieval, focusing on the verification of arbitrary natural language claims, independent of their source. To study this setting, we introduce a comprehensive evaluation framework focused on generalization, testing robustness to (i) long-tail knowledge, (ii) variation in claim sources, (iii) multilinguality, and (iv) long-form generation. Across 9 datasets, 18 methods and 3 models, our experiments indicate that logit-based approaches often underperform compared to those that leverage internal model representations. Building on this finding, we introduce INTRA, a method that exploits interactions between internal representations and achieves state-of-the-art performance with strong generalization. More broadly, our work establishes fact-checking without retrieval as a promising research direction that can complement retrieval-based frameworks, improve scalability, and enable the use of such systems as reward signals during training or as components integrated into the generation process.

标签

事实核查 LLM 内部表示 泛化性

arXiv 分类

cs.CL cs.AI