LLM Reasoning 相关度: 9/10

SAFE: Stepwise Atomic Feedback for Error correction in Multi-hop Reasoning

Daeyong Kwon, Soyoung Yoon, Seung-won Hwang
arXiv: 2604.01993v1 发布: 2026-04-02 更新: 2026-04-02

AI 摘要

SAFE框架通过KG验证的逐步反馈纠正LLM多跳推理中的错误,提升推理的可靠性和准确性。

主要贡献

  • 提出了SAFE框架,用于动态评估和纠正LLM多跳推理错误
  • 建立了原子错误分类体系和KG验证流程,用于识别和消除训练数据中的噪声
  • 设计了反馈模型,用于在推理时动态检测未根据知识图谱的推理步骤

方法论

SAFE框架包含训练时验证和推理时验证两个阶段,利用KG进行事实验证,并训练反馈模型进行动态纠错。

原文摘要

Multi-hop QA benchmarks frequently reward Large Language Models (LLMs) for spurious correctness, masking ungrounded or flawed reasoning steps. To shift toward rigorous reasoning, we propose SAFE, a dynamic benchmarking framework that replaces the ungrounded Chain-of-Thought (CoT) with a strictly verifiable sequence of grounded entities. Our framework operates across two phases: (1) train-time verification, where we establish an atomic error taxonomy and a Knowledge Graph (KG)-grounded verification pipeline to eliminate noisy supervision in standard benchmarks, identifying up to 14% of instances as unanswerable, and (2) inference-time verification, where a feedback model trained on this verified dataset dynamically detects ungrounded steps in real-time. Experimental results demonstrate that SAFE not only exposes the critical flaws of existing benchmarks at train-time, but also significantly outperforms standard baselines, achieving an average accuracy gain of 8.4 pp while guaranteeing verifiable trajectories at inference-time.

标签

多跳推理 知识图谱 错误纠正 可验证推理

arXiv 分类

cs.CL cs.AI