SAFE: Stepwise Atomic Feedback for Error correction in Multi-hop Reasoning
AI 摘要
SAFE框架通过KG验证的逐步反馈纠正LLM多跳推理中的错误,提升推理的可靠性和准确性。
主要贡献
- 提出了SAFE框架,用于动态评估和纠正LLM多跳推理错误
- 建立了原子错误分类体系和KG验证流程,用于识别和消除训练数据中的噪声
- 设计了反馈模型,用于在推理时动态检测未根据知识图谱的推理步骤
方法论
SAFE框架包含训练时验证和推理时验证两个阶段,利用KG进行事实验证,并训练反馈模型进行动态纠错。
原文摘要
Multi-hop QA benchmarks frequently reward Large Language Models (LLMs) for spurious correctness, masking ungrounded or flawed reasoning steps. To shift toward rigorous reasoning, we propose SAFE, a dynamic benchmarking framework that replaces the ungrounded Chain-of-Thought (CoT) with a strictly verifiable sequence of grounded entities. Our framework operates across two phases: (1) train-time verification, where we establish an atomic error taxonomy and a Knowledge Graph (KG)-grounded verification pipeline to eliminate noisy supervision in standard benchmarks, identifying up to 14% of instances as unanswerable, and (2) inference-time verification, where a feedback model trained on this verified dataset dynamically detects ungrounded steps in real-time. Experimental results demonstrate that SAFE not only exposes the critical flaws of existing benchmarks at train-time, but also significantly outperforms standard baselines, achieving an average accuracy gain of 8.4 pp while guaranteeing verifiable trajectories at inference-time.