LLM Reasoning 相关度: 8/10

Decomposed Prompting Does Not Fix Knowledge Gaps, But Helps Models Say "I Don't Know"

Dhruv Madhwal, Lyuxin David Zhang, Dan Roth, Tomer Wolfson, Vivek Gupta
arXiv: 2602.04853v1 发布: 2026-02-04 更新: 2026-02-04

AI 摘要

分解提示不能弥补知识差距,但能帮助模型表达“我不知道”。

主要贡献

  • 揭示分解提示对模型可靠性的影响
  • 提出基于提示方式不一致性的不确定性信号
  • 提出无需训练的基于不一致性的拒绝策略

方法论

对比直接、辅助和递增三种提示方式,通过模型间的意见分歧来判断知识不确定性,并利用该信号实现拒绝回答。

原文摘要

Large language models often struggle to recognize their knowledge limits in closed-book question answering, leading to confident hallucinations. While decomposed prompting is typically used to improve accuracy, we investigate its impact on reliability. We evaluate three task-equivalent prompting regimes: Direct, Assistive, and Incremental, across different model scales and multi-hop QA benchmarks. We find that although accuracy gains from decomposition diminish in frontier models, disagreements between prompting regimes remain highly indicative of potential errors. Because factual knowledge is stable while hallucinations are stochastic, cross-regime agreement provides a precise signal of internal uncertainty. We leverage this signal to implement a training-free abstention policy that requires no retrieval or fine-tuning. Our results show that disagreement-based abstention outperforms standard uncertainty baselines as an error detector, improving both F1 and AUROC across settings. This demonstrates that decomposition-based prompting can serve as a practical diagnostic probe for model reliability in closed-book QA.

标签

LLM Prompting Uncertainty Closed-book QA

arXiv 分类

cs.CL