LLM Reasoning 相关度: 9/10

More Test-Time Compute Can Hurt: Overestimation Bias in LLM Beam Search

Gal Dalal, Assaf Hallak, Gal Chechik, Yftach Ziser
arXiv: 2603.15377v1 发布: 2026-03-16 更新: 2026-03-16

AI 摘要

更大beam size可能损害LLM推理性能,论文分析了过估计偏差问题并提出了最佳beam size选择方法。

主要贡献

  • 揭示了beam search中的过估计偏差问题
  • 提出了基于信号噪声比的最大有效beam width理论
  • 提出了实际应用中选择beam width的诊断指标

方法论

论文基于极值理论分析beam search的偏差,并通过实验验证了不同scorer对最佳beam width的影响。

原文摘要

Wider beam search should improve LLM reasoning, but when should you stop widening? Prior work on beam width selection has focused on inference efficiency \citep{qin2025dsbd, freitag2017beam}, without analyzing whether wider search can \emph{hurt} output quality. We present an analysis, grounded in Extreme Value Theory, that answers this question. Beam selection over noisy scorer outputs introduces a systematic overestimation bias that grows with the candidate pool size, and we derive a maximum useful beam width $\hat{k}$ beyond which search degrades performance. This critical width depends on the signal-to-noise ratio of the scorer: $\hat{k}$ grows exponentially with $(Δ/σ)^2$, where $Δ> 0$ is the quality advantage of correct paths over incorrect ones and $σ$ is the scorer noise. We validate this theory by comparing perplexity-guided and PRM-guided beam search across three 7B-parameter models and ten domains on MR-BEN (5,975 questions). Perplexity scoring, with its high noise, yields $\hat{k} = 1$: search provides no benefit at any width tested. PRM scoring, with lower noise, yields $\hat{k} \geq 4$, with gains of up to 8.9 percentage points. The same model, the same algorithm, but different scorers place $\hat{k}$ at opposite ends of the beam width range. Our analysis identifies the scorer's signal-to-noise ratio as the key quantity governing beam width selection, and we propose diagnostic indicators for choosing the beam width in practice.

标签

LLM Beam Search Reasoning Overestimation Bias Extreme Value Theory

arXiv 分类

cs.LG cs.AI