LLM Reasoning 相关度: 8/10

HindSight: Evaluating Research Idea Generation via Future Impact

Bo Jiang
arXiv: 2603.15164v1 发布: 2026-03-16 更新: 2026-03-16

AI 摘要

提出HindSight框架,通过未来影响评估AI生成研究想法的质量,揭示了LLM评估与实际影响的差异。

主要贡献

  • 提出HindSight评估框架
  • 揭示LLM评估与实际研究影响的差异
  • 发现LLM倾向于高估不切实际的新颖想法

方法论

利用时间分割,将生成想法与未来实际发表论文匹配,并根据引用和接收度进行评分。

原文摘要

Evaluating AI-generated research ideas typically relies on LLM judges or human panels -- both subjective and disconnected from actual research impact. We introduce \hs{}, a time-split evaluation framework that measures idea quality by matching generated ideas against real future publications and scoring them by citation impact and venue acceptance. Using a temporal cutoff~$T$, we restrict an idea generation system to pre-$T$ literature, then evaluate its outputs against papers published in the subsequent 30 months. Experiments across 10 AI/ML research topics reveal a striking disconnect: LLM-as-Judge finds no significant difference between retrieval-augmented and vanilla idea generation ($p{=}0.584$), while \hs{} shows the retrieval-augmented system produces 2.5$\times$ higher-scoring ideas ($p{<}0.001$). Moreover, \hs{} scores are \emph{negatively} correlated with LLM-judged novelty ($ρ{=}{-}0.29$, $p{<}0.01$), suggesting that LLMs systematically overvalue novel-sounding ideas that never materialize in real research.

标签

评估框架 研究想法生成 LLM评估 未来影响

arXiv 分类

cs.CL cs.AI cs.LG