HindSight: Evaluating Research Idea Generation via Future Impact
AI 摘要
提出HindSight框架,通过未来影响评估AI生成研究想法的质量,揭示了LLM评估与实际影响的差异。
主要贡献
- 提出HindSight评估框架
- 揭示LLM评估与实际研究影响的差异
- 发现LLM倾向于高估不切实际的新颖想法
方法论
利用时间分割,将生成想法与未来实际发表论文匹配,并根据引用和接收度进行评分。
原文摘要
Evaluating AI-generated research ideas typically relies on LLM judges or human panels -- both subjective and disconnected from actual research impact. We introduce \hs{}, a time-split evaluation framework that measures idea quality by matching generated ideas against real future publications and scoring them by citation impact and venue acceptance. Using a temporal cutoff~$T$, we restrict an idea generation system to pre-$T$ literature, then evaluate its outputs against papers published in the subsequent 30 months. Experiments across 10 AI/ML research topics reveal a striking disconnect: LLM-as-Judge finds no significant difference between retrieval-augmented and vanilla idea generation ($p{=}0.584$), while \hs{} shows the retrieval-augmented system produces 2.5$\times$ higher-scoring ideas ($p{<}0.001$). Moreover, \hs{} scores are \emph{negatively} correlated with LLM-judged novelty ($ρ{=}{-}0.29$, $p{<}0.01$), suggesting that LLMs systematically overvalue novel-sounding ideas that never materialize in real research.