A Systematic Study of Pseudo-Relevance Feedback with LLMs
AI 摘要
系统性研究LLM伪相关反馈,揭示反馈源和反馈模型对效果的影响,提供设计指导。
主要贡献
- 系统分析反馈源和反馈模型对PRF的影响
- 揭示LLM生成文本作为反馈源的有效性
- 强调反馈模型在PRF中的关键作用
方法论
通过控制实验,在低资源BEIR数据集上评估五种LLM PRF方法,分析不同反馈源和反馈模型的性能。
原文摘要
Pseudo-relevance feedback (PRF) methods built on large language models (LLMs) can be organized along two key design dimensions: the feedback source, which is where the feedback text is derived from and the feedback model, which is how the given feedback text is used to refine the query representation. However, the independent role that each dimension plays is unclear, as both are often entangled in empirical evaluations. In this paper, we address this gap by systematically studying how the choice of feedback source and feedback model impact PRF effectiveness through controlled experimentation. Across 13 low-resource BEIR tasks with five LLM PRF methods, our results show: (1) the choice of feedback model can play a critical role in PRF effectiveness; (2) feedback derived solely from LLM-generated text provides the most cost-effective solution; and (3) feedback derived from the corpus is most beneficial when utilizing candidate documents from a strong first-stage retriever. Together, our findings provide a better understanding of which elements in the PRF design space are most important.