菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-03-11
📄 Abstract - A Systematic Study of Pseudo-Relevance Feedback with LLMs

Pseudo-relevance feedback (PRF) methods built on large language models (LLMs) can be organized along two key design dimensions: the feedback source, which is where the feedback text is derived from and the feedback model, which is how the given feedback text is used to refine the query representation. However, the independent role that each dimension plays is unclear, as both are often entangled in empirical evaluations. In this paper, we address this gap by systematically studying how the choice of feedback source and feedback model impact PRF effectiveness through controlled experimentation. Across 13 low-resource BEIR tasks with five LLM PRF methods, our results show: (1) the choice of feedback model can play a critical role in PRF effectiveness; (2) feedback derived solely from LLM-generated text provides the most cost-effective solution; and (3) feedback derived from the corpus is most beneficial when utilizing candidate documents from a strong first-stage retriever. Together, our findings provide a better understanding of which elements in the PRF design space are most important.

顶级标签: llm natural language processing model evaluation
详细标签: pseudo-relevance feedback information retrieval query refinement retrieval-augmented generation beir benchmark 或 搜索:

基于大语言模型的伪相关反馈系统化研究 / A Systematic Study of Pseudo-Relevance Feedback with LLMs


1️⃣ 一句话总结

这篇论文通过系统实验发现,在使用大语言模型进行伪相关反馈时,反馈模型的选择对效果至关重要,仅用模型生成的文本作为反馈源最具性价比,而使用外部文档库作为反馈源则需依赖强大的初始检索器才能发挥最大效益。

源自 arXiv: 2603.11008