基于自动构建领域内范例与多LLM扩展精化的查询扩展方法 / Automatic In-Domain Exemplar Construction and LLM-Based Refinement of Multi-LLM Expansions for Query Expansion
1️⃣ 一句话总结
这篇论文提出了一种全自动的查询扩展框架,它能够自动为不同领域的问题找到合适的参考例子,并巧妙地组合多个大语言模型的优势来生成更准确、更稳定的查询扩展,从而有效提升信息检索的效果。
Query expansion with large language models is promising but often relies on hand-crafted prompts, manually chosen exemplars, or a single LLM, making it non-scalable and sensitive to domain shift. We present an automated, domain-adaptive QE framework that builds in-domain exemplar pools by harvesting pseudo-relevant passages using a BM25-MonoT5 pipeline. A training-free cluster-based strategy selects diverse demonstrations, yielding strong and stable in-context QE without supervision. To further exploit model complementarity, we introduce a two-LLM ensemble in which two heterogeneous LLMs independently generate expansions and a refinement LLM consolidates them into one coherent expansion. Across TREC DL20, DBPedia, and SciFact, the refined ensemble delivers consistent and statistically significant gains over BM25, Rocchio, zero-shot, and fixed few-shot baselines. The framework offers a reproducible testbed for exemplar selection and multi-LLM generation, and a practical, label-free solution for real-world QE.
基于自动构建领域内范例与多LLM扩展精化的查询扩展方法 / Automatic In-Domain Exemplar Construction and LLM-Based Refinement of Multi-LLM Expansions for Query Expansion
这篇论文提出了一种全自动的查询扩展框架,它能够自动为不同领域的问题找到合适的参考例子,并巧妙地组合多个大语言模型的优势来生成更准确、更稳定的查询扩展,从而有效提升信息检索的效果。
源自 arXiv: 2602.08917