菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-02-09
📄 Abstract - Automatic In-Domain Exemplar Construction and LLM-Based Refinement of Multi-LLM Expansions for Query Expansion

Query expansion with large language models is promising but often relies on hand-crafted prompts, manually chosen exemplars, or a single LLM, making it non-scalable and sensitive to domain shift. We present an automated, domain-adaptive QE framework that builds in-domain exemplar pools by harvesting pseudo-relevant passages using a BM25-MonoT5 pipeline. A training-free cluster-based strategy selects diverse demonstrations, yielding strong and stable in-context QE without supervision. To further exploit model complementarity, we introduce a two-LLM ensemble in which two heterogeneous LLMs independently generate expansions and a refinement LLM consolidates them into one coherent expansion. Across TREC DL20, DBPedia, and SciFact, the refined ensemble delivers consistent and statistically significant gains over BM25, Rocchio, zero-shot, and fixed few-shot baselines. The framework offers a reproducible testbed for exemplar selection and multi-LLM generation, and a practical, label-free solution for real-world QE.

顶级标签: llm natural language processing systems
详细标签: query expansion exemplar selection multi-llm ensemble domain adaptation retrieval 或 搜索:

基于自动构建领域内范例与多LLM扩展精化的查询扩展方法 / Automatic In-Domain Exemplar Construction and LLM-Based Refinement of Multi-LLM Expansions for Query Expansion


1️⃣ 一句话总结

这篇论文提出了一种全自动的查询扩展框架,它能够自动为不同领域的问题找到合适的参考例子,并巧妙地组合多个大语言模型的优势来生成更准确、更稳定的查询扩展,从而有效提升信息检索的效果。

源自 arXiv: 2602.08917