检索增强生成中的“少即是多”:基于信息增益剪枝的生成器对齐重排序与证据选择 / Less is More for RAG: Information Gain Pruning for Generator-Aligned Reranking and Evidence Selection
1️⃣ 一句话总结
这篇论文提出了一种名为‘信息增益剪枝’的新方法,它能在不改变现有系统接口的前提下,智能地筛选出对生成答案真正有用的检索文本,从而在显著减少输入信息量的同时,大幅提升问答系统的准确率。
Retrieval-augmented generation (RAG) grounds large language models with external evidence, but under a limited context budget, the key challenge is deciding which retrieved passages should be injected. We show that retrieval relevance metrics (e.g., NDCG) correlate weakly with end-to-end QA quality and can even become negatively correlated under multi-passage injection, where redundancy and mild conflicts destabilize generation. We propose \textbf{Information Gain Pruning (IGP)}, a deployment-friendly reranking-and-pruning module that selects evidence using a generator-aligned utility signal and filters weak or harmful passages before truncation, without changing existing budget interfaces. Across five open-domain QA benchmarks and multiple retrievers and generators, IGP consistently improves the quality--cost trade-off. In a representative multi-evidence setting, IGP delivers about +12--20% relative improvement in average F1 while reducing final-stage input tokens by roughly 76--79% compared to retriever-only baselines.
检索增强生成中的“少即是多”:基于信息增益剪枝的生成器对齐重排序与证据选择 / Less is More for RAG: Information Gain Pruning for Generator-Aligned Reranking and Evidence Selection
这篇论文提出了一种名为‘信息增益剪枝’的新方法,它能在不改变现有系统接口的前提下,智能地筛选出对生成答案真正有用的检索文本,从而在显著减少输入信息量的同时,大幅提升问答系统的准确率。
源自 arXiv: 2601.17532