通过早期知识对齐实现多跳推理 / Multi-hop Reasoning via Early Knowledge Alignment
1️⃣ 一句话总结
这篇论文提出了一种名为‘早期知识对齐’的简单有效方法,让大语言模型在分解复杂问题之前先了解可用的知识库信息,从而显著提升了多步问答系统的准确性和效率,减少了推理过程中的错误传递。
Retrieval-Augmented Generation (RAG) has emerged as a powerful paradigm for Large Language Models (LLMs) to address knowledge-intensive queries requiring domain-specific or up-to-date information. To handle complex multi-hop questions that are challenging for single-step retrieval, iterative RAG approaches incorporating reinforcement learning have been proposed. However, existing iterative RAG systems typically plan to decompose questions without leveraging information about the available retrieval corpus, leading to inefficient retrieval and reasoning chains that cascade into suboptimal performance. In this paper, we introduce Early Knowledge Alignment (EKA), a simple but effective module that aligns LLMs with retrieval set before planning in iterative RAG systems with contextually relevant retrieved knowledge. Extensive experiments on six standard RAG datasets demonstrate that by establishing a stronger reasoning foundation, EKA significantly improves retrieval precision, reduces cascading errors, and enhances both performance and efficiency. Our analysis from an entropy perspective demonstrate that incorporating early knowledge reduces unnecessary exploration during the reasoning process, enabling the model to focus more effectively on relevant information subsets. Moreover, EKA proves effective as a versatile, training-free inference strategy that scales seamlessly to large models. Generalization tests across diverse datasets and retrieval corpora confirm the robustness of our approach. Overall, EKA advances the state-of-the-art in iterative RAG systems while illuminating the critical interplay between structured reasoning and efficient exploration in reinforcement learning-augmented frameworks. The code is released at \href{this https URL}{Github}.
通过早期知识对齐实现多跳推理 / Multi-hop Reasoning via Early Knowledge Alignment
这篇论文提出了一种名为‘早期知识对齐’的简单有效方法,让大语言模型在分解复杂问题之前先了解可用的知识库信息,从而显著提升了多步问答系统的准确性和效率,减少了推理过程中的错误传递。
源自 arXiv: 2512.20144