菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-01-26
📄 Abstract - FABLE: Forest-Based Adaptive Bi-Path LLM-Enhanced Retrieval for Multi-Document Reasoning

The rapid expansion of long-context Large Language Models (LLMs) has reignited debate on whether Retrieval-Augmented Generation (RAG) remains necessary. However, empirical evidence reveals persistent limitations of long-context inference, including the lost-in-the-middle phenomenon, high computational cost, and poor scalability for multi-document reasoning. Conversely, traditional RAG systems, while efficient, are constrained by flat chunk-level retrieval that introduces semantic noise and fails to support structured cross-document synthesis. We present \textbf{FABLE}, a \textbf{F}orest-based \textbf{A}daptive \textbf{B}i-path \textbf{L}LM-\textbf{E}nhanced retrieval framework that integrates LLMs into both knowledge organization and retrieval. FABLE constructs LLM-enhanced hierarchical forest indexes with multi-granularity semantic structures, then employs a bi-path strategy combining LLM-guided hierarchical traversal with structure-aware propagation for fine-grained evidence acquisition, with explicit budget control for adaptive efficiency trade-offs. Extensive experiments demonstrate that FABLE consistently outperforms SOTA RAG methods and achieves comparable accuracy to full-context LLM inference with up to 94\% token reduction, showing that long-context LLMs amplify rather than fully replace the need for structured retrieval.

顶级标签: llm natural language processing systems
详细标签: retrieval-augmented generation multi-document reasoning hierarchical indexing long-context llms efficiency optimization 或 搜索:

FABLE:一种基于森林结构的自适应双路径大语言模型增强检索框架,用于多文档推理 / FABLE: Forest-Based Adaptive Bi-Path LLM-Enhanced Retrieval for Multi-Document Reasoning


1️⃣ 一句话总结

这篇论文提出了一个名为FABLE的新型检索框架,它通过让大语言模型参与构建层次化的文档‘森林’索引,并结合两种智能检索路径,在显著减少计算开销的同时,实现了与让大模型直接阅读全部长文档相媲美的多文档推理效果。

源自 arXiv: 2601.18116