Legal-DC:面向法律文档的检索增强生成基准评测 / Legal-DC: Benchmarking Retrieval-Augmented Generation for Legal Documents
1️⃣ 一句话总结
本研究针对中文法律场景,创建了一个专门的评测数据集并提出了一个能更好理解法律条文结构的智能问答框架,显著提升了法律文件咨询的准确性和可靠性。
Retrieval-Augmented Generation (RAG) has emerged as a promising technology for legal document consultation, yet its application in Chinese legal scenarios faces two key limitations: existing benchmarks lack specialized support for joint retriever-generator evaluation, and mainstream RAG systems often fail to accommodate the structured nature of legal provisions. To address these gaps, this study advances two core contributions: First, we constructed the Legal-DC benchmark dataset, comprising 480 legal documents (covering areas such as market regulation and contract management) and 2,475 refined question-answer pairs, each annotated with clause-level references, filling the gap for specialized evaluation resources in Chinese legal RAG. Second, we propose the LegRAG framework, which integrates legal adaptive indexing (clause-boundary segmentation) with a dual-path self-reflection mechanism to ensure clause integrity while enhancing answer accuracy. Third, we introduce automated evaluation methods for large language models to meet the high-reliability demands of legal retrieval scenarios. LegRAG outperforms existing state-of-the-art methods by 1.3% to 5.6% across key evaluation metrics. This research provides a specialized benchmark, practical framework, and empirical insights to advance the development of Chinese legal RAG systems. Our code and data are available at this https URL.
Legal-DC:面向法律文档的检索增强生成基准评测 / Legal-DC: Benchmarking Retrieval-Augmented Generation for Legal Documents
本研究针对中文法律场景,创建了一个专门的评测数据集并提出了一个能更好理解法律条文结构的智能问答框架,显著提升了法律文件咨询的准确性和可靠性。
源自 arXiv: 2603.11772