长文本语言模型输出的细粒度不确定性量化:一项比较研究 / Fine-Grained Uncertainty Quantification for Long-Form Language Model Outputs: A Comparative Study
1️⃣ 一句话总结
这篇论文提出了一个专门用于评估长文本AI生成内容可信度的新框架,通过比较不同方法发现,基于简单“主张-回应”逻辑关系的一致性检验效果最好,并且这种不确定性评估能有效提升AI生成长文本的事实准确性。
Uncertainty quantification has emerged as an effective approach to closed-book hallucination detection for LLMs, but existing methods are largely designed for short-form outputs and do not generalize well to long-form generation. We introduce a taxonomy for fine-grained uncertainty quantification in long-form LLM outputs that distinguishes methods by design choices at three stages: response decomposition, unit-level scoring, and response-level aggregation. We formalize several families of consistency-based black-box scorers, providing generalizations and extensions of existing methods. In our experiments across multiple LLMs and datasets, we find 1) claim-response entailment consistently performs better or on par with more complex claim-level scorers, 2) claim-level scoring generally yields better results than sentence-level scoring, and 3) uncertainty-aware decoding is highly effective for improving the factuality of long-form outputs. Our framework clarifies relationships between prior methods, enables apples-to-apples comparisons, and provides practical guidance for selecting components for fine-grained UQ.
长文本语言模型输出的细粒度不确定性量化:一项比较研究 / Fine-Grained Uncertainty Quantification for Long-Form Language Model Outputs: A Comparative Study
这篇论文提出了一个专门用于评估长文本AI生成内容可信度的新框架,通过比较不同方法发现,基于简单“主张-回应”逻辑关系的一致性检验效果最好,并且这种不确定性评估能有效提升AI生成长文本的事实准确性。
源自 arXiv: 2602.17431