菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-03-26
📄 Abstract - The Anatomy of Uncertainty in LLMs

Understanding why a large language model (LLM) is uncertain about the response is important for their reliable deployment. Current approaches, which either provide a single uncertainty score or rely on the classical aleatoric-epistemic dichotomy, fail to offer actionable insights for improving the generative model. Recent studies have also shown that such methods are not enough for understanding uncertainty in LLMs. In this work, we advocate for an uncertainty decomposition framework that dissects LLM uncertainty into three distinct semantic components: (i) input ambiguity, arising from ambiguous prompts; (ii) knowledge gaps, caused by insufficient parametric evidence; and (iii) decoding randomness, stemming from stochastic sampling. Through a series of experiments we demonstrate that the dominance of these components can shift across model size and task. Our framework provides a better understanding to audit LLM reliability and detect hallucinations, paving the way for targeted interventions and more trustworthy systems.

顶级标签: llm model evaluation theory
详细标签: uncertainty decomposition hallucination detection model reliability generative models trustworthy ai 或 搜索:

大语言模型不确定性的解剖分析 / The Anatomy of Uncertainty in LLMs


1️⃣ 一句话总结

这篇论文提出了一个将大语言模型的不确定性分解为输入歧义、知识缺失和随机解码三个具体来源的新框架,以帮助更好地评估模型可靠性并减少幻觉,从而构建更可信的系统。

源自 arXiv: 2603.24967