菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-01-27
📄 Abstract - Semantic Uncertainty Quantification of Hallucinations in LLMs: A Quantum Tensor Network Based Method

Large language models (LLMs) exhibit strong generative capabilities but remain vulnerable to confabulations, fluent yet unreliable outputs that vary arbitrarily even under identical prompts. Leveraging a quantum tensor network based pipeline, we propose a quantum physics inspired uncertainty quantification framework that accounts for aleatoric uncertainty in token sequence probability for semantic equivalence based clustering of LLM generations. This offers a principled and interpretable scheme for hallucination detection. We further introduce an entropy maximization strategy that prioritizes high certainty, semantically coherent outputs and highlights entropy regions where LLM decisions are likely to be unreliable, offering practical guidelines for when human oversight is warranted. We evaluate the robustness of our scheme under different generation lengths and quantization levels, dimensions overlooked in prior studies, demonstrating that our approach remains reliable even in resource constrained deployments. A total of 116 experiments on TriviaQA, NQ, SVAMP, and SQuAD across multiple architectures including Mistral-7B, Mistral-7B-instruct, Falcon-rw-1b, LLaMA-3.2-1b, LLaMA-2-13b-chat, LLaMA-2-7b-chat, LLaMA-2-13b, and LLaMA-2-7b show consistent improvements in AUROC and AURAC over state of the art baselines.

顶级标签: llm model evaluation theory
详细标签: uncertainty quantification hallucination detection quantum tensor networks semantic equivalence entropy maximization 或 搜索:

大语言模型幻觉的语义不确定性量化:一种基于量子张量网络的方法 / Semantic Uncertainty Quantification of Hallucinations in LLMs: A Quantum Tensor Network Based Method


1️⃣ 一句话总结

这篇论文提出了一种受量子物理启发的、基于张量网络的不确定性量化方法,用于检测大语言模型生成的不可靠内容(幻觉),并通过最大化熵的策略来优先输出高确定性、语义连贯的结果,从而为需要人工监督的场景提供实用指导。

源自 arXiv: 2601.20026