GSAR:面向多智能体大语言模型幻觉检测与恢复的带类型依据评分框架 / GSAR: Typed Grounding for Hallucination Detection and Recovery in Multi-Agent LLMs
1️⃣ 一句话总结
本文提出GSAR框架,通过将模型陈述分为四类依据(有据、无据、矛盾、补充),并引入加权评分与三级决策(继续、重生成、重规划)机制,有效检测和修复多智能体大语言模型生成内容中的幻觉问题,在多种独立评估模型上均验证了其稳定性和有效性。
Autonomous multi-agent LLM systems are increasingly deployed to investigate operational incidents and produce structured diagnostic reports. Their trustworthiness hinges on whether each claim is grounded in observed evidence rather than model-internal inference. Existing groundedness evaluators (binary classifiers, LLM-as-judge scalars, self-correction loops) treat supporting evidence as interchangeable and emit a single signal that offers no principled control over downstream action. We present GSAR, a grounding-evaluation and replanning framework that (i) partitions claims into a four-way typology (grounded, ungrounded, contradicted, complementary), giving first-class standing to non-redundant alternative perspectives; (ii) assigns evidence-type-specific weights reflecting epistemic strength; (iii) computes an asymmetric contradiction-penalised weighted groundedness score; and (iv) couples that score to a three-tier decision function (proceed, regenerate, replan) driving a bounded-iteration outer loop under an explicit compute budget. We formalise the algorithm, prove six structural properties, and evaluate five design claims on FEVER with gold Wikipedia evidence under four independently-trained LLM judges (gpt-5.4, claude-sonnet-4-6, claude-opus-4-7, gemini-2.5-pro). Every ablation reproduces in the same direction on every judge: bootstrap 95% CIs on the rho=0 effect exclude 0 on all four; the no-complementary ablation under Opus 4.7 has CI [-96,-68] of 200; at n=1000 three independent judges converge to DeltaS(rho=0)=+0.058. A head-to-head against Vectara HHEM-2.1-Open is included. To our knowledge, GSAR is the first published groundedness framework coupling evidence-typed scoring with tiered recovery under an explicit compute budget.
GSAR:面向多智能体大语言模型幻觉检测与恢复的带类型依据评分框架 / GSAR: Typed Grounding for Hallucination Detection and Recovery in Multi-Agent LLMs
本文提出GSAR框架,通过将模型陈述分为四类依据(有据、无据、矛盾、补充),并引入加权评分与三级决策(继续、重生成、重规划)机制,有效检测和修复多智能体大语言模型生成内容中的幻觉问题,在多种独立评估模型上均验证了其稳定性和有效性。
源自 arXiv: 2604.23366