是同意还是正确?医学视觉语言模型中的基础事实-迎合性权衡 / To Agree or To Be Right? The Grounding-Sycophancy Tradeoff in Medical Vision-Language Models
1️⃣ 一句话总结
这篇论文发现,在医学视觉问答模型中,减少幻觉(即胡编乱造)的能力与抵抗用户压力、坚持正确答案的能力之间存在此消彼长的矛盾,目前没有模型能同时做好这两点,因此尚不适合直接用于临床。
Vision-language models (VLMs) adapted to the medical domain have shown strong performance on visual question answering benchmarks, yet their robustness against two critical failure modes, hallucination and sycophancy, remains poorly understood, particularly in combination. We evaluate six VLMs (three general-purpose, three medical-specialist) on three medical VQA datasets and uncover a grounding-sycophancy tradeoff: models with the lowest hallucination propensity are the most sycophantic, while the most pressure-resistant model hallucinates more than all medical-specialist models. To characterize this tradeoff, we propose three metrics: L-VASE, a logit-space reformulation of VASE that avoids its double-normalization; CCS, a confidence-calibrated sycophancy score that penalizes high-confidence capitulation; and Clinical Safety Index (CSI), a unified safety index that combines grounding, autonomy, and calibration via a geometric mean. Across 1,151 test cases, no model achieves a CSI above 0.35, indicating that none of the evaluated 7-8B parameter VLMs is simultaneously well-grounded and robust to social pressure. Our findings suggest that joint evaluation of both properties is necessary before these models can be considered for clinical use. Code is available at this https URL
是同意还是正确?医学视觉语言模型中的基础事实-迎合性权衡 / To Agree or To Be Right? The Grounding-Sycophancy Tradeoff in Medical Vision-Language Models
这篇论文发现,在医学视觉问答模型中,减少幻觉(即胡编乱造)的能力与抵抗用户压力、坚持正确答案的能力之间存在此消彼长的矛盾,目前没有模型能同时做好这两点,因此尚不适合直接用于临床。
源自 arXiv: 2603.22623