LATA:用于医学视觉语言模型置信度预测的拉普拉斯辅助直推式适应方法 / LATA: Laplacian-Assisted Transductive Adaptation for Conformal Uncertainty in Medical VLMs
1️⃣ 一句话总结
本文提出了一种名为LATA的新方法,它能在不重新训练模型、也几乎不需要额外标注的情况下,有效提升医学视觉语言模型在陌生数据上预测结果的可信度,使其预测更准确、更稳定。
Medical vision-language models (VLMs) are strong zero-shot recognizers for medical imaging, but their reliability under domain shift hinges on calibrated uncertainty with guarantees. Split conformal prediction (SCP) offers finite-sample coverage, yet prediction sets often become large (low efficiency) and class-wise coverage unbalanced-high class-conditioned coverage gap (CCV), especially in few-shot, imbalanced regimes; moreover, naively adapting to calibration labels breaks exchangeability and voids guarantees. We propose \texttt{\textbf{LATA}} (Laplacian-Assisted Transductive Adaptation), a \textit{training- and label-free} refinement that operates on the joint calibration and test pool by smoothing zero-shot probabilities over an image-image k-NN graph using a small number of CCCP mean-field updates, preserving SCP validity via a deterministic transform. We further introduce a \textit{failure-aware} conformal score that plugs into the vision-language uncertainty (ViLU) framework, providing instance-level difficulty and label plausibility to improve prediction set efficiency and class-wise balance at fixed coverage. \texttt{\textbf{LATA}} is black-box (no VLM updates), compute-light (windowed transduction, no backprop), and includes an optional prior knob that can run strictly label-free or, if desired, in a label-informed variant using calibration marginals once. Across \textbf{three} medical VLMs and \textbf{nine} downstream tasks, \texttt{\textbf{LATA}} consistently reduces set size and CCV while matching or tightening target coverage, outperforming prior transductive baselines and narrowing the gap to label-using methods, while using far less compute. Comprehensive ablations and qualitative analyses show that \texttt{\textbf{LATA}} sharpens zero-shot predictions without compromising exchangeability.
LATA:用于医学视觉语言模型置信度预测的拉普拉斯辅助直推式适应方法 / LATA: Laplacian-Assisted Transductive Adaptation for Conformal Uncertainty in Medical VLMs
本文提出了一种名为LATA的新方法,它能在不重新训练模型、也几乎不需要额外标注的情况下,有效提升医学视觉语言模型在陌生数据上预测结果的可信度,使其预测更准确、更稳定。
源自 arXiv: 2602.17535