菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-01-29
📄 Abstract - Epistemic Uncertainty Quantification for Pre-trained VLMs via Riemannian Flow Matching

Vision-Language Models (VLMs) are typically deterministic in nature and lack intrinsic mechanisms to quantify epistemic uncertainty, which reflects the model's lack of knowledge or ignorance of its own representations. We theoretically motivate negative log-density of an embedding as a proxy for the epistemic uncertainty, where low-density regions signify model ignorance. The proposed method REPVLM computes the probability density on the hyperspherical manifold of the VLM embeddings using Riemannian Flow Matching. We empirically demonstrate that REPVLM achieves near-perfect correlation between uncertainty and prediction error, significantly outperforming existing baselines. Beyond classification, we also demonstrate that the model also provides a scalable metric for out-of-distribution detection and automated data curation.

顶级标签: multi-modal model evaluation machine learning
详细标签: uncertainty quantification vision-language models riemannian flow matching out-of-distribution detection embedding density 或 搜索:

基于黎曼流匹配的预训练视觉语言模型认知不确定性量化 / Epistemic Uncertainty Quantification for Pre-trained VLMs via Riemannian Flow Matching


1️⃣ 一句话总结

这篇论文提出了一种名为REPVLM的新方法,它通过黎曼流匹配技术来量化视觉语言模型在预测时的认知不确定性(即模型对自身知识盲区的认知),实验表明该方法能近乎完美地反映预测错误,并可用于识别未知数据和自动化数据筛选。

源自 arXiv: 2601.21662