菜单

关于 🐙 GitHub
arXiv 提交日期: 2025-12-17
📄 Abstract - EmoCaliber: Advancing Reliable Visual Emotion Comprehension via Confidence Verbalization and Calibration

Visual Emotion Comprehension (VEC) aims to infer sentiment polarities or emotion categories from affective cues embedded in images. In recent years, Multimodal Large Language Models (MLLMs) have established a popular paradigm in VEC, leveraging their generalizability to unify VEC tasks defined under diverse emotion taxonomies. While this paradigm achieves notable success, it typically formulates VEC as a deterministic task, requiring the model to output a single, definitive emotion label for each image. Such a formulation insufficiently accounts for the inherent subjectivity of emotion perception, overlooking alternative interpretations that may be equally plausible to different viewers. To address this limitation, we propose equipping MLLMs with capabilities to verbalize their confidence in emotion predictions. This additional signal provides users with an estimate of both the plausibility of alternative interpretations and the MLLMs' self-assessed competence, thereby enhancing reliability in practice. Building on this insight, we introduce a three-stage training framework that progressively endows with structured reasoning, teaches to verbalize confidence, and calibrates confidence expression, culminating in EmoCaliber, a confidence-aware MLLM for VEC. Through fair and comprehensive evaluations on the unified benchmark VECBench, EmoCaliber demonstrates overall superiority against existing methods in both emotion prediction and confidence estimation. These results validate the effectiveness of our approach and mark a feasible step toward more reliable VEC systems. Project page: this https URL.

顶级标签: multi-modal model evaluation natural language processing
详细标签: visual emotion comprehension confidence calibration multimodal llm emotion prediction reliability 或 搜索:

EmoCaliber:通过置信度表达与校准推进可靠的视觉情感理解 / EmoCaliber: Advancing Reliable Visual Emotion Comprehension via Confidence Verbalization and Calibration


1️⃣ 一句话总结

这篇论文提出了一个名为EmoCaliber的新模型,它通过让多模态大语言模型学会表达自己对情感预测的置信度,并校准这种表达,来应对视觉情感理解任务中固有的主观性,从而构建更可靠的系统。


源自 arXiv: 2512.15528