当视觉不是问题:评估视觉语言模型在误导性数据可视化上的表现 / When Visuals Aren't the Problem: Evaluating Vision-Language Models on Misleading Data Visualizations
1️⃣ 一句话总结
这篇论文通过创建一个包含多种误导性图表和说明的基准测试,评估了当前主流视觉语言模型,发现它们能较好地识别图表设计错误,但难以检测由说明文字中的逻辑推理错误引发的误导信息。
Visualizations help communicate data insights, but deceptive data representations can distort their interpretation and propagate misinformation. While recent Vision Language Models (VLMs) perform well on many chart understanding tasks, their ability to detect misleading visualizations, especially when deception arises from subtle reasoning errors in captions, remains poorly understood. Here, we evaluate VLMs on misleading visualization-caption pairs grounded in a fine-grained taxonomy of reasoning errors (e.g., Cherry-picking, Causal inference) and visualization design errors (e.g., Truncated axis, Dual axis, inappropriate encodings). To this end, we develop a benchmark that combines real-world visualization with human-authored, curated misleading captions designed to elicit specific reasoning and visualization error types, enabling controlled analysis across error categories and modalities of misleadingness. Evaluating many commercial and open-source VLMs, we find that models detect visual design errors substantially more reliably than reasoning-based misinformation, and frequently misclassify non-misleading visualizations as deceptive. Overall, our work fills a gap between coarse detection of misleading content and the attribution of the specific reasoning or visualization errors that give rise to it.
当视觉不是问题:评估视觉语言模型在误导性数据可视化上的表现 / When Visuals Aren't the Problem: Evaluating Vision-Language Models on Misleading Data Visualizations
这篇论文通过创建一个包含多种误导性图表和说明的基准测试,评估了当前主流视觉语言模型,发现它们能较好地识别图表设计错误,但难以检测由说明文字中的逻辑推理错误引发的误导信息。
源自 arXiv: 2603.22368