视觉说服力:什么影响了视觉-语言模型的决策? / Visual Persuasion: What Influences Decisions of Vision-Language Models?
1️⃣ 一句话总结
这篇论文通过让视觉-语言模型在受控的图片选择任务中进行测试,并系统性地修改图片内容,揭示并解释了这些AI模型在做视觉决策(如点击、推荐)时,其选择偏好会受到图片中构图、光线、背景等具体视觉元素的显著影响,从而为提前发现和审计AI的视觉决策漏洞提供了一种方法。
The web is littered with images, once created for human consumption and now increasingly interpreted by agents using vision-language models (VLMs). These agents make visual decisions at scale, deciding what to click, recommend, or buy. Yet, we know little about the structure of their visual preferences. We introduce a framework for studying this by placing VLMs in controlled image-based choice tasks and systematically perturbing their inputs. Our key idea is to treat the agent's decision function as a latent visual utility that can be inferred through revealed preference: choices between systematically edited images. Starting from common images, such as product photos, we propose methods for visual prompt optimization, adapting text optimization methods to iteratively propose and apply visually plausible modifications using an image generation model (such as in composition, lighting, or background). We then evaluate which edits increase selection probability. Through large-scale experiments on frontier VLMs, we demonstrate that optimized edits significantly shift choice probabilities in head-to-head comparisons. We develop an automatic interpretability pipeline to explain these preferences, identifying consistent visual themes that drive selection. We argue that this approach offers a practical and efficient way to surface visual vulnerabilities, safety concerns that might otherwise be discovered implicitly in the wild, supporting more proactive auditing and governance of image-based AI agents.
视觉说服力:什么影响了视觉-语言模型的决策? / Visual Persuasion: What Influences Decisions of Vision-Language Models?
这篇论文通过让视觉-语言模型在受控的图片选择任务中进行测试,并系统性地修改图片内容,揭示并解释了这些AI模型在做视觉决策(如点击、推荐)时,其选择偏好会受到图片中构图、光线、背景等具体视觉元素的显著影响,从而为提前发现和审计AI的视觉决策漏洞提供了一种方法。
源自 arXiv: 2602.15278