菜单

🤖 系统
📄 Abstract - Aligned but Stereotypical? The Hidden Influence of System Prompts on Social Bias in LVLM-Based Text-to-Image Models

Large vision-language model (LVLM) based text-to-image (T2I) systems have become the dominant paradigm in image generation, yet whether they amplify social biases remains insufficiently understood. In this paper, we show that LVLM-based models produce markedly more socially biased images than non-LVLM-based models. We introduce a 1,024 prompt benchmark spanning four levels of linguistic complexity and evaluate demographic bias across multiple attributes in a systematic manner. Our analysis identifies system prompts, the predefined instructions guiding LVLMs, as a primary driver of biased behavior. Through decoded intermediate representations, token-probability diagnostics, and embedding-association analyses, we reveal how system prompts encode demographic priors that propagate into image synthesis. To this end, we propose FairPro, a training-free meta-prompting framework that enables LVLMs to self-audit and construct fairness-aware system prompts at test time. Experiments on two LVLM-based T2I models, SANA and Qwen-Image, show that FairPro substantially reduces demographic bias while preserving text-image alignment. We believe our findings provide deeper insight into the central role of system prompts in bias propagation and offer a practical, deployable approach for building more socially responsible T2I systems.

顶级标签: llm multi-modal model evaluation
详细标签: social bias text-to-image system prompts fairness benchmark 或 搜索:

对齐但刻板?系统提示对基于LVLM的文生图模型中社会偏见的隐性影响 / Aligned but Stereotypical? The Hidden Influence of System Prompts on Social Bias in LVLM-Based Text-to-Image Models


1️⃣ 一句话总结

这篇论文发现,基于大视觉语言模型的文生图系统会因内置的‘系统提示’而产生严重的刻板印象偏见,并提出了一个无需额外训练的‘FairPro’框架,让模型能自我审查并生成更公平的图像。


📄 打开原文 PDF