菜单

关于 🐙 GitHub
arXiv 提交日期: 2025-12-15
📄 Abstract - MineTheGap: Automatic Mining of Biases in Text-to-Image Models

Text-to-Image (TTI) models generate images based on text prompts, which often leave certain aspects of the desired image ambiguous. When faced with these ambiguities, TTI models have been shown to exhibit biases in their interpretations. These biases can have societal impacts, e.g., when showing only a certain race for a stated occupation. They can also affect user experience when creating redundancy within a set of generated images instead of spanning diverse possibilities. Here, we introduce MineTheGap - a method for automatically mining prompts that cause a TTI model to generate biased outputs. Our method goes beyond merely detecting bias for a given prompt. Rather, it leverages a genetic algorithm to iteratively refine a pool of prompts, seeking for those that expose biases. This optimization process is driven by a novel bias score, which ranks biases according to their severity, as we validate on a dataset with known biases. For a given prompt, this score is obtained by comparing the distribution of generated images to the distribution of LLM-generated texts that constitute variations on the prompt. Code and examples are available on the project's webpage.

顶级标签: aigc model evaluation multi-modal
详细标签: text-to-image bias detection genetic algorithm bias severity llm evaluation 或 搜索:

挖掘差距:文本到图像模型中偏见的自动挖掘 / MineTheGap: Automatic Mining of Biases in Text-to-Image Models


1️⃣ 一句话总结

这篇论文提出了一种名为MineTheGap的自动化方法,它利用遗传算法和一种新的偏见评分机制,主动寻找并评估文本到图像模型在生成图片时可能暴露出的社会偏见(如职业与种族的刻板关联)或多样性不足等问题,而不仅仅是检测已知提示下的偏见。

源自 arXiv: 2512.13427