菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-01-29
📄 Abstract - Investigating Associational Biases in Inter-Model Communication of Large Generative Models

Social bias in generative AI can manifest not only as performance disparities but also as associational bias, whereby models learn and reproduce stereotypical associations between concepts and demographic groups, even in the absence of explicit demographic information (e.g., associating doctors with men). These associations can persist, propagate, and potentially amplify across repeated exchanges in inter-model communication pipelines, where one generative model's output becomes another's input. This is especially salient for human-centred perception tasks, such as human activity recognition and affect prediction, where inferences about behaviour and internal states can lead to errors or stereotypical associations that propagate into unequal treatment. In this work, focusing on human activity and affective expression, we study how such associations evolve within an inter-model communication pipeline that alternates between image generation and image description. Using the RAF-DB and PHASE datasets, we quantify demographic distribution drift induced by model-to-model information exchange and assess whether these drifts are systematic using an explainability pipeline. Our results reveal demographic drifts toward younger representations for both actions and emotions, as well as toward more female-presenting representations, primarily for emotions. We further find evidence that some predictions are supported by spurious visual regions (e.g., background or hair) rather than concept-relevant cues (e.g., body or face). We also examine whether these demographic drifts translate into measurable differences in downstream behaviour, i.e., while predicting activity and emotion labels. Finally, we outline mitigation strategies spanning data-centric, training and deployment interventions, and emphasise the need for careful safeguards when deploying interconnected models in human-centred AI systems.

顶级标签: multi-modal model evaluation aigc
详细标签: social bias associational bias inter-model communication demographic drift image generation 或 搜索:

探究大型生成模型间通信中的关联性偏见 / Investigating Associational Biases in Inter-Model Communication of Large Generative Models


1️⃣ 一句话总结

这项研究发现,在图像生成和描述交替进行的AI模型间通信中,模型会学习并传播关于人种、性别和年龄的刻板关联,导致下游任务(如识别人类活动和情绪)出现系统性偏见,并提出了相应的缓解策略。

源自 arXiv: 2601.22093