菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-05-04
📄 Abstract - MooD: An Efficient VA-Driven Affective Image Editing Framework via Fine-Grained Semantic Control

Affective image editing (AIE) aims to edit visual content to evoke target emotions. However, existing methods often overlook inference efficiency and predominantly depend on discrete emotion representations, which to some extent limits their practical applicability and makes it challenging to capture complex and subtle human emotions. To tackle these gaps, we propose MooD, the first framework that directly leverages continuous Valence-Arousal (VA) values for fine-grained and efficient AIE. Specifically, we first introduce a VA-Aware retrieval strategy to bridge vague affective values and concrete visual semantics. Building upon this, MooD integrates visual transfer and semantic guidance to achieve controllable AIE. Furthermore, we construct AffectSet, a VA-annotated dataset to support model optimization and evaluation. Extensive qualitative and quantitative experimental results demonstrate that our MooD achieves superior performance in both affective controllability and visual fidelity while maintaining high efficiency. A series of ablation studies further reveal the crucial factors of our design. Our code and data will be made publicly open soon.

顶级标签: computer vision machine learning aigc
详细标签: affective image editing valence-arousal fine-grained control dataset efficiency 或 搜索:

MooD:一种基于效价-唤醒度驱动的高效情感图像编辑框架,实现细粒度语义控制 / MooD: An Efficient VA-Driven Affective Image Editing Framework via Fine-Grained Semantic Control


1️⃣ 一句话总结

本文提出了MooD,一个利用连续效价-唤醒度(VA)数值来高效编辑图像情感的新框架,通过VA感知检索和视觉-语义引导,实现了比传统离散情绪方法更精细、更可控的情感图像调整,并配套构建了VA标注数据集AffectSet来验证效果。

源自 arXiv: 2605.02521