PixelSmile:迈向细粒度面部表情编辑 / PixelSmile: Toward Fine-Grained Facial Expression Editing
1️⃣ 一句话总结
这篇论文提出了一个名为PixelSmile的扩散模型框架,通过构建新数据集和采用对称联合训练等方法,解决了细粒度面部表情编辑中语义混淆的难题,实现了对表情强度连续、精确且不改变人物身份的线性控制。
Fine-grained facial expression editing has long been limited by intrinsic semantic overlap. To address this, we construct the Flex Facial Expression (FFE) dataset with continuous affective annotations and establish FFE-Bench to evaluate structural confusion, editing accuracy, linear controllability, and the trade-off between expression editing and identity preservation. We propose PixelSmile, a diffusion framework that disentangles expression semantics via fully symmetric joint training. PixelSmile combines intensity supervision with contrastive learning to produce stronger and more distinguishable expressions, achieving precise and stable linear expression control through textual latent interpolation. Extensive experiments demonstrate that PixelSmile achieves superior disentanglement and robust identity preservation, confirming its effectiveness for continuous, controllable, and fine-grained expression editing, while naturally supporting smooth expression blending.
PixelSmile:迈向细粒度面部表情编辑 / PixelSmile: Toward Fine-Grained Facial Expression Editing
这篇论文提出了一个名为PixelSmile的扩散模型框架,通过构建新数据集和采用对称联合训练等方法,解决了细粒度面部表情编辑中语义混淆的难题,实现了对表情强度连续、精确且不改变人物身份的线性控制。
源自 arXiv: 2603.25728