具有可控细粒度表情的高保真3D面部虚拟形象合成 / High-Fidelity 3D Facial Avatar Synthesis with Controllable Fine-Grained Expressions
1️⃣ 一句话总结
这篇论文提出了一种新方法,通过同时优化纹理和面部网格的生成,并利用文本提示进行引导,实现了对3D面部虚拟形象细微表情的精确控制。
Facial expression editing methods can be mainly categorized into two types based on their architectures: 2D-based and 3D-based methods. The former lacks 3D face modeling capabilities, making it difficult to edit 3D factors effectively. The latter has demonstrated superior performance in generating high-quality and view-consistent renderings using single-view 2D face images. Although these methods have successfully used animatable models to control facial expressions, they still have limitations in achieving precise control over fine-grained expressions. To address this issue, in this paper, we propose a novel approach by simultaneously refining both the latent code of a pretrained 3D-Aware GAN model for texture editing and the expression code of the driven 3DMM model for mesh editing. Specifically, we introduce a Dual Mappers module, comprising Texture Mapper and Emotion Mapper, to learn the transformations of the given latent code for textures and the expression code for meshes, respectively. To optimize the Dual Mappers, we propose a Text-Guided Optimization method, leveraging a CLIP-based objective function with expression text prompts as targets, while integrating a SubSpace Projection mechanism to project the text embedding to the expression subspace such that we can have more precise control over fine-grained expressions. Extensive experiments and comparative analyses demonstrate the effectiveness and superiority of our proposed method.
具有可控细粒度表情的高保真3D面部虚拟形象合成 / High-Fidelity 3D Facial Avatar Synthesis with Controllable Fine-Grained Expressions
这篇论文提出了一种新方法,通过同时优化纹理和面部网格的生成,并利用文本提示进行引导,实现了对3D面部虚拟形象细微表情的精确控制。
源自 arXiv: 2603.14781