GenLCA:基于扩散模型从真实世界视频生成全身虚拟化身 / GenLCA: 3D Diffusion for Full-Body Avatars from In-the-Wild Videos
1️⃣ 一句话总结
这篇论文提出了一个名为GenLCA的新方法,它能够利用海量的普通网络视频,训练出一个高质量的3D扩散模型,从而仅凭文字或图片就能生成并编辑逼真且能流畅动画的全身虚拟数字人。
We present GenLCA, a diffusion-based generative model for generating and editing photorealistic full-body avatars from text and image inputs. The generated avatars are faithful to the inputs, while supporting high-fidelity facial and full-body animations. The core idea is a novel paradigm that enables training a full-body 3D diffusion model from partially observable 2D data, allowing the training dataset to scale to millions of real-world videos. This scalability contributes to the superior photorealism and generalizability of GenLCA. Specifically, we scale up the dataset by repurposing a pretrained feed-forward avatar reconstruction model as an animatable 3D tokenizer, which encodes unstructured video frames into structured 3D tokens. However, most real-world videos only provide partial observations of body parts, resulting in excessive blurring or transparency artifacts in the 3D tokens. To address this, we propose a novel visibility-aware diffusion training strategy that replaces invalid regions with learnable tokens and computes losses only over valid regions. We then train a flow-based diffusion model on the token dataset, inherently maintaining the photorealism and animatability provided by the pretrained avatar reconstruction model. Our approach effectively enables the use of large-scale real-world video data to train a diffusion model natively in 3D. We demonstrate the efficacy of our method through diverse and high-fidelity generation and editing results, outperforming existing solutions by a large margin. The project page is available at this https URL.
GenLCA:基于扩散模型从真实世界视频生成全身虚拟化身 / GenLCA: 3D Diffusion for Full-Body Avatars from In-the-Wild Videos
这篇论文提出了一个名为GenLCA的新方法,它能够利用海量的普通网络视频,训练出一个高质量的3D扩散模型,从而仅凭文字或图片就能生成并编辑逼真且能流畅动画的全身虚拟数字人。
源自 arXiv: 2604.07273