菜单

关于 🐙 GitHub
arXiv 提交日期: 2025-12-12
📄 Abstract - PersonaLive! Expressive Portrait Image Animation for Live Streaming

Current diffusion-based portrait animation models predominantly focus on enhancing visual quality and expression realism, while overlooking generation latency and real-time performance, which restricts their application range in the live streaming scenario. We propose PersonaLive, a novel diffusion-based framework towards streaming real-time portrait animation with multi-stage training recipes. Specifically, we first adopt hybrid implicit signals, namely implicit facial representations and 3D implicit keypoints, to achieve expressive image-level motion control. Then, a fewer-step appearance distillation strategy is proposed to eliminate appearance redundancy in the denoising process, greatly improving inference efficiency. Finally, we introduce an autoregressive micro-chunk streaming generation paradigm equipped with a sliding training strategy and a historical keyframe mechanism to enable low-latency and stable long-term video generation. Extensive experiments demonstrate that PersonaLive achieves state-of-the-art performance with up to 7-22x speedup over prior diffusion-based portrait animation models.

顶级标签: computer vision video generation aigc
详细标签: portrait animation diffusion models real-time generation live streaming appearance distillation 或 搜索:

PersonaLive!:面向直播的富有表现力的人像图像动画 / PersonaLive! Expressive Portrait Image Animation for Live Streaming


1️⃣ 一句话总结

这篇论文提出了一种名为PersonaLive的新方法,它通过结合隐式面部控制、精简去噪步骤和流式生成技术,在保持人像动画高表现力的同时,实现了比现有方法快7到22倍的实时生成速度,从而解决了AI人像动画在直播场景中延迟过高、无法实时应用的核心难题。


源自 arXiv: 2512.11253