菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-04-28
📄 Abstract - Mutual Forcing: Dual-Mode Self-Evolution for Fast Autoregressive Audio-Video Character Generation

In this work, we propose Mutual Forcing, a framework for fast autoregressive audio-video generation with long-horizon audio-video synchronization. Our approach addresses two key challenges: joint audio-video modeling and fast autoregressive generation. To ease joint audio-video optimization, we adopt a two-stage training strategy: we first train uni-modal generators and then couple them into a unified audio-video model for joint training on paired data. For streaming generation, we ask whether a native fast causal audio-video model can be trained directly, instead of following existing streaming distillation pipelines that typically train a bidirectional model first and then convert it into a causal generator through multiple distillation stages. Our answer is Mutual Forcing, which builds directly on native autoregressive model and integrates few-step and multi-step generation within a single weight-shared model, enabling self-distillation and improved training-inference consistency. The multi-step mode improves the few-step mode via self-distillation, while the few-step mode generates historical context during training to improve training-inference consistency; because the two modes share parameters, these two effects reinforce each other within a single model. Compared with prior approaches such as Self-Forcing, Mutual Forcing removes the need for an additional bidirectional teacher model, supports more flexible training sequence lengths, reduces training overhead, and allows the model to improve directly from real paired data rather than a fixed teacher. Experiments show that Mutual Forcing matches or surpasses strong baselines that require around 50 sampling steps while using only 4 to 8 steps, demonstrating substantial advantages in both efficiency and quality. The project page is available at this https URL.

顶级标签: multi-modal model training systems
详细标签: audio-video generation autoregressive model self-distillation fast generation synchronization 或 搜索:

相互强制:面向快速自回归音视频人物生成的雙模自演化框架 / Mutual Forcing: Dual-Mode Self-Evolution for Fast Autoregressive Audio-Video Character Generation


1️⃣ 一句话总结

本文提出了一种名为“相互强制”的新方法,能够直接训练一个快速、自回归式的音视频联合生成模型,通过让模型同时使用少量步骤和多个步骤两种生成模式并共享参数、互相促进,从而在仅需4到8步采样的情况下,达到或超越传统需要约50步采样的方法,显著提升了音视频生成的效率和质量。

源自 arXiv: 2604.25819