菜单

关于 🐙 GitHub
arXiv 提交日期: 2025-12-18
📄 Abstract - TurboDiffusion: Accelerating Video Diffusion Models by 100-200 Times

We introduce TurboDiffusion, a video generation acceleration framework that can speed up end-to-end diffusion generation by 100-200x while maintaining video quality. TurboDiffusion mainly relies on several components for acceleration: (1) Attention acceleration: TurboDiffusion uses low-bit SageAttention and trainable Sparse-Linear Attention (SLA) to speed up attention computation. (2) Step distillation: TurboDiffusion adopts rCM for efficient step distillation. (3) W8A8 quantization: TurboDiffusion quantizes model parameters and activations to 8 bits to accelerate linear layers and compress the model. In addition, TurboDiffusion incorporates several other engineering optimizations. We conduct experiments on the Wan2.2-I2V-14B-720P, Wan2.1-T2V-1.3B-480P, Wan2.1-T2V-14B-720P, and Wan2.1-T2V-14B-480P models. Experimental results show that TurboDiffusion achieves 100-200x speedup for video generation even on a single RTX 5090 GPU, while maintaining comparable video quality. The GitHub repository, which includes model checkpoints and easy-to-use code, is available at this https URL.

顶级标签: video generation model training aigc
详细标签: diffusion acceleration attention optimization model quantization video diffusion inference speedup 或 搜索:

TurboDiffusion:将视频扩散模型加速100-200倍 / TurboDiffusion: Accelerating Video Diffusion Models by 100-200 Times


1️⃣ 一句话总结

这篇论文提出了一个名为TurboDiffusion的框架,它通过优化注意力计算、减少生成步骤和压缩模型等技术,在保证视频质量的同时,将视频生成速度提升了100到200倍。

源自 arXiv: 2512.16093