菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-02-10
📄 Abstract - Causality in Video Diffusers is Separable from Denoising

Causality -- referring to temporal, uni-directional cause-effect relationships between components -- underlies many complex generative processes, including videos, language, and robot trajectories. Current causal diffusion models entangle temporal reasoning with iterative denoising, applying causal attention across all layers, at every denoising step, and over the entire context. In this paper, we show that the causal reasoning in these models is separable from the multi-step denoising process. Through systematic probing of autoregressive video diffusers, we uncover two key regularities: (1) early layers produce highly similar features across denoising steps, indicating redundant computation along the diffusion trajectory; and (2) deeper layers exhibit sparse cross-frame attention and primarily perform intra-frame rendering. Motivated by these findings, we introduce Separable Causal Diffusion (SCD), a new architecture that explicitly decouples once-per-frame temporal reasoning, via a causal transformer encoder, from multi-step frame-wise rendering, via a lightweight diffusion decoder. Extensive experiments on both pretraining and post-training tasks across synthetic and real benchmarks show that SCD significantly improves throughput and per-frame latency while matching or surpassing the generation quality of strong causal diffusion baselines.

顶级标签: video generation model training theory
详细标签: causal reasoning video diffusion architecture design computational efficiency temporal modeling 或 搜索:

视频扩散模型中的因果性与去噪过程是可分离的 / Causality in Video Diffusers is Separable from Denoising


1️⃣ 一句话总结

这篇论文发现,在生成视频的扩散模型中,负责理解时间先后顺序的因果推理部分,可以和负责逐帧绘制细节的多步去噪过程分离开来,并据此提出了一种新的高效架构,在保证生成质量的同时大幅提升了处理速度。

源自 arXiv: 2602.10095