菜单

关于 🐙 GitHub
arXiv 提交日期: 2025-12-12
📄 Abstract - Flowception: Temporally Expansive Flow Matching for Video Generation

We present Flowception, a novel non-autoregressive and variable-length video generation framework. Flowception learns a probability path that interleaves discrete frame insertions with continuous frame denoising. Compared to autoregressive methods, Flowception alleviates error accumulation/drift as the frame insertion mechanism during sampling serves as an efficient compression mechanism to handle long-term context. Compared to full-sequence flows, our method reduces FLOPs for training three-fold, while also being more amenable to local attention variants, and allowing to learn the length of videos jointly with their content. Quantitative experimental results show improved FVD and VBench metrics over autoregressive and full-sequence baselines, which is further validated with qualitative results. Finally, by learning to insert and denoise frames in a sequence, Flowception seamlessly integrates different tasks such as image-to-video generation and video interpolation.

顶级标签: video generation model training aigc
详细标签: flow matching non-autoregressive video interpolation temporal modeling variable-length generation 或 搜索:

Flowception:用于视频生成的时间扩展流匹配方法 / Flowception: Temporally Expansive Flow Matching for Video Generation


1️⃣ 一句话总结

这篇论文提出了一种名为Flowception的新型视频生成方法,它通过交替插入新帧和优化已有帧来高效生成高质量、长度可变的视频,相比传统方法减少了计算开销和误差累积,并能同时处理图像生成视频和视频插帧等任务。


源自 arXiv: 2512.11438