MemFlow:用于一致且高效长视频叙事的自适应流动记忆 / MemFlow: Flowing Adaptive Memory for Consistent and Efficient Long Video Narratives
1️⃣ 一句话总结
这篇论文提出了一个名为MemFlow的新方法,它通过动态选择和激活与当前生成内容最相关的历史画面来管理记忆,从而在生成超长视频时,既能高效地保持故事内容的连贯性,又几乎不影响生成速度。
The core challenge for streaming video generation is maintaining the content consistency in long context, which poses high requirement for the memory design. Most existing solutions maintain the memory by compressing historical frames with predefined strategies. However, different to-generate video chunks should refer to different historical cues, which is hard to satisfy with fixed strategies. In this work, we propose MemFlow to address this problem. Specifically, before generating the coming chunk, we dynamically update the memory bank by retrieving the most relevant historical frames with the text prompt of this chunk. This design enables narrative coherence even if new event happens or scenario switches in future frames. In addition, during generation, we only activate the most relevant tokens in the memory bank for each query in the attention layers, which effectively guarantees the generation efficiency. In this way, MemFlow achieves outstanding long-context consistency with negligible computation burden (7.9% speed reduction compared with the memory-free baseline) and keeps the compatibility with any streaming video generation model with KV cache.
MemFlow:用于一致且高效长视频叙事的自适应流动记忆 / MemFlow: Flowing Adaptive Memory for Consistent and Efficient Long Video Narratives
这篇论文提出了一个名为MemFlow的新方法,它通过动态选择和激活与当前生成内容最相关的历史画面来管理记忆,从而在生成超长视频时,既能高效地保持故事内容的连贯性,又几乎不影响生成速度。
源自 arXiv: 2512.14699