IC-Effect:基于上下文学习的精确高效视频特效编辑 / IC-Effect: Precise and Efficient Video Effects Editing via In-Context Learning
1️⃣ 一句话总结
这篇论文提出了一个名为IC-Effect的新框架,它能够根据简单的文字指令,仅用少量示例就能在视频中精准地添加火焰、粒子等复杂特效,同时完美保持背景不变和视频在时间上的连贯性,大大降低了高质量视频特效制作的门槛。
We propose \textbf{IC-Effect}, an instruction-guided, DiT-based framework for few-shot video VFX editing that synthesizes complex effects (\eg flames, particles and cartoon characters) while strictly preserving spatial and temporal consistency. Video VFX editing is highly challenging because injected effects must blend seamlessly with the background, the background must remain entirely unchanged, and effect patterns must be learned efficiently from limited paired data. However, existing video editing models fail to satisfy these requirements. IC-Effect leverages the source video as clean contextual conditions, exploiting the contextual learning capability of DiT models to achieve precise background preservation and natural effect injection. A two-stage training strategy, consisting of general editing adaptation followed by effect-specific learning via Effect-LoRA, ensures strong instruction following and robust effect modeling. To further improve efficiency, we introduce spatiotemporal sparse tokenization, enabling high fidelity with substantially reduced computation. We also release a paired VFX editing dataset spanning $15$ high-quality visual styles. Extensive experiments show that IC-Effect delivers high-quality, controllable, and temporally consistent VFX editing, opening new possibilities for video creation.
IC-Effect:基于上下文学习的精确高效视频特效编辑 / IC-Effect: Precise and Efficient Video Effects Editing via In-Context Learning
这篇论文提出了一个名为IC-Effect的新框架,它能够根据简单的文字指令,仅用少量示例就能在视频中精准地添加火焰、粒子等复杂特效,同时完美保持背景不变和视频在时间上的连贯性,大大降低了高质量视频特效制作的门槛。
源自 arXiv: 2512.15635