菜单

关于 🐙 GitHub
arXiv 提交日期: 2025-12-22
📄 Abstract - Over++: Generative Video Compositing for Layer Interaction Effects

In professional video compositing workflows, artists must manually create environmental interactions-such as shadows, reflections, dust, and splashes-between foreground subjects and background layers. Existing video generative models struggle to preserve the input video while adding such effects, and current video inpainting methods either require costly per-frame masks or yield implausible results. We introduce augmented compositing, a new task that synthesizes realistic, semi-transparent environmental effects conditioned on text prompts and input video layers, while preserving the original scene. To address this task, we present Over++, a video effect generation framework that makes no assumptions about camera pose, scene stationarity, or depth supervision. We construct a paired effect dataset tailored for this task and introduce an unpaired augmentation strategy that preserves text-driven editability. Our method also supports optional mask control and keyframe guidance without requiring dense annotations. Despite training on limited data, Over++ produces diverse and realistic environmental effects and outperforms existing baselines in both effect generation and scene preservation.

顶级标签: video generation computer vision multi-modal
详细标签: video compositing layer interaction environmental effects text-to-video video inpainting 或 搜索:

Over++:用于图层交互效果生成的视频合成框架 / Over++: Generative Video Compositing for Layer Interaction Effects


1️⃣ 一句话总结

这篇论文提出了一个名为Over++的视频生成框架,它能够根据文字描述,在输入视频的前景和背景图层之间自动合成逼真的半透明环境交互效果(如阴影、倒影),同时保持原始场景内容不变,解决了现有方法难以兼顾效果生成与场景保留的难题。

源自 arXiv: 2512.19661