菜单

🤖 系统
📄 Abstract - Layer-Aware Video Composition via Split-then-Merge

We present Split-then-Merge (StM), a novel framework designed to enhance control in generative video composition and address its data scarcity problem. Unlike conventional methods relying on annotated datasets or handcrafted rules, StM splits a large corpus of unlabeled videos into dynamic foreground and background layers, then self-composes them to learn how dynamic subjects interact with diverse scenes. This process enables the model to learn the complex compositional dynamics required for realistic video generation. StM introduces a novel transformation-aware training pipeline that utilizes a multi-layer fusion and augmentation to achieve affordance-aware composition, alongside an identity-preservation loss that maintains foreground fidelity during blending. Experiments show StM outperforms SoTA methods in both quantitative benchmarks and in humans/VLLM-based qualitative evaluations. More details are available at our project page: this https URL

顶级标签: video generation aigc computer vision
详细标签: video composition layer-aware generation self-supervised learning foreground-background fusion affordance-aware composition 或 搜索:

基于拆分与合并的分层感知视频合成 / Layer-Aware Video Composition via Split-then-Merge


1️⃣ 一句话总结

这篇论文提出了一个名为‘拆分与合并’的新框架,它通过将大量无标签视频自动拆分为前景和背景层,并让它们相互组合来学习动态物体与场景的交互,从而无需人工标注就能生成更真实、可控的视频。


📄 打开原文 PDF