菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-01-07
📄 Abstract - Choreographing a World of Dynamic Objects

Dynamic objects in our physical 4D (3D + time) world are constantly evolving, deforming, and interacting with other objects, leading to diverse 4D scene dynamics. In this paper, we present a universal generative pipeline, CHORD, for CHOReographing Dynamic objects and scenes and synthesizing this type of phenomena. Traditional rule-based graphics pipelines to create these dynamics are based on category-specific heuristics, yet are labor-intensive and not scalable. Recent learning-based methods typically demand large-scale datasets, which may not cover all object categories in interest. Our approach instead inherits the universality from the video generative models by proposing a distillation-based pipeline to extract the rich Lagrangian motion information hidden in the Eulerian representations of 2D videos. Our method is universal, versatile, and category-agnostic. We demonstrate its effectiveness by conducting experiments to generate a diverse range of multi-body 4D dynamics, show its advantage compared to existing methods, and demonstrate its applicability in generating robotics manipulation policies. Project page: this https URL

顶级标签: computer vision video generation multi-modal
详细标签: 4d scene generation motion distillation lagrangian motion dynamic objects robotics manipulation 或 搜索:

编排动态物体的世界 / Choreographing a World of Dynamic Objects


1️⃣ 一句话总结

这篇论文提出了一个名为CHORD的通用生成式方法,能够从普通2D视频中提取物体运动信息,从而自动生成和模拟各种动态物体与场景的复杂运动和交互,无需依赖大量特定类别的数据或人工规则。

源自 arXiv: 2601.04194