运动3到4:用于4D合成的3D运动重建 / Motion 3-to-4: 3D Motion Reconstruction for 4D Synthesis
1️⃣ 一句话总结
这篇论文提出了一个名为Motion 3-to-4的新方法,它能够仅用一段普通手机拍摄的视频,就能自动生成一个既包含三维形状又能流畅运动的动态数字物体,解决了从单一视角视频中同时还原物体形状和运动轨迹的难题。
We present Motion 3-to-4, a feed-forward framework for synthesising high-quality 4D dynamic objects from a single monocular video and an optional 3D reference mesh. While recent advances have significantly improved 2D, video, and 3D content generation, 4D synthesis remains difficult due to limited training data and the inherent ambiguity of recovering geometry and motion from a monocular viewpoint. Motion 3-to-4 addresses these challenges by decomposing 4D synthesis into static 3D shape generation and motion reconstruction. Using a canonical reference mesh, our model learns a compact motion latent representation and predicts per-frame vertex trajectories to recover complete, temporally coherent geometry. A scalable frame-wise transformer further enables robustness to varying sequence lengths. Evaluations on both standard benchmarks and a new dataset with accurate ground-truth geometry show that Motion 3-to-4 delivers superior fidelity and spatial consistency compared to prior work. Project page is available at this https URL.
运动3到4:用于4D合成的3D运动重建 / Motion 3-to-4: 3D Motion Reconstruction for 4D Synthesis
这篇论文提出了一个名为Motion 3-to-4的新方法,它能够仅用一段普通手机拍摄的视频,就能自动生成一个既包含三维形状又能流畅运动的动态数字物体,解决了从单一视角视频中同时还原物体形状和运动轨迹的难题。
源自 arXiv: 2601.14253