菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-01-14
📄 Abstract - V-DPM: 4D Video Reconstruction with Dynamic Point Maps

Powerful 3D representations such as DUSt3R invariant point maps, which encode 3D shape and camera parameters, have significantly advanced feed forward 3D reconstruction. While point maps assume static scenes, Dynamic Point Maps (DPMs) extend this concept to dynamic 3D content by additionally representing scene motion. However, existing DPMs are limited to image pairs and, like DUSt3R, require post processing via optimization when more than two views are involved. We argue that DPMs are more useful when applied to videos and introduce V-DPM to demonstrate this. First, we show how to formulate DPMs for video input in a way that maximizes representational power, facilitates neural prediction, and enables reuse of pretrained models. Second, we implement these ideas on top of VGGT, a recent and powerful 3D reconstructor. Although VGGT was trained on static scenes, we show that a modest amount of synthetic data is sufficient to adapt it into an effective V-DPM predictor. Our approach achieves state of the art performance in 3D and 4D reconstruction for dynamic scenes. In particular, unlike recent dynamic extensions of VGGT such as P3, DPMs recover not only dynamic depth but also the full 3D motion of every point in the scene.

顶级标签: computer vision multi-modal video
详细标签: 4d reconstruction dynamic scenes point maps video understanding 3d motion 或 搜索:

V-DPM:利用动态点图进行4D视频重建 / V-DPM: 4D Video Reconstruction with Dynamic Point Maps


1️⃣ 一句话总结

这篇论文提出了一种名为V-DPM的新方法,它通过将静态场景的3D重建技术扩展到视频领域,不仅能从视频中重建出动态场景的3D形状,还能精确追踪场景中每个点的完整3D运动轨迹,实现了更先进的4D(3D+时间)动态场景重建。

源自 arXiv: 2601.09499