自由几何:从自身更长版本中优化三维重建 / Free Geometry: Refining 3D Reconstruction from Longer Versions of Itself
1️⃣ 一句话总结
这篇论文提出了一种名为‘自由几何’的新方法,能让现成的三维重建AI模型在遇到新场景时,通过观察更多角度的图像来自我学习和快速调整,从而显著提升重建的准确性和稳定性,整个过程无需额外的三维数据标注。
Feed-forward 3D reconstruction models are efficient but rigid: once trained, they perform inference in a zero-shot manner and cannot adapt to the test scene. As a result, visually plausible reconstructions often contain errors, particularly under occlusions, specularities, and ambiguous cues. To address this, we introduce Free Geometry, a framework that enables feed-forward 3D reconstruction models to self-evolve at test time without any 3D ground truth. Our key insight is that, when the model receives more views, it produces more reliable and view-consistent reconstructions. Leveraging this property, given a testing sequence, we mask a subset of frames to construct a self-supervised task. Free Geometry enforces cross-view feature consistency between representations from full and partial observations, while maintaining the pairwise relations implied by the held-out frames. This self-supervision allows for fast recalibration via lightweight LoRA updates, taking less than 2 minutes per dataset on a single GPU. Our approach consistently improves state-of-the-art foundation models, including Depth Anything 3 and VGGT, across 4 benchmark datasets, yielding an average improvement of 3.73% in camera pose accuracy and 2.88% in point map prediction. Code is available at this https URL .
自由几何:从自身更长版本中优化三维重建 / Free Geometry: Refining 3D Reconstruction from Longer Versions of Itself
这篇论文提出了一种名为‘自由几何’的新方法,能让现成的三维重建AI模型在遇到新场景时,通过观察更多角度的图像来自我学习和快速调整,从而显著提升重建的准确性和稳定性,整个过程无需额外的三维数据标注。
源自 arXiv: 2604.14048