菜单

🤖 系统
📄 Abstract - SAM 3D: 3Dfy Anything in Images

We present SAM 3D, a generative model for visually grounded 3D object reconstruction, predicting geometry, texture, and layout from a single image. SAM 3D excels in natural images, where occlusion and scene clutter are common and visual recognition cues from context play a larger role. We achieve this with a human- and model-in-the-loop pipeline for annotating object shape, texture, and pose, providing visually grounded 3D reconstruction data at unprecedented scale. We learn from this data in a modern, multi-stage training framework that combines synthetic pretraining with real-world alignment, breaking the 3D "data barrier". We obtain significant gains over recent work, with at least a 5:1 win rate in human preference tests on real-world objects and scenes. We will release our code and model weights, an online demo, and a new challenging benchmark for in-the-wild 3D object reconstruction.

顶级标签: computer vision model training aigc
详细标签: 3d reconstruction single image generative model object geometry texture prediction 或 搜索:

📄 论文总结

SAM 3D:图像中任意物体的三维化 / SAM 3D: 3Dfy Anything in Images


1️⃣ 一句话总结

这篇论文提出了一个名为SAM 3D的生成模型,能够仅凭一张图片就重建出物体的三维形状、纹理和布局,尤其在处理遮挡多、背景复杂的真实场景时表现优异,并通过创新的数据标注和训练方法大幅提升了重建效果。


📄 打开原文 PDF