超越体素的三维编辑:从三维掩码与自建数据中学习 / Beyond Voxel 3D Editing: Learning from 3D Masks and Self-Constructed Data
1️⃣ 一句话总结
这篇论文提出了一个名为BVE的新框架,它通过创建一个专门的大规模数据集并引入无需标注的三维掩码技术,解决了现有三维编辑方法在修改范围、保真度和数据稀缺方面的难题,能够更高效、精准地根据文字提示编辑三维模型,同时保持未修改区域的原貌。
3D editing refers to the ability to apply local or global modifications to 3D assets. Effective 3D editing requires maintaining semantic consistency by performing localized changes according to prompts, while also preserving local invariance so that unchanged regions remain consistent with the original. However, existing approaches have significant limitations: multi-view editing methods incur losses when projecting back to 3D, while voxel-based editing is constrained in both the regions that can be modified and the scale of modifications. Moreover, the lack of sufficiently large editing datasets for training and evaluation remains a challenge. To address these challenges, we propose a Beyond Voxel 3D Editing (BVE) framework with a self-constructed large-scale dataset specifically tailored for 3D editing. Building upon this dataset, our model enhances a foundational image-to-3D generative architecture with lightweight, trainable modules, enabling efficient injection of textual semantics without the need for expensive full-model retraining. Furthermore, we introduce an annotation-free 3D masking strategy to preserve local invariance, maintaining the integrity of unchanged regions during editing. Extensive experiments demonstrate that BVE achieves superior performance in generating high-quality, text-aligned 3D assets, while faithfully retaining the visual characteristics of the original input.
超越体素的三维编辑:从三维掩码与自建数据中学习 / Beyond Voxel 3D Editing: Learning from 3D Masks and Self-Constructed Data
这篇论文提出了一个名为BVE的新框架,它通过创建一个专门的大规模数据集并引入无需标注的三维掩码技术,解决了现有三维编辑方法在修改范围、保真度和数据稀缺方面的难题,能够更高效、精准地根据文字提示编辑三维模型,同时保持未修改区域的原貌。
源自 arXiv: 2604.13688