菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-03-19
📄 Abstract - CustomTex: High-fidelity Indoor Scene Texturing via Multi-Reference Customization

The creation of high-fidelity, customizable 3D indoor scene textures remains a significant challenge. While text-driven methods offer flexibility, they lack the precision for fine-grained, instance-level control, and often produce textures with insufficient quality, artifacts, and baked-in shading. To overcome these limitations, we introduce CustomTex, a novel framework for instance-level, high-fidelity scene texturing driven by reference images. CustomTex takes an untextured 3D scene and a set of reference images specifying the desired appearance for each object instance, and generates a unified, high-resolution texture map. The core of our method is a dual-distillation approach that separates semantic control from pixel-level enhancement. We employ semantic-level distillation, equipped with an instance cross-attention, to ensure semantic plausibility and ``reference-instance'' alignment, and pixel-level distillation to enforce high visual fidelity. Both are unified within a Variational Score Distillation (VSD) optimization framework. Experiments demonstrate that CustomTex achieves precise instance-level consistency with reference images and produces textures with superior sharpness, reduced artifacts, and minimal baked-in shading compared to state-of-the-art methods. Our work establishes a more direct and user-friendly path to high-quality, customizable 3D scene appearance editing.

顶级标签: computer vision multi-modal model training
详细标签: 3d scene texturing reference image customization variational score distillation instance-level control texture generation 或 搜索:

CustomTex:基于多参考图像定制的高保真室内场景纹理生成 / CustomTex: High-fidelity Indoor Scene Texturing via Multi-Reference Customization


1️⃣ 一句话总结

这篇论文提出了一个名为CustomTex的新方法,它能让用户通过提供几张参考图片,就能为3D室内场景中的每个物体自动生成高清晰、无瑕疵且符合用户指定外观的高质量纹理。

源自 arXiv: 2603.19121