DiffStyle3D:通过注意力优化实现一致的3D高斯风格化 / DiffStyle3D: Consistent 3D Gaussian Stylization via Attention Optimization
1️⃣ 一句话总结
这篇论文提出了一种名为DiffStyle3D的新方法,它通过直接优化扩散模型的注意力空间,并结合几何信息来保持多视角一致性,从而解决了现有3D风格迁移技术中风格不一致和训练不稳定的问题,能生成更高质量、更逼真的3D风格化内容。
3D style transfer enables the creation of visually expressive 3D content, enriching the visual appearance of 3D scenes and objects. However, existing VGG- and CLIP-based methods struggle to model multi-view consistency within the model itself, while diffusion-based approaches can capture such consistency but rely on denoising directions, leading to unstable training. To address these limitations, we propose DiffStyle3D, a novel diffusion-based paradigm for 3DGS style transfer that directly optimizes in the latent space. Specifically, we introduce an Attention-Aware Loss that performs style transfer by aligning style features in the self-attention space, while preserving original content through content feature alignment. Inspired by the geometric invariance of 3D stylization, we propose a Geometry-Guided Multi-View Consistency method that integrates geometric information into self-attention to enable cross-view correspondence modeling. Based on geometric information, we additionally construct a geometry-aware mask to prevent redundant optimization in overlapping regions across views, which further improves multi-view consistency. Extensive experiments show that DiffStyle3D outperforms state-of-the-art methods, achieving higher stylization quality and visual realism.
DiffStyle3D:通过注意力优化实现一致的3D高斯风格化 / DiffStyle3D: Consistent 3D Gaussian Stylization via Attention Optimization
这篇论文提出了一种名为DiffStyle3D的新方法,它通过直接优化扩散模型的注意力空间,并结合几何信息来保持多视角一致性,从而解决了现有3D风格迁移技术中风格不一致和训练不稳定的问题,能生成更高质量、更逼真的3D风格化内容。
源自 arXiv: 2601.19717