菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-02-03
📄 Abstract - AnyStyle: Single-Pass Multimodal Stylization for 3D Gaussian Splatting

The growing demand for rapid and scalable 3D asset creation has driven interest in feed-forward 3D reconstruction methods, with 3D Gaussian Splatting (3DGS) emerging as an effective scene representation. While recent approaches have demonstrated pose-free reconstruction from unposed image collections, integrating stylization or appearance control into such pipelines remains underexplored. Existing attempts largely rely on image-based conditioning, which limits both controllability and flexibility. In this work, we introduce AnyStyle, a feed-forward 3D reconstruction and stylization framework that enables pose-free, zero-shot stylization through multimodal conditioning. Our method supports both textual and visual style inputs, allowing users to control the scene appearance using natural language descriptions or reference images. We propose a modular stylization architecture that requires only minimal architectural modifications and can be integrated into existing feed-forward 3D reconstruction backbones. Experiments demonstrate that AnyStyle improves style controllability over prior feed-forward stylization methods while preserving high-quality geometric reconstruction. A user study further confirms that AnyStyle achieves superior stylization quality compared to an existing state-of-the-art approach. Repository: this https URL.

顶级标签: computer vision aigc model training
详细标签: 3d reconstruction 3d gaussian splatting multimodal stylization feed-forward zero-shot 或 搜索:

AnyStyle:面向3D高斯泼溅的单次多模态风格化方法 / AnyStyle: Single-Pass Multimodal Stylization for 3D Gaussian Splatting


1️⃣ 一句话总结

这篇论文提出了一个名为AnyStyle的前馈式3D重建与风格化框架,它能够仅通过一次处理,就利用文本描述或参考图片等不同模态的输入,为3D高斯泼溅模型实现无需特定拍摄姿态的、高质量的零样本风格化控制。

源自 arXiv: 2602.04043