菜单

关于 🐙 GitHub
arXiv 提交日期: 2025-12-09
📄 Abstract - WonderZoom: Multi-Scale 3D World Generation

We present WonderZoom, a novel approach to generating 3D scenes with contents across multiple spatial scales from a single image. Existing 3D world generation models remain limited to single-scale synthesis and cannot produce coherent scene contents at varying granularities. The fundamental challenge is the lack of a scale-aware 3D representation capable of generating and rendering content with largely different spatial sizes. WonderZoom addresses this through two key innovations: (1) scale-adaptive Gaussian surfels for generating and real-time rendering of multi-scale 3D scenes, and (2) a progressive detail synthesizer that iteratively generates finer-scale 3D contents. Our approach enables users to "zoom into" a 3D region and auto-regressively synthesize previously non-existent fine details from landscapes to microscopic features. Experiments demonstrate that WonderZoom significantly outperforms state-of-the-art video and 3D models in both quality and alignment, enabling multi-scale 3D world creation from a single image. We show video results and an interactive viewer of generated multi-scale 3D worlds in this https URL

顶级标签: computer vision aigc multi-modal
详细标签: 3d generation multi-scale synthesis gaussian splatting scene representation image-to-3d 或 搜索:

WonderZoom:多尺度三维世界生成 / WonderZoom: Multi-Scale 3D World Generation


1️⃣ 一句话总结

这篇论文提出了一种名为WonderZoom的新方法,能够仅从一张图片生成包含从宏观景观到微观细节的多尺度、连贯的三维场景,解决了现有技术只能生成单一尺度内容的局限。


源自 arXiv: 2512.09164