用于稀疏视角新视图合成的分层高斯泼溅中的多模态先验引导重要性采样 / Multimodal-Prior-Guided Importance Sampling for Hierarchical Gaussian Splatting in Sparse-View Novel View Synthesis
1️⃣ 一句话总结
这篇论文提出了一种新方法,通过融合图像颜色、语义和几何等多种信息来智能判断3D场景中哪些细节可以准确重建,从而在仅有少数几张照片的情况下,生成更清晰、更逼真的新视角图像。
We present multimodal-prior-guided importance sampling as the central mechanism for hierarchical 3D Gaussian Splatting (3DGS) in sparse-view novel view synthesis. Our sampler fuses complementary cues { -- } photometric rendering residuals, semantic priors, and geometric priors { -- } to produce a robust, local recoverability estimate that directly drives where to inject fine Gaussians. Built around this sampling core, our framework comprises (1) a coarse-to-fine Gaussian representation that encodes global shape with a stable coarse layer and selectively adds fine primitives where the multimodal metric indicates recoverable detail; and (2) a geometric-aware sampling and retention policy that concentrates refinement on geometrically critical and complex regions while protecting newly added primitives in underconstrained areas from premature pruning. By prioritizing regions supported by consistent multimodal evidence rather than raw residuals alone, our method alleviates overfitting texture-induced errors and suppresses noise from pose/appearance inconsistencies. Experiments on diverse sparse-view benchmarks demonstrate state-of-the-art reconstructions, with up to +0.3 dB PSNR on DTU.
用于稀疏视角新视图合成的分层高斯泼溅中的多模态先验引导重要性采样 / Multimodal-Prior-Guided Importance Sampling for Hierarchical Gaussian Splatting in Sparse-View Novel View Synthesis
这篇论文提出了一种新方法,通过融合图像颜色、语义和几何等多种信息来智能判断3D场景中哪些细节可以准确重建,从而在仅有少数几张照片的情况下,生成更清晰、更逼真的新视角图像。
源自 arXiv: 2603.02866