VSDiffusion:通过可见性约束扩散驯服不适定的阴影生成问题 / VSDiffusion: Taming Ill-Posed Shadow Generation via Visibility-Constrained Diffusion
1️⃣ 一句话总结
这篇论文提出了一个名为VSDiffusion的两阶段AI框架,它通过引入可见性先验来约束生成过程,从而为合成图像中的前景物体生成几何上更准确、更逼真的投射阴影。
Generating realistic cast shadows for inserted foreground objects is a crucial yet challenging problem in image composition, where maintaining geometric consistency of shadow and object in complex scenes remains difficult due to the ill-posed nature of shadow formation. To address this issue, we propose VSDiffusion, a visibility-constrained two-stage framework designed to narrow the solution space by incorporating visibility priors. In Stage I, we predict a coarse shadow mask to localize plausible shadow generated regions. And in Stage II, conditional diffusion is performed guided by lighting and depth cues estimated from the composite to generate accurate shadows. In VSDiffusion, we inject visibility priors through two complementary pathways. First, a visibility control branch with shadow-gated cross attention that provides multi-scale structural guidance. Then, a learned soft prior map that reweights training loss in error-prone regions to enhance geometric correction. Additionally, we also introduce high-frequency guided enhancement module to sharpen boundaries and improve texture interaction with the background. Experiments on widely used public DESOBAv2 dataset demonstrated that our proposed VSDiffusion can generate accurate shadow, and establishes new SOTA results across most evaluation metrics.
VSDiffusion:通过可见性约束扩散驯服不适定的阴影生成问题 / VSDiffusion: Taming Ill-Posed Shadow Generation via Visibility-Constrained Diffusion
这篇论文提出了一个名为VSDiffusion的两阶段AI框架,它通过引入可见性先验来约束生成过程,从而为合成图像中的前景物体生成几何上更准确、更逼真的投射阴影。
源自 arXiv: 2603.08020