菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-02-24
📄 Abstract - Deep unfolding of MCMC kernels: scalable, modular & explainable GANs for high-dimensional posterior sampling

Markov chain Monte Carlo (MCMC) methods are fundamental to Bayesian computation, but can be computationally intensive, especially in high-dimensional settings. Push-forward generative models, such as generative adversarial networks (GANs), variational auto-encoders and normalising flows offer a computationally efficient alternative for posterior sampling. However, push-forward models are opaque as they lack the modularity of Bayes Theorem, leading to poor generalisation with respect to changes in the likelihood function. In this work, we introduce a novel approach to GAN architecture design by applying deep unfolding to Langevin MCMC algorithms. This paradigm maps fixed-step iterative algorithms onto modular neural networks, yielding architectures that are both flexible and amenable to interpretation. Crucially, our design allows key model parameters to be specified at inference time, offering robustness to changes in the likelihood parameters. We train these unfolded samplers end-to-end using a supervised regularized Wasserstein GAN framework for posterior sampling. Through extensive Bayesian imaging experiments, we demonstrate that our proposed approach achieves high sampling accuracy and excellent computational efficiency, while retaining the physics consistency, adaptability and interpretability of classical MCMC strategies.

顶级标签: machine learning model training theory
详细标签: bayesian computation generative adversarial networks posterior sampling deep unfolding mcmc 或 搜索:

MCMC核的深度展开:用于高维后验采样的可扩展、模块化且可解释的生成对抗网络 / Deep unfolding of MCMC kernels: scalable, modular & explainable GANs for high-dimensional posterior sampling


1️⃣ 一句话总结

这篇论文提出了一种新方法,通过将传统的马尔可夫链蒙特卡洛采样算法‘展开’成模块化的神经网络结构,构建出既高效又易于理解的生成对抗网络,从而在贝叶斯计算中实现快速、准确且能适应不同参数的后验分布采样。

源自 arXiv: 2602.20758