LLM Reasoning 相关度: 6/10

Deep unfolding of MCMC kernels: scalable, modular & explainable GANs for high-dimensional posterior sampling

Jonathan Spence, Tobías I. Liaudat, Konstantinos Zygalakis, Marcelo Pereyra
arXiv: 2602.20758v1 发布: 2026-02-24 更新: 2026-02-24

AI 摘要

该论文提出了一种基于深度展开MCMC核的GAN架构,用于高效、模块化和可解释的高维后验采样。

主要贡献

  • 提出基于深度展开Langevin MCMC算法的GAN架构
  • 设计了一种监督正则化Wasserstein GAN框架用于后验采样
  • 在贝叶斯成像实验中验证了方法的有效性

方法论

通过深度展开将MCMC算法映射到模块化神经网络,利用监督正则化Wasserstein GAN框架进行端到端训练。

原文摘要

Markov chain Monte Carlo (MCMC) methods are fundamental to Bayesian computation, but can be computationally intensive, especially in high-dimensional settings. Push-forward generative models, such as generative adversarial networks (GANs), variational auto-encoders and normalising flows offer a computationally efficient alternative for posterior sampling. However, push-forward models are opaque as they lack the modularity of Bayes Theorem, leading to poor generalisation with respect to changes in the likelihood function. In this work, we introduce a novel approach to GAN architecture design by applying deep unfolding to Langevin MCMC algorithms. This paradigm maps fixed-step iterative algorithms onto modular neural networks, yielding architectures that are both flexible and amenable to interpretation. Crucially, our design allows key model parameters to be specified at inference time, offering robustness to changes in the likelihood parameters. We train these unfolded samplers end-to-end using a supervised regularized Wasserstein GAN framework for posterior sampling. Through extensive Bayesian imaging experiments, we demonstrate that our proposed approach achieves high sampling accuracy and excellent computational efficiency, while retaining the physics consistency, adaptability and interpretability of classical MCMC strategies.

标签

MCMC GAN 深度展开 贝叶斯推断 后验采样

arXiv 分类

cs.LG