Neural Galerkin Normalizing Flow for Transition Probability Density Functions of Diffusion Models
AI 摘要
提出一种新的Neural Galerkin Normalizing Flow框架,近似扩散过程的转移概率密度函数。
主要贡献
- 提出Neural Galerkin Normalizing Flow框架
- 使用Normalizing Flows求解Fokker-Planck方程
- 离线训练后可进行高效的在线评估
方法论
通过Normalizing Flows变换参考过程的转移概率密度函数,并使用Neural Galerkin方法求解参数的ODE系统。
原文摘要
We propose a new Neural Galerkin Normalizing Flow framework to approximate the transition probability density function of a diffusion process by solving the corresponding Fokker-Planck equation with an atomic initial distribution, parametrically with respect to the location of the initial mass. By using Normalizing Flows, we look for the solution as a transformation of the transition probability density function of a reference stochastic process, ensuring that our approximation is structure-preserving and automatically satisfies positivity and mass conservation constraints. By extending Neural Galerkin schemes to the context of Normalizing Flows, we derive a system of ODEs for the time evolution of the Normalizing Flow's parameters. Adaptive sampling routines are used to evaluate the Fokker-Planck residual in meaningful locations, which is of vital importance to address high-dimensional PDEs. Numerical results show that this strategy captures key features of the true solution and enforces the causal relationship between the initial datum and the density function at subsequent times. After completing an offline training phase, online evaluation becomes significantly more cost-effective than solving the PDE from scratch. The proposed method serves as a promising surrogate model, which could be deployed in many-query problems associated with stochastic differential equations, like Bayesian inference, simulation, and diffusion bridge generation.