Smoothing the Landscape: Causal Structure Learning via Diffusion Denoising Objectives
AI 摘要
提出DDCD框架,利用扩散模型的目标函数学习因果结构,解决高维数据下的可扩展性和稳定性问题。
主要贡献
- 利用扩散模型的去噪目标函数平滑梯度,加速收敛
- 提出自适应的k-hop无环约束,提高运行效率
- 在合成数据和真实数据上验证了DDCD的有效性
方法论
使用扩散模型的逆向去噪过程推断参数化的因果结构,并结合自适应k-hop无环约束。
原文摘要
Understanding causal dependencies in observational data is critical for informing decision-making. These relationships are often modeled as Bayesian Networks (BNs) and Directed Acyclic Graphs (DAGs). Existing methods, such as NOTEARS and DAG-GNN, often face issues with scalability and stability in high-dimensional data, especially when there is a feature-sample imbalance. Here, we show that the denoising score matching objective of diffusion models could smooth the gradients for faster, more stable convergence. We also propose an adaptive k-hop acyclicity constraint that improves runtime over existing solutions that require matrix inversion. We name this framework Denoising Diffusion Causal Discovery (DDCD). Unlike generative diffusion models, DDCD utilizes the reverse denoising process to infer a parameterized causal structure rather than to generate data. We demonstrate the competitive performance of DDCDs on synthetic benchmarking data. We also show that our methods are practically useful by conducting qualitative analyses on two real-world examples. Code is available at this url: https://github.com/haozhu233/ddcd.