LLM Reasoning 相关度: 7/10

Frequentist Consistency of Prior-Data Fitted Networks for Causal Inference

Valentyn Melnychuk, Vahid Balazadeh, Stefan Feuerriegel, Rahul G. Krishnan
arXiv: 2603.12037v1 发布: 2026-03-12 更新: 2026-03-12

AI 摘要

论文分析了基于PFN的因果推断方法的一致性问题,并提出了校准方法。

主要贡献

  • 指出现有PFN方法存在先验诱导的混淆偏差。
  • 提出基于单步后验校正(OSPC)的校准程序。
  • 通过tailoring martingale posteriors实现了OSPC,并验证了其有效性。

方法论

通过理论分析,证明了OSPC可以恢复频率一致性,并通过半合成实验验证了其性能。

原文摘要

Foundation models based on prior-data fitted networks (PFNs) have shown strong empirical performance in causal inference by framing the task as an in-context learning problem.However, it is unclear whether PFN-based causal estimators provide uncertainty quantification that is consistent with classical frequentist estimators. In this work, we address this gap by analyzing the frequentist consistency of PFN-based estimators for the average treatment effect (ATE). (1) We show that existing PFNs, when interpreted as Bayesian ATE estimators, can exhibit prior-induced confounding bias: the prior is not asymptotically overwritten by data, which, in turn, prevents frequentist consistency. (2) As a remedy, we suggest employing a calibration procedure based on a one-step posterior correction (OSPC). We show that the OSPC helps to restore frequentist consistency and can yield a semi-parametric Bernstein-von Mises theorem for calibrated PFNs (i.e., both the calibrated PFN-based estimators and the classical semi-parametric efficient estimators converge in distribution with growing data size). (3) Finally, we implement OSPC through tailoring martingale posteriors on top of the PFNs. In this way, we are able to recover functional nuisance posteriors from PFNs, required by the OSPC. In multiple (semi-)synthetic experiments, PFNs calibrated with our martingale posterior OSPC produce ATE uncertainty that (i) asymptotically matches frequentist uncertainty and (ii) is well calibrated in finite samples in comparison to other Bayesian ATE estimators.

标签

因果推断 Foundation Model 频率一致性 贝叶斯方法

arXiv 分类

cs.LG