LLM Reasoning 相关度: 7/10

Use What You Know: Causal Foundation Models with Partial Graphs

Arik Reuter, Anish Dhir, Cristiana Diaconu, Jake Robertson, Ole Ossen, Frank Hutter, Adrian Weller, Mark van der Wilk, Bernhard Schölkopf
arXiv: 2602.14972v1 发布: 2026-02-16 更新: 2026-02-16

AI 摘要

论文提出了一种将因果信息融入因果基础模型(CFMs)的方法,提升模型性能。

主要贡献

  • 提出在CFMs中融入因果信息的框架
  • 提出利用完整或部分因果图信息的策略
  • 实验证明了该方法可以使通用CFM达到特定模型的性能

方法论

通过将可学习的偏差注入到注意力机制中,有效地利用了完整和部分的因果信息。

原文摘要

Estimating causal quantities traditionally relies on bespoke estimators tailored to specific assumptions. Recently proposed Causal Foundation Models (CFMs) promise a more unified approach by amortising causal discovery and inference in a single step. However, in their current state, they do not allow for the incorporation of any domain knowledge, which can lead to suboptimal predictions. We bridge this gap by introducing methods to condition CFMs on causal information, such as the causal graph or more readily available ancestral information. When access to complete causal graph information is too strict a requirement, our approach also effectively leverages partial causal information. We systematically evaluate conditioning strategies and find that injecting learnable biases into the attention mechanism is the most effective method to utilise full and partial causal information. Our experiments show that this conditioning allows a general-purpose CFM to match the performance of specialised models trained on specific causal structures. Overall, our approach addresses a central hurdle on the path towards all-in-one causal foundation models: the capability to answer causal queries in a data-driven manner while effectively leveraging any amount of domain expertise.

标签

因果推断 因果基础模型 领域知识 注意力机制

arXiv 分类

cs.LG