LLM Reasoning 相关度: 6/10

Amortising Inference and Meta-Learning Priors in Neural Networks

Tommy Rochussen, Vincent Fortuin
arXiv: 2602.08782v1 发布: 2026-02-09 更新: 2026-02-09

AI 摘要

该论文提出了一种学习神经网络权重先验的方法,结合了贝叶斯深度学习和概率元学习。

主要贡献

  • 提出了一种学习权重先验的方法
  • 实现了数据集级别的摊销变分推断
  • 利用贝叶斯神经网络作为灵活的生成模型

方法论

通过引入数据集级别的摊销变分推断,学习BNN权重的先验,并将模型视为神经过程。

原文摘要

One of the core facets of Bayesianism is in the updating of prior beliefs in light of new evidence$\text{ -- }$so how can we maintain a Bayesian approach if we have no prior beliefs in the first place? This is one of the central challenges in the field of Bayesian deep learning, where it is not clear how to represent beliefs about a prediction task by prior distributions over model parameters. Bridging the fields of Bayesian deep learning and probabilistic meta-learning, we introduce a way to $\textit{learn}$ a weights prior from a collection of datasets by introducing a way to perform per-dataset amortised variational inference. The model we develop can be viewed as a neural process whose latent variable is the set of weights of a BNN and whose decoder is the neural network parameterised by a sample of the latent variable itself. This unique model allows us to study the behaviour of Bayesian neural networks under well-specified priors, use Bayesian neural networks as flexible generative models, and perform desirable but previously elusive feats in neural processes such as within-task minibatching or meta-learning under extreme data-starvation.

标签

贝叶斯深度学习 元学习 变分推断 神经网络

arXiv 分类

stat.ML cs.LG