LLM Reasoning 相关度: 6/10

Latent Matters: Learning Deep State-Space Models

Alexej Klushyn, Richard Kurle, Maximilian Soelch, Botond Cseke, Patrick van der Smagt
arXiv: 2602.23050v1 发布: 2026-02-26 更新: 2026-02-26

AI 摘要

论文提出一种约束优化框架训练深度状态空间模型,并提出EKVAE模型,在系统辨识和预测方面表现优异。

主要贡献

  • 提出一种约束优化框架训练DSSM
  • 提出 extended Kalman VAE (EKVAE) 模型
  • EKVAE在系统辨识和预测精度上优于现有模型

方法论

通过约束优化框架训练DSSM,结合变分推断和贝叶斯滤波/平滑方法,更准确地建模动态系统。

原文摘要

Deep state-space models (DSSMs) enable temporal predictions by learning the underlying dynamics of observed sequence data. They are often trained by maximising the evidence lower bound. However, as we show, this does not ensure the model actually learns the underlying dynamics. We therefore propose a constrained optimisation framework as a general approach for training DSSMs. Building upon this, we introduce the extended Kalman VAE (EKVAE), which combines amortised variational inference with classic Bayesian filtering/smoothing to model dynamics more accurately than RNN-based DSSMs. Our results show that the constrained optimisation framework significantly improves system identification and prediction accuracy on the example of established state-of-the-art DSSMs. The EKVAE outperforms previous models w.r.t. prediction accuracy, achieves remarkable results in identifying dynamical systems, and can furthermore successfully learn state-space representations where static and dynamic features are disentangled.

标签

深度状态空间模型 变分推断 贝叶斯滤波 系统辨识

arXiv 分类

cs.LG