Agent Tuning & Optimization 相关度: 5/10

Tuning the burn-in phase in training recurrent neural networks improves their performance

Julian D. Schiller, Malte Heinrich, Victor G. Lopez, Matthias A. Müller
arXiv: 2602.10911v1 发布: 2026-02-11 更新: 2026-02-11

AI 摘要

研究了RNN训练中burn-in阶段对性能的影响,并通过实验验证其重要性。

主要贡献

  • 理论分析了截断BPTT的误差界限
  • 强调了RNN训练中burn-in阶段的重要性
  • 实验证明适当调整burn-in阶段可显著提高性能

方法论

通过理论分析建立了截断BPTT的误差界限,并通过系统辨识和时间序列预测任务的实验验证。

原文摘要

Training recurrent neural networks (RNNs) with standard backpropagation through time (BPTT) can be challenging, especially in the presence of long input sequences. A practical alternative to reduce computational and memory overhead is to perform BPTT repeatedly over shorter segments of the training data set, corresponding to truncated BPTT. In this paper, we examine the training of RNNs when using such a truncated learning approach for time series tasks. Specifically, we establish theoretical bounds on the accuracy and performance loss when optimizing over subsequences instead of the full data sequence. This reveals that the burn-in phase of the RNN is an important tuning knob in its training, with significant impact on the performance guarantees. We validate our theoretical results through experiments on standard benchmarks from the fields of system identification and time series forecasting. In all experiments, we observe a strong influence of the burn-in phase on the training process, and proper tuning can lead to a reduction of the prediction error on the training and test data of more than 60% in some cases.

标签

RNN BPTT Truncated BPTT Burn-in Time Series

arXiv 分类

cs.LG eess.SY math.OC