Temporal Credit Is Free
arXiv: 2603.28750v1
发布: 2026-03-30
更新: 2026-03-30
AI 摘要
循环神经网络无需完整雅可比传播即可在线学习,仅用即时导数和梯度归一化即可。
主要贡献
- 提出了一种新的训练循环神经网络的方法,无需完整RTRL
- 提出一种架构规则预测何时需要梯度归一化
- 实验证明该方法在多个任务上性能优异且更节省内存
方法论
利用即时导数和梯度归一化替代完整雅可比传播,加速循环神经网络的在线学习过程。
原文摘要
Recurrent networks do not need Jacobian propagation to adapt online. The hidden state already carries temporal credit through the forward pass; immediate derivatives suffice if you stop corrupting them with stale trace memory and normalize gradient scales across parameter groups. An architectural rule predicts when normalization is needed: \b{eta}2 is required when gradients must pass through a nonlinear state update with no output bypass, and unnecessary otherwise. Across ten architectures, real primate neural data, and streaming ML benchmarks, immediate derivatives with RMSprop match or exceed full RTRL, scaling to n = 1024 at 1000x less memory.