Agent Tuning & Optimization 相关度: 5/10

Unbiased Approximate Vector-Jacobian Products for Efficient Backpropagation

Killian Bakong, Laurent Massoulié, Edouard Oyallon, Kevin Scaman
arXiv: 2602.14701v1 发布: 2026-02-16 更新: 2026-02-16

AI 摘要

提出一种基于随机无偏近似向量-雅可比积的反向传播方法,以降低深度学习的计算和内存成本。

主要贡献

  • 提出随机无偏近似向量-雅可比积的反向传播方法
  • 分析了精度与成本之间的权衡
  • 在多种架构上验证了该方法的有效性

方法论

通过随机无偏近似向量-雅可比积替换精确计算,理论分析精度-成本权衡,并在MLP、BagNets和Visual Transformers上进行实验验证。

原文摘要

In this work we introduce methods to reduce the computational and memory costs of training deep neural networks. Our approach consists in replacing exact vector-jacobian products by randomized, unbiased approximations thereof during backpropagation. We provide a theoretical analysis of the trade-off between the number of epochs needed to achieve a target precision and the cost reduction for each epoch. We then identify specific unbiased estimates of vector-jacobian products for which we establish desirable optimality properties of minimal variance under sparsity constraints. Finally we provide in-depth experiments on multi-layer perceptrons, BagNets and Visual Transfomers architectures. These validate our theoretical results, and confirm the potential of our proposed unbiased randomized backpropagation approach for reducing the cost of deep learning.

标签

反向传播 向量-雅可比积 深度学习 优化 无偏估计

arXiv 分类

cs.LG stat.ML