Unbiased Approximate Vector-Jacobian Products for Efficient Backpropagation
AI 摘要
提出一种基于随机无偏近似向量-雅可比积的反向传播方法,以降低深度学习的计算和内存成本。
主要贡献
- 提出随机无偏近似向量-雅可比积的反向传播方法
- 分析了精度与成本之间的权衡
- 在多种架构上验证了该方法的有效性
方法论
通过随机无偏近似向量-雅可比积替换精确计算,理论分析精度-成本权衡,并在MLP、BagNets和Visual Transformers上进行实验验证。
原文摘要
In this work we introduce methods to reduce the computational and memory costs of training deep neural networks. Our approach consists in replacing exact vector-jacobian products by randomized, unbiased approximations thereof during backpropagation. We provide a theoretical analysis of the trade-off between the number of epochs needed to achieve a target precision and the cost reduction for each epoch. We then identify specific unbiased estimates of vector-jacobian products for which we establish desirable optimality properties of minimal variance under sparsity constraints. Finally we provide in-depth experiments on multi-layer perceptrons, BagNets and Visual Transfomers architectures. These validate our theoretical results, and confirm the potential of our proposed unbiased randomized backpropagation approach for reducing the cost of deep learning.