菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-02-16
📄 Abstract - Unbiased Approximate Vector-Jacobian Products for Efficient Backpropagation

In this work we introduce methods to reduce the computational and memory costs of training deep neural networks. Our approach consists in replacing exact vector-jacobian products by randomized, unbiased approximations thereof during backpropagation. We provide a theoretical analysis of the trade-off between the number of epochs needed to achieve a target precision and the cost reduction for each epoch. We then identify specific unbiased estimates of vector-jacobian products for which we establish desirable optimality properties of minimal variance under sparsity constraints. Finally we provide in-depth experiments on multi-layer perceptrons, BagNets and Visual Transfomers architectures. These validate our theoretical results, and confirm the potential of our proposed unbiased randomized backpropagation approach for reducing the cost of deep learning.

顶级标签: model training machine learning theory
详细标签: backpropagation computational efficiency variance reduction neural network training randomized optimization 或 搜索:

用于高效反向传播的无偏近似向量-雅可比积方法 / Unbiased Approximate Vector-Jacobian Products for Efficient Backpropagation


1️⃣ 一句话总结

这篇论文提出了一种通过随机无偏近似来替代精确计算,从而显著降低深度神经网络训练的计算和内存成本的新方法。

源自 arXiv: 2602.14701