在序列蒙特卡洛框架中研究神经网络的批次推断 / Investigating Batch Inference in a Sequential Monte Carlo Framework for Neural Networks
1️⃣ 一句话总结
这篇论文提出了一种改进的序列蒙特卡洛采样方法,通过逐步引入小批量数据来近似贝叶斯神经网络的后验分布,在图像分类基准测试中实现了高达6倍的训练加速,且精度损失极小。
Bayesian inference allows us to define a posterior distribution over the weights of a generic neural network (NN). Exact posteriors are usually intractable, in which case approximations can be employed. One such approximation - variational inference - is computationally efficient when using mini-batch stochastic gradient descent as subsets of the data are used for likelihood and gradient evaluations, though the approach relies on the selection of a variational distribution which sufficiently matches the form of the posterior. Particle-based methods such as Markov chain Monte Carlo and Sequential Monte Carlo (SMC) do not assume a parametric family for the posterior by typically require higher computational cost. These sampling methods typically use the full-batch of data for likelihood and gradient evaluations, which contributes to this computational expense. We explore several methods of gradually introducing more mini-batches of data (data annealing) into likelihood and gradient evaluations of an SMC sampler. We find that we can achieve up to $6\times$ faster training with minimal loss in accuracy on benchmark image classification problems using NNs.
在序列蒙特卡洛框架中研究神经网络的批次推断 / Investigating Batch Inference in a Sequential Monte Carlo Framework for Neural Networks
这篇论文提出了一种改进的序列蒙特卡洛采样方法,通过逐步引入小批量数据来近似贝叶斯神经网络的后验分布,在图像分类基准测试中实现了高达6倍的训练加速,且精度损失极小。
源自 arXiv: 2601.21983