菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-01-29
📄 Abstract - Investigating Batch Inference in a Sequential Monte Carlo Framework for Neural Networks

Bayesian inference allows us to define a posterior distribution over the weights of a generic neural network (NN). Exact posteriors are usually intractable, in which case approximations can be employed. One such approximation - variational inference - is computationally efficient when using mini-batch stochastic gradient descent as subsets of the data are used for likelihood and gradient evaluations, though the approach relies on the selection of a variational distribution which sufficiently matches the form of the posterior. Particle-based methods such as Markov chain Monte Carlo and Sequential Monte Carlo (SMC) do not assume a parametric family for the posterior by typically require higher computational cost. These sampling methods typically use the full-batch of data for likelihood and gradient evaluations, which contributes to this computational expense. We explore several methods of gradually introducing more mini-batches of data (data annealing) into likelihood and gradient evaluations of an SMC sampler. We find that we can achieve up to $6\times$ faster training with minimal loss in accuracy on benchmark image classification problems using NNs.

顶级标签: machine learning model training theory
详细标签: bayesian inference sequential monte carlo neural networks data annealing particle methods 或 搜索:

在序列蒙特卡洛框架中研究神经网络的批次推断 / Investigating Batch Inference in a Sequential Monte Carlo Framework for Neural Networks


1️⃣ 一句话总结

这篇论文提出了一种改进的序列蒙特卡洛采样方法,通过逐步引入小批量数据来近似贝叶斯神经网络的后验分布,在图像分类基准测试中实现了高达6倍的训练加速,且精度损失极小。

源自 arXiv: 2601.21983