📄
Abstract - On the Equivalence of Random Network Distillation, Deep Ensembles, and Bayesian Inference
Uncertainty quantification is central to safe and efficient deployments of deep learning models, yet many computationally practical methods lack lacking rigorous theoretical motivation. Random network distillation (RND) is a lightweight technique that measures novelty via prediction errors against a fixed random target. While empirically effective, it has remained unclear what uncertainties RND measures and how its estimates relate to other approaches, e.g. Bayesian inference or deep ensembles. This paper establishes these missing theoretical connections by analyzing RND within the neural tangent kernel framework in the limit of infinite network width. Our analysis reveals two central findings in this limit: (1) The uncertainty signal from RND -- its squared self-predictive error -- is equivalent to the predictive variance of a deep ensemble. (2) By constructing a specific RND target function, we show that the RND error distribution can be made to mirror the centered posterior predictive distribution of Bayesian inference with wide neural networks. Based on this equivalence, we moreover devise a posterior sampling algorithm that generates i.i.d. samples from an exact Bayesian posterior predictive distribution using this modified \textit{Bayesian RND} model. Collectively, our findings provide a unified theoretical perspective that places RND within the principled frameworks of deep ensembles and Bayesian inference, and offer new avenues for efficient yet theoretically grounded uncertainty quantification methods.
论随机网络蒸馏、深度集成与贝叶斯推断的等价性 /
On the Equivalence of Random Network Distillation, Deep Ensembles, and Bayesian Inference
1️⃣ 一句话总结
这篇论文在无限宽神经网络的极限下,从理论上证明了随机网络蒸馏(RND)这种轻量级不确定性估计方法,其核心信号等价于深度集成的预测方差,并且通过特定设计可以使其误差分布与贝叶斯推断的后验预测分布一致,从而为RND提供了坚实的理论依据并开辟了高效不确定性量化的新途径。