菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-02-16
📄 Abstract - Activation-Space Uncertainty Quantification for Pretrained Networks

Reliable uncertainty estimates are crucial for deploying pretrained models; yet, many strong methods for quantifying uncertainty require retraining, Monte Carlo sampling, or expensive second-order computations and may alter a frozen backbone's predictions. To address this, we introduce Gaussian Process Activations (GAPA), a post-hoc method that shifts Bayesian modeling from weights to activations. GAPA replaces standard nonlinearities with Gaussian-process activations whose posterior mean exactly matches the original activation, preserving the backbone's point predictions by construction while providing closed-form epistemic variances in activation space. To scale to modern architectures, we use a sparse variational inducing-point approximation over cached training activations, combined with local k-nearest-neighbor subset conditioning, enabling deterministic single-pass uncertainty propagation without sampling, backpropagation, or second-order information. Across regression, classification, image segmentation, and language modeling, GAPA matches or outperforms strong post-hoc baselines in calibration and out-of-distribution detection while remaining efficient at test time.

顶级标签: model evaluation machine learning theory
详细标签: uncertainty quantification bayesian methods gaussian processes post-hoc analysis activation space 或 搜索:

预训练网络激活空间的不确定性量化 / Activation-Space Uncertainty Quantification for Pretrained Networks


1️⃣ 一句话总结

这篇论文提出了一种名为GAPA的新方法,它能在不改变预训练模型预测结果的前提下,高效地为模型输出提供可靠的不确定性估计,适用于多种任务且计算成本低。

源自 arXiv: 2602.14934