预训练网络激活空间的不确定性量化 / Activation-Space Uncertainty Quantification for Pretrained Networks
1️⃣ 一句话总结
这篇论文提出了一种名为GAPA的新方法,它能在不改变预训练模型预测结果的前提下,高效地为模型输出提供可靠的不确定性估计,适用于多种任务且计算成本低。
Reliable uncertainty estimates are crucial for deploying pretrained models; yet, many strong methods for quantifying uncertainty require retraining, Monte Carlo sampling, or expensive second-order computations and may alter a frozen backbone's predictions. To address this, we introduce Gaussian Process Activations (GAPA), a post-hoc method that shifts Bayesian modeling from weights to activations. GAPA replaces standard nonlinearities with Gaussian-process activations whose posterior mean exactly matches the original activation, preserving the backbone's point predictions by construction while providing closed-form epistemic variances in activation space. To scale to modern architectures, we use a sparse variational inducing-point approximation over cached training activations, combined with local k-nearest-neighbor subset conditioning, enabling deterministic single-pass uncertainty propagation without sampling, backpropagation, or second-order information. Across regression, classification, image segmentation, and language modeling, GAPA matches or outperforms strong post-hoc baselines in calibration and out-of-distribution detection while remaining efficient at test time.
预训练网络激活空间的不确定性量化 / Activation-Space Uncertainty Quantification for Pretrained Networks
这篇论文提出了一种名为GAPA的新方法,它能在不改变预训练模型预测结果的前提下,高效地为模型输出提供可靠的不确定性估计,适用于多种任务且计算成本低。
源自 arXiv: 2602.14934