基于图像锐化的高效先发制人鲁棒化方法 / Efficient Preemptive Robustification with Image Sharpening
1️⃣ 一句话总结
这篇论文提出了一种简单高效的图像锐化方法,能在攻击发生前主动提升图像对微小扰动的抵抗能力,无需依赖复杂的模型或计算,就能显著增强深度学习模型的安全性。
Despite their great success, deep neural networks rely on high-dimensional, non-robust representations, making them vulnerable to imperceptible perturbations, even in transfer scenarios. To address this, both training-time defenses (e.g., adversarial training and robust architecture design) and post-attack defenses (e.g., input purification and adversarial detection) have been extensively studied. Recently, a limited body of work has preliminarily explored a pre-attack defense paradigm, termed preemptive robustification, which introduces subtle modifications to benign samples prior to attack to proactively resist adversarial perturbations. Unfortunately, their practical applicability remains questionable due to several limitations, including (1) reliance on well-trained classifiers as surrogates to provide robustness priors, (2) substantial computational overhead arising from iterative optimization or trained generators for robustification, and (3) limited interpretability of the optimization- or generation-based robustification processes. Inspired by recent studies revealing a positive correlation between texture intensity and the robustness of benign samples, we show that image sharpening alone can efficiently robustify images. To the best of our knowledge, this is the first surrogate-free, optimization-free, generator-free, and human-interpretable robustification approach. Extensive experiments demonstrate that sharpening yields remarkable robustness gains with low computational cost, especially in transfer scenarios.
基于图像锐化的高效先发制人鲁棒化方法 / Efficient Preemptive Robustification with Image Sharpening
这篇论文提出了一种简单高效的图像锐化方法,能在攻击发生前主动提升图像对微小扰动的抵抗能力,无需依赖复杂的模型或计算,就能显著增强深度学习模型的安全性。
源自 arXiv: 2603.25244