大多数卷积神经网络都受到微小对抗性扰动的影响 / Most Convolutional Networks Suffer from Small Adversarial Perturbations
1️⃣ 一句话总结
这篇论文证明,即使是微小的、几乎无法察觉的输入扰动,也足以欺骗随机构建的卷积神经网络,并且这种‘对抗性攻击’可以通过简单的梯度下降一步实现。
The existence of adversarial examples is relatively understood for random fully connected neural networks, but much less so for convolutional neural networks (CNNs). The recent work [Daniely, 2025] establishes that adversarial examples can be found in CNNs, in some non-optimal distance from the input. We extend over this work and prove that adversarial examples in random CNNs with input dimension $d$ can be found already in $\ell_2$-distance of order $\lVert x \rVert /\sqrt{d}$ from the input $x$, which is essentially the nearest possible. We also show that such adversarial small perturbations can be found using a single step of gradient descent. To derive our results we use Fourier decomposition to efficiently bound the singular values of a random linear convolutional operator, which is the main ingredient of a CNN layer. This bound might be of independent interest.
大多数卷积神经网络都受到微小对抗性扰动的影响 / Most Convolutional Networks Suffer from Small Adversarial Perturbations
这篇论文证明,即使是微小的、几乎无法察觉的输入扰动,也足以欺骗随机构建的卷积神经网络,并且这种‘对抗性攻击’可以通过简单的梯度下降一步实现。
源自 arXiv: 2602.03415