菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-02-03
📄 Abstract - Most Convolutional Networks Suffer from Small Adversarial Perturbations

The existence of adversarial examples is relatively understood for random fully connected neural networks, but much less so for convolutional neural networks (CNNs). The recent work [Daniely, 2025] establishes that adversarial examples can be found in CNNs, in some non-optimal distance from the input. We extend over this work and prove that adversarial examples in random CNNs with input dimension $d$ can be found already in $\ell_2$-distance of order $\lVert x \rVert /\sqrt{d}$ from the input $x$, which is essentially the nearest possible. We also show that such adversarial small perturbations can be found using a single step of gradient descent. To derive our results we use Fourier decomposition to efficiently bound the singular values of a random linear convolutional operator, which is the main ingredient of a CNN layer. This bound might be of independent interest.

顶级标签: theory machine learning model evaluation
详细标签: adversarial examples convolutional networks robustness random networks fourier analysis 或 搜索:

大多数卷积神经网络都受到微小对抗性扰动的影响 / Most Convolutional Networks Suffer from Small Adversarial Perturbations


1️⃣ 一句话总结

这篇论文证明,即使是微小的、几乎无法察觉的输入扰动,也足以欺骗随机构建的卷积神经网络,并且这种‘对抗性攻击’可以通过简单的梯度下降一步实现。

源自 arXiv: 2602.03415