通过SVDA在图像分类中实现可解释的视觉Transformer / Interpretable Vision Transformers in Image Classification via SVDA
1️⃣ 一句话总结
这篇论文提出了一种名为SVDA的新型注意力机制,它能让视觉Transformer在保持图像分类高精度的同时,生成更清晰、更易于理解的注意力模式,从而提升模型的可解释性。
Vision Transformers (ViTs) have achieved state-of-the-art performance in image classification, yet their attention mechanisms often remain opaque and exhibit dense, non-structured behaviors. In this work, we adapt our previously proposed SVD-Inspired Attention (SVDA) mechanism to the ViT architecture, introducing a geometrically grounded formulation that enhances interpretability, sparsity, and spectral structure. We apply the use of interpretability indicators -- originally proposed with SVDA -- to monitor attention dynamics during training and assess structural properties of the learned representations. Experimental evaluations on four widely used benchmarks -- CIFAR-10, FashionMNIST, CIFAR-100, and ImageNet-100 -- demonstrate that SVDA consistently yields more interpretable attention patterns without sacrificing classification accuracy. While the current framework offers descriptive insights rather than prescriptive guidance, our results establish SVDA as a comprehensive and informative tool for analyzing and developing structured attention models in computer vision. This work lays the foundation for future advances in explainable AI, spectral diagnostics, and attention-based model compression.
通过SVDA在图像分类中实现可解释的视觉Transformer / Interpretable Vision Transformers in Image Classification via SVDA
这篇论文提出了一种名为SVDA的新型注意力机制,它能让视觉Transformer在保持图像分类高精度的同时,生成更清晰、更易于理解的注意力模式,从而提升模型的可解释性。
源自 arXiv: 2602.10994