计算特征方向以实现旋转无关的图像分析 / Computing a Characteristic Orientation for Rotation-Independent Image Analysis
1️⃣ 一句话总结
这篇论文提出了一种名为‘通用强度方向’的预处理方法,它能为每张图像计算一个全局方向并将其对齐到标准参考系,从而让普通的深度学习模型无需改动结构就能更好地处理旋转后的图像,在多个数据集上取得了比专用旋转不变模型更高的准确率。
Handling geometric transformations, particularly rotations, remains a challenge in deep learning for computer vision. Standard neural networks lack inherent rotation invariance and typically rely on data augmentation or architectural modifications to improve robustness. Although effective, these approaches increase computational demands, require specialised implementations, or alter network structures, limiting their applicability. This paper introduces General Intensity Direction (GID), a preprocessing method that improves rotation robustness without modifying the network architecture. The method estimates a global orientation for each image and aligns it to a canonical reference frame, allowing standard models to process inputs more consistently across different rotations. Unlike moment-based approaches that extract invariant descriptors, this method directly transforms the image while preserving spatial structure, making it compatible with convolutional networks. Experimental evaluation on the rotated MNIST dataset shows that the proposed method achieves higher accuracy than state-of-the-art rotation-invariant architectures. Additional experiments on the CIFAR-10 dataset, confirm that the method remains effective under more complex conditions.
计算特征方向以实现旋转无关的图像分析 / Computing a Characteristic Orientation for Rotation-Independent Image Analysis
这篇论文提出了一种名为‘通用强度方向’的预处理方法,它能为每张图像计算一个全局方向并将其对齐到标准参考系,从而让普通的深度学习模型无需改动结构就能更好地处理旋转后的图像,在多个数据集上取得了比专用旋转不变模型更高的准确率。
源自 arXiv: 2602.20930