菜单

关于 🐙 GitHub
arXiv 提交日期: 2025-12-21
📄 Abstract - CosineGate: Semantic Dynamic Routing via Cosine Incompatibility in Residual Networks

Modern deep residual networks perform substantial redundant computation by evaluating all residual blocks for every input, even when identity mappings suffice. We introduce CosineGate, an end-to-end differentiable architecture for dynamic routing in residual networks that uses cosine incompatibility between identity and residual feature representations as a self-supervised skip signal. CosineGate measures semantic redundancy through the Cosine Incompatibility Ratio (CIR), defined as 1 - cos(x, F(x)), and uses Gumbel-Softmax relaxation to enable per-sample, per-block gating during training. A progressive FLOPs regularization term controls average compute usage without destabilizing optimization. On CIFAR-10, CosineGate spans the accuracy-efficiency Pareto frontier: an aggressive configuration achieves 89.9 percent accuracy with 24.1 percent FLOPs savings, a balanced configuration achieves 91.3 percent accuracy with 28.5 percent savings at epoch 160, and a conservative configuration reaches a peak of 93.2 percent accuracy with minimal compute reduction. These results match or exceed ResNet-20 (91.3 percent) while reducing computation, without auxiliary supervision, distillation, or task-specific heuristics. Our results demonstrate that simple geometric measures of feature incompatibility provide a principled and effective signal for dynamic residual routing.

顶级标签: model training systems machine learning
详细标签: dynamic routing residual networks computational efficiency gating mechanism self-supervised learning 或 搜索:

CosineGate:基于余弦不兼容性的残差网络语义动态路由方法 / CosineGate: Semantic Dynamic Routing via Cosine Incompatibility in Residual Networks


1️⃣ 一句话总结

这篇论文提出了一种名为CosineGate的智能方法,它能让深度神经网络根据输入图片的难易程度,自动跳过一些不必要的计算步骤,从而在保持甚至提高识别准确率的同时,显著减少计算量,实现更高效的推理。

源自 arXiv: 2512.22206