顺序训练的早期退出神经网络中稳定性与可塑性的平衡 / Balancing Stability and Plasticity in Sequentially Trained Early-Exiting Neural Networks
1️⃣ 一句话总结
这篇论文提出两种方法,在逐步增加早期退出分类器时,通过保护关键参数或保留原有输出分布,解决新分类器干扰旧分类器性能的问题,从而在节省计算资源的同时保持高准确性。
Early-exiting neural networks enable adaptive inference by allowing inputs to exit at intermediate classifiers, reducing computation for easy samples while maintaining high accuracy. In practice, exits can be trained sequentially by incrementally adding them to a shared backbone; however, this sequential training can cause newly introduced exits to interfere with previously learned ones, degrading the performance of earlier classifiers. We address this problem by retaining the knowledge embedded in existing exits while allowing new ones to specialize. We propose two alternative approaches that operate at different levels of the model. The first constrains learning by protecting parameters that are important for previously trained exits, while the second preserves the output distributions of earlier exits as the network adapts. These alternatives directly reflect the stability-plasticity trade-off studied in continual learning. Accordingly, we leverage \textit{Elastic Weight Consolidation} to constrain critical weights and \textit{Learning without Forgetting} to preserve output distributions. Experiments on standard benchmarks show that our approaches consistently improve early-exit performance, achieving higher accuracy over existing sequential training methods and significant performance speedups at low computational budgets.
顺序训练的早期退出神经网络中稳定性与可塑性的平衡 / Balancing Stability and Plasticity in Sequentially Trained Early-Exiting Neural Networks
这篇论文提出两种方法,在逐步增加早期退出分类器时,通过保护关键参数或保留原有输出分布,解决新分类器干扰旧分类器性能的问题,从而在节省计算资源的同时保持高准确性。
源自 arXiv: 2605.05358