基于辅助损失的解耦式分割学习 / Decoupled Split Learning via Auxiliary Loss
1️⃣ 一句话总结
这篇论文提出了一种新的分割学习方法,通过在客户端增加一个辅助分类器来提供本地训练信号,从而让客户端和服务器能够半独立地训练各自的模型部分,这种方法在保持与标准方法相当性能的同时,将通信开销减少了一半,并显著降低了内存使用。
Split learning is a distributed training paradigm where a neural network is partitioned between clients and a server, which allows data to remain at the client while only intermediate activations are shared. Traditional split learning relies on end-to-end backpropagation across the client-server split point. This incurs a large communication overhead (i.e., forward activations and backward gradients need to be exchanged every iteration) and significant memory use (for storing activations and gradients). In this paper, we develop a beyond-backpropagation training method for split learning. In this approach, the client and server train their model partitions semi-independently, using local loss signals instead of propagated gradients. In particular, the client's network is augmented with a small auxiliary classifier at the split point to provide a local error signal, while the server trains on the client's transmitted activations using the true loss function. This decoupling removes the need to send backward gradients, which cuts communication costs roughly in half and also reduces memory overhead (as each side only stores local activations for its own backward pass). We evaluate our approach on CIFAR-10 and CIFAR-100. Our experiments show two key results. First, the proposed approach achieves performance on par with standard split learning that uses backpropagation. Second, it significantly reduces communication (of transmitting activations/gradient) by 50% and peak memory usage by up to 58%.
基于辅助损失的解耦式分割学习 / Decoupled Split Learning via Auxiliary Loss
这篇论文提出了一种新的分割学习方法,通过在客户端增加一个辅助分类器来提供本地训练信号,从而让客户端和服务器能够半独立地训练各自的模型部分,这种方法在保持与标准方法相当性能的同时,将通信开销减少了一半,并显著降低了内存使用。
源自 arXiv: 2601.19261