基于本地训练数据统计的客户端条件化联邦学习 / Client-Conditional Federated Learning via Local Training Data Statistics
1️⃣ 一句话总结
这篇论文提出了一种新的联邦学习方法,它通过分析每个客户端本地数据的统计特征来个性化调整一个全局模型,无需额外通信,就能在各种复杂的数据差异场景下稳定地达到甚至超过已知最优方法的性能。
Federated learning (FL) under data heterogeneity remains challenging: existing methods either ignore client differences (FedAvg), require costly cluster discovery (IFCA), or maintain per-client models (Ditto). All degrade when data is sparse or heterogeneity is multi-dimensional. We propose conditioning a single global model on locally-computed PCA statistics of each client's training data, requiring zero additional communication. Evaluating across 97~configurations spanning four heterogeneity types (label shift, covariate shift, concept shift, and combined heterogeneity), four datasets (MNIST, Fashion-MNIST, CIFAR-10, CIFAR-100), and seven FL baseline methods, we find that our method matches the Oracle baseline -- which knows true cluster assignments -- across all settings, surpasses it by 1--6% on combined heterogeneity where continuous statistics are richer than discrete cluster identifiers, and is uniquely sparsity-robust among all tested methods.
基于本地训练数据统计的客户端条件化联邦学习 / Client-Conditional Federated Learning via Local Training Data Statistics
这篇论文提出了一种新的联邦学习方法,它通过分析每个客户端本地数据的统计特征来个性化调整一个全局模型,无需额外通信,就能在各种复杂的数据差异场景下稳定地达到甚至超过已知最优方法的性能。
源自 arXiv: 2603.11307