菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-03-11
📄 Abstract - Client-Conditional Federated Learning via Local Training Data Statistics

Federated learning (FL) under data heterogeneity remains challenging: existing methods either ignore client differences (FedAvg), require costly cluster discovery (IFCA), or maintain per-client models (Ditto). All degrade when data is sparse or heterogeneity is multi-dimensional. We propose conditioning a single global model on locally-computed PCA statistics of each client's training data, requiring zero additional communication. Evaluating across 97~configurations spanning four heterogeneity types (label shift, covariate shift, concept shift, and combined heterogeneity), four datasets (MNIST, Fashion-MNIST, CIFAR-10, CIFAR-100), and seven FL baseline methods, we find that our method matches the Oracle baseline -- which knows true cluster assignments -- across all settings, surpasses it by 1--6% on combined heterogeneity where continuous statistics are richer than discrete cluster identifiers, and is uniquely sparsity-robust among all tested methods.

顶级标签: machine learning systems model training
详细标签: federated learning data heterogeneity client conditioning pca statistics sparsity robustness 或 搜索:

基于本地训练数据统计的客户端条件化联邦学习 / Client-Conditional Federated Learning via Local Training Data Statistics


1️⃣ 一句话总结

这篇论文提出了一种新的联邦学习方法,它通过分析每个客户端本地数据的统计特征来个性化调整一个全局模型,无需额外通信,就能在各种复杂的数据差异场景下稳定地达到甚至超过已知最优方法的性能。

源自 arXiv: 2603.11307