菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-02-23
📄 Abstract - DP-FedAdamW: An Efficient Optimizer for Differentially Private Federated Large Models

Balancing convergence efficiency and robustness under Differential Privacy (DP) is a central challenge in Federated Learning (FL). While AdamW accelerates training and fine-tuning in large-scale models, we find that directly applying it to Differentially Private FL (DPFL) suffers from three major issues: (i) data heterogeneity and privacy noise jointly amplify the variance of second-moment estimator, (ii) DP perturbations bias the second-moment estimator, and (iii) DP amplify AdamW sensitivity to local overfitting, worsening client drift. We propose DP-FedAdamW, the first AdamW-based optimizer for DPFL. It restores AdamW under DP by stabilizing second-moment variance, removing DP-induced bias, and aligning local updates to the global descent to curb client drift. Theoretically, we establish an unbiased second-moment estimator and prove a linearly accelerated convergence rate without any heterogeneity assumption, while providing tighter $(\varepsilon,\delta)$-DP guarantees. Our empirical results demonstrate the effectiveness of DP-FedAdamW across language and vision Transformers and ResNet-18. On Tiny-ImageNet (Swin-Base, $\varepsilon=1$), DP-FedAdamW outperforms the state-of-the-art (SOTA) by 5.83\%. The code is available in Appendix.

顶级标签: machine learning model training systems
详细标签: federated learning differential privacy optimizer adamw privacy-preserving ml 或 搜索:

DP-FedAdamW:一种用于差分隐私联邦大模型的高效优化器 / DP-FedAdamW: An Efficient Optimizer for Differentially Private Federated Large Models


1️⃣ 一句话总结

这篇论文提出了一种名为DP-FedAdamW的新型优化器,它成功解决了差分隐私联邦学习中AdamW优化器因数据差异和隐私噪声导致的性能下降问题,在保证隐私的同时显著提升了大型模型的训练效率和精度。

源自 arXiv: 2602.19945