菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-03-03
📄 Abstract - Adaptive Methods Are Preferable in High Privacy Settings: An SDE Perspective

Differential Privacy (DP) is becoming central to large-scale training as privacy regulations tighten. We revisit how DP noise interacts with adaptivity in optimization through the lens of stochastic differential equations, providing the first SDE-based analysis of private optimizers. Focusing on DP-SGD and DP-SignSGD under per-example clipping, we show a sharp contrast under fixed hyperparameters: DP-SGD converges at a Privacy-Utility Trade-Off of $\mathcal{O}(1/\varepsilon^2)$ with speed independent of $\varepsilon$, while DP-SignSGD converges at a speed linear in $\varepsilon$ with an $\mathcal{O}(1/\varepsilon)$ trade-off, dominating in high-privacy or large batch noise regimes. By contrast, under optimal learning rates, both methods achieve comparable theoretical asymptotic performance; however, the optimal learning rate of DP-SGD scales linearly with $\varepsilon$, while that of DP-SignSGD is essentially $\varepsilon$-independent. This makes adaptive methods far more practical, as their hyperparameters transfer across privacy levels with little or no re-tuning. Empirical results confirm our theory across training and test metrics, and empirically extend from DP-SignSGD to DP-Adam.

顶级标签: machine learning theory model training
详细标签: differential privacy optimization stochastic differential equations adaptive methods privacy-utility trade-off 或 搜索:

在高隐私设置下自适应方法更优:一个随机微分方程的视角 / Adaptive Methods Are Preferable in High Privacy Settings: An SDE Perspective


1️⃣ 一句话总结

这篇论文通过随机微分方程分析发现,在严格的差分隐私训练中,自适应优化方法(如DP-SignSGD、DP-Adam)因其超参数对隐私级别不敏感而比传统方法(如DP-SGD)更实用,尤其在隐私要求高或噪声大的场景下表现更优。

源自 arXiv: 2603.03226