菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-05-04
📄 Abstract - A Parameter-Free First-Order Algorithm for Non-Convex Optimization with $\tilde{\mkern1mu O}(ε^{-5/3})$ Global Rate

We introduce PF-AGD, the first parameter-free, deterministic, accelerated first-order method to achieve $O(\epsilon^{-5/3}\log(1/\epsilon))$ oracle complexity bound when minimizing sufficiently smooth, non-convex functions; this is the best-known bound for first-order methods on smooth non-convex objectives. Unlike existing methods possessing this rate that require a priori knowledge of smoothness constants, we use an adaptive backtracking scheme and a gradient-based restart mechanism to estimate local curvature. This yields a practical algorithm that matches best-known theoretical rates. Empirically, PF-AGD outperforms the practical variant of AGD-Until-Guilty (Carmon et al., 2017), as well as other parameter-free variants, and is a viable alternative to nonlinear conjugate gradient methods.

顶级标签: machine learning theory
详细标签: first-order optimization non-convex optimization parameter-free accelerated gradient descent global rate 或 搜索:

一种无参数的一阶算法,用于非凸优化,达到全局收敛速率~O(ε^{-5/3}) / A Parameter-Free First-Order Algorithm for Non-Convex Optimization with $\tilde{\mkern1mu O}(ε^{-5/3})$ Global Rate


1️⃣ 一句话总结

本文提出了一种名为PF-AGD的算法,它无需用户手动调整参数,就能在优化非凸函数时达到目前已知最快的计算效率,并且在实际测试中比同类方法表现更好。

源自 arXiv: 2605.02127