菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-02-17
📄 Abstract - POP: Prior-fitted Optimizer Policies

Optimization refers to the task of finding extrema of an objective function. Classical gradient-based optimizers are highly sensitive to hyperparameter choices. In highly non-convex settings their performance relies on carefully tuned learning rates, momentum, and gradient accumulation. To address these limitations, we introduce POP (Prior-fitted Optimizer Policies), a meta-learned optimizer that predicts coordinate-wise step sizes conditioned on the contextual information provided in the optimization trajectory. Our model is learned on millions of synthetic optimization problems sampled from a novel prior spanning both convex and non-convex objectives. We evaluate POP on an established benchmark including 47 optimization functions of various complexity, where it consistently outperforms first-order gradient-based methods, non-convex optimization approaches (e.g., evolutionary strategies), Bayesian optimization, and a recent meta-learned competitor under matched budget constraints. Our evaluation demonstrates strong generalization capabilities without task-specific tuning.

顶级标签: machine learning model training theory
详细标签: optimization meta-learning hyperparameter tuning gradient descent non-convex optimization 或 搜索:

POP:基于先验拟合的优化器策略 / POP: Prior-fitted Optimizer Policies


1️⃣ 一句话总结

这篇论文提出了一种名为POP的元学习优化器,它通过从大量合成优化问题中学习,能够自动预测每一步的调整步长,从而在各种复杂函数优化任务中,无需手动调参就显著超越了传统梯度方法和其他先进优化算法。

源自 arXiv: 2602.15473