菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-02-23
📄 Abstract - Wasserstein Distributionally Robust Online Learning

We study distributionally robust online learning, where a risk-averse learner updates decisions sequentially to guard against worst-case distributions drawn from a Wasserstein ambiguity set centered at past observations. While this paradigm is well understood in the offline setting through Wasserstein Distributionally Robust Optimization (DRO), its online extension poses significant challenges in both convergence and computation. In this paper, we address these challenges. First, we formulate the problem as an online saddle-point stochastic game between a decision maker and an adversary selecting worst-case distributions, and propose a general framework that converges to a robust Nash equilibrium coinciding with the solution of the corresponding offline Wasserstein DRO problem. Second, we address the main computational bottleneck, which is the repeated solution of worst-case expectation problems. For the important class of piecewise concave loss functions, we propose a tailored algorithm that exploits problem geometry to achieve substantial speedups over state-of-the-art solvers such as Gurobi. The key insight is a novel connection between the worst-case expectation problem, an inherently infinite-dimensional optimization problem, and a classical and tractable budget allocation problem, which is of independent interest.

顶级标签: machine learning theory
详细标签: distributionally robust optimization online learning wasserstein distance saddle-point optimization worst-case expectation 或 搜索:

Wasserstein分布鲁棒在线学习 / Wasserstein Distributionally Robust Online Learning


1️⃣ 一句话总结

这篇论文提出了一种新的在线学习方法,让风险厌恶的决策者能够根据历史数据动态调整策略,以抵御最坏情况的数据分布变化,并通过将复杂的无限维优化问题转化为经典的预算分配问题,大幅提升了计算效率。

源自 arXiv: 2602.20403