菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-02-18
📄 Abstract - Learning Distributed Equilibria in Linear-Quadratic Stochastic Differential Games: An $α$-Potential Approach

We analyze independent policy-gradient (PG) learning in $N$-player linear-quadratic (LQ) stochastic differential games. Each player employs a distributed policy that depends only on its own state and updates the policy independently using the gradient of its own objective. We establish global linear convergence of these methods to an equilibrium by showing that the LQ game admits an $\alpha$-potential structure, with $\alpha$ determined by the degree of pairwise interaction asymmetry. For pairwise-symmetric interactions, we construct an affine distributed equilibrium by minimizing the potential function and show that independent PG methods converge globally to this equilibrium, with complexity scaling linearly in the population size and logarithmically in the desired accuracy. For asymmetric interactions, we prove that independent projected PG algorithms converge linearly to an approximate equilibrium, with suboptimality proportional to the degree of asymmetry. Numerical experiments confirm the theoretical results across both symmetric and asymmetric interaction networks.

顶级标签: theory multi-agents reinforcement learning
详细标签: stochastic differential games policy gradient distributed equilibrium linear-quadratic games convergence analysis 或 搜索:

线性二次随机微分博弈中的分布式均衡学习:一种α-势能方法 / Learning Distributed Equilibria in Linear-Quadratic Stochastic Differential Games: An $α$-Potential Approach


1️⃣ 一句话总结

这篇论文证明了在多智能体线性二次随机微分博弈中,每个智能体仅依赖自身状态并独立更新策略的梯度学习方法,能够高效收敛到一个均衡点,其收敛速度与智能体数量线性相关,且均衡的近似程度取决于智能体间交互的不对称性。

源自 arXiv: 2602.16555