菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-05-11
📄 Abstract - Muown: Row-Norm Control for Muon Optimization

Muon has emerged as a strong competitor to AdamW for language model pre-training, yet its behavior at scale is sensitive to weight decay. Recent work has observed that, for Muon without decoupled weight decay, the spectral norm of weight matrices drifts upward over training. Through a decomposition of the spectral norm into a row-magnitude factor and a row-coherence factor, we identify the former as the empirical driver of this drift under Muon, while the latter remains well-behaved along the trajectory. Motivated by this diagnosis, we introduce Muown, a drop-in replacement for Muon that treats the row-magnitude vector as an explicit optimizer variable, updating it under the $\ell_\infty$ geometry induced by the decomposition, while applying Muon unchanged to the remaining direction component. We prove that Muown attains the optimal non-convex rates in both deterministic and stochastic regimes under a dual norm aligned with the underlying geometries and with a stochastic noise coefficient that empirically remains below that of Muon throughout training. Across GPT-style pre-training on FineWeb-Edu with model sizes from 124M up to 2.7B parameters, Muown improves perplexity over Muon, SOAP, AdamW, and Lion. It also widens the plateau of near-optimal learning rates across model scales, reduces sensitivity to weight decay, and avoids the spectral norm drift at negligible step-time overhead when appropriately sharded.

顶级标签: llm model training machine learning
详细标签: optimization weight decay spectral norm muon optimizer language model pre-training 或 搜索:

Muown:用于缪子优化的行范数控制 / Muown: Row-Norm Control for Muon Optimization


1️⃣ 一句话总结

本文提出了一种名为Muown的新优化器,它通过将权重矩阵的行范数作为独立变量进行显式控制,解决了Muon优化器在大规模语言模型训练中遇到的谱范数漂移问题,在多个模型规模下均取得了比Muon、AdamW等优化器更好的性能,且对超参数(如学习率和权重衰减)的敏感性更低。

源自 arXiv: 2605.10797