菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-02-04
📄 Abstract - Gradient Flow Through Diagram Expansions: Learning Regimes and Explicit Solutions

We develop a general mathematical framework to analyze scaling regimes and derive explicit analytic solutions for gradient flow (GF) in large learning problems. Our key innovation is a formal power series expansion of the loss evolution, with coefficients encoded by diagrams akin to Feynman diagrams. We show that this expansion has a well-defined large-size limit that can be used to reveal different learning phases and, in some cases, to obtain explicit solutions of the nonlinear GF. We focus on learning Canonical Polyadic (CP) decompositions of high-order tensors, and show that this model has several distinct extreme lazy and rich GF regimes such as free evolution, NTK and under- and over-parameterized mean-field. We show that these regimes depend on the parameter scaling, tensor order, and symmetry of the model in a specific and subtle way. Moreover, we propose a general approach to summing the formal loss expansion by reducing it to a PDE; in a wide range of scenarios, it turns out to be 1st order and solvable by the method of characteristics. We observe a very good agreement of our theoretical predictions with experiment.

顶级标签: theory model training machine learning
详细标签: gradient flow scaling regimes tensor decomposition explicit solutions diagram expansion 或 搜索:

通过图展开的梯度流:学习机制与显式解 / Gradient Flow Through Diagram Expansions: Learning Regimes and Explicit Solutions


1️⃣ 一句话总结

这篇论文建立了一个用类似费曼图的展开方法来分析大规模机器学习中梯度流行为的数学框架,揭示了不同参数设置下模型的学习阶段,并能为某些非线性问题提供显式解。

源自 arXiv: 2602.04548