菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-03-16
📄 Abstract - Deep learning and the rate of approximation by flows

We investigate the dependence of the approximation capacity of deep residual networks on its depth in a continuous dynamical systems setting. This can be formulated as the general problem of quantifying the minimal time-horizon required to approximate a diffeomorphism by flows driven by a given family $\mathcal F$ of vector fields. We show that this minimal time can be identified as a geodesic distance on a sub-Finsler manifold of diffeomorphisms, where the local geometry is characterised by a variational principle involving $\mathcal F$. This connects the learning efficiency of target relationships to their compatibility with the learning architectural choice. Further, the results suggest that the key approximation mechanism in deep learning, namely the approximation of functions by composition or dynamics, differs in a fundamental way from linear approximation theory, where linear spaces and norm-based rate estimates are replaced by manifolds and geodesic distances.

顶级标签: theory machine learning model training
详细标签: deep residual networks approximation theory dynamical systems sub-finsler geometry geodesic distance 或 搜索:

深度学习与流逼近的速率 / Deep learning and the rate of approximation by flows


1️⃣ 一句话总结

这篇论文通过将深度残差网络的逼近能力与微分同胚流所需的最短时间联系起来,揭示了深度学习的核心逼近机制本质上是基于流形上的测地距离,而非传统的线性空间和范数估计,从而阐明了网络架构与目标函数兼容性对学习效率的决定性影响。

源自 arXiv: 2603.15363