菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-03-30
📄 Abstract - The Unreasonable Effectiveness of Scaling Laws in AI

Classical AI scaling laws, especially for pre-training, describe how training loss decreases with compute in a power-law form. Their effectiveness has a basic and very practical sense: they make progress predictable, albeit at a declining rate. Yet their effectiveness is also unreasonable in two further senses. First, these laws are largely empirical and observational, but they appear repeatedly across model families and increasingly across training-adjacent regimes. Second, despite the diminishing returns they predict, progress in practice has often continued through rapidly improving efficiency, visible for example in falling cost per token. This paper argues that both features arise from the same source: scaling laws are unusually effective because they abstract away from many realization details. The compute variable is best understood as logical compute, an implementation-agnostic notion of model-side work, while the practical burden of scaling depends on how efficiently real resources are converted into that compute. This abstraction helps explain both why the laws travel so well across settings and why they give rise to a persistent efficiency game in hardware, algorithms, and systems. Once efficiency is made explicit, the main practical question becomes how many efficiency doublings are required to keep scaling productive despite diminishing returns. Under that view, diminishing returns are not only a geometric flattening of the loss curve, but also rising pressure for cost reduction, system-level innovation, and the breakthroughs needed to sustain Moore-like efficiency doublings.

顶级标签: machine learning model training theory
详细标签: scaling laws compute efficiency power-law model performance diminishing returns 或 搜索:

AI中缩放定律的惊人有效性 / The Unreasonable Effectiveness of Scaling Laws in AI


1️⃣ 一句话总结

这篇论文认为,AI缩放定律之所以能广泛适用并持续推动进步,是因为它抽象掉了具体实现细节,将计算视为一种与实现方式无关的“逻辑计算”,从而既解释了其跨场景的普适性,也揭示了硬件、算法和系统层面持续进行效率竞赛的内在动力。

源自 arXiv: 2603.28507