菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-03-30
📄 Abstract - Temporal Credit Is Free

Recurrent networks do not need Jacobian propagation to adapt online. The hidden state already carries temporal credit through the forward pass; immediate derivatives suffice if you stop corrupting them with stale trace memory and normalize gradient scales across parameter groups. An architectural rule predicts when normalization is needed: \b{eta}2 is required when gradients must pass through a nonlinear state update with no output bypass, and unnecessary otherwise. Across ten architectures, real primate neural data, and streaming ML benchmarks, immediate derivatives with RMSprop match or exceed full RTRL, scaling to n = 1024 at 1000x less memory.

顶级标签: machine learning model training theory
详细标签: recurrent neural networks online learning gradient propagation rtrl memory efficiency 或 搜索:

时间信用是免费的 / Temporal Credit Is Free


1️⃣ 一句话总结

这篇论文提出了一种新方法,让循环神经网络在在线学习时无需复杂的反向传播,只需利用前向传播中已有的信息,就能高效地更新网络参数,从而大大降低了计算和内存消耗,并在多种任务上取得了与传统方法相当甚至更好的效果。

源自 arXiv: 2603.28750