当学习带来伤害:用于实时在线训练的固定极点循环神经网络 / When Learning Hurts: Fixed-Pole RNN for Real-Time Online Training
1️⃣ 一句话总结
这篇论文通过理论和实验证明,在数据有限、需要实时学习的场景下,固定循环神经网络内部动态的极点参数,只训练输出层,比同时训练所有参数效果更好、更稳定且计算成本更低。
Recurrent neural networks (RNNs) can be interpreted as discrete-time state-space models, where the state evolution corresponds to an infinite-impulse-response (IIR) filtering operation governed by both feedforward weights and recurrent poles. While, in principle, all parameters including pole locations can be optimized via backpropagation through time (BPTT), such joint learning incurs substantial computational overhead and is often impractical for applications with limited training data. Echo state networks (ESNs) mitigate this limitation by fixing the recurrent dynamics and training only a linear readout, enabling efficient and stable online adaptation. In this work, we analytically and empirically examine why learning recurrent poles does not provide tangible benefits in data-constrained, real-time learning scenarios. Our analysis shows that pole learning renders the weight optimization problem highly non-convex, requiring significantly more training samples and iterations for gradient-based methods to converge to meaningful solutions. Empirically, we observe that for complex-valued data, gradient descent frequently exhibits prolonged plateaus, and advanced optimizers offer limited improvement. In contrast, fixed-pole architectures induce stable and well-conditioned state representations even with limited training data. Numerical results demonstrate that fixed-pole networks achieve superior performance with lower training complexity, making them more suitable for online real-time tasks.
当学习带来伤害:用于实时在线训练的固定极点循环神经网络 / When Learning Hurts: Fixed-Pole RNN for Real-Time Online Training
这篇论文通过理论和实验证明,在数据有限、需要实时学习的场景下,固定循环神经网络内部动态的极点参数,只训练输出层,比同时训练所有参数效果更好、更稳定且计算成本更低。
源自 arXiv: 2602.21454