FIRE:用于平衡稳定性-可塑性权衡的Frobenius等距重初始化方法 / FIRE: Frobenius-Isometry Reinitialization for Balancing the Stability-Plasticity Tradeoff
1️⃣ 一句话总结
这篇论文提出了一种名为FIRE的新方法,它通过一个数学优化问题来智能地调整神经网络权重,从而在持续学习任务中巧妙地平衡了‘记住旧知识’和‘学习新任务’之间的矛盾,并在图像、语言和强化学习等多个领域取得了更好的效果。
Deep neural networks trained on nonstationary data must balance stability (i.e., retaining prior knowledge) and plasticity (i.e., adapting to new tasks). Standard reinitialization methods, which reinitialize weights toward their original values, are widely used but difficult to tune: conservative reinitializations fail to restore plasticity, while aggressive ones erase useful knowledge. We propose FIRE, a principled reinitialization method that explicitly balances the stability-plasticity tradeoff. FIRE quantifies stability through Squared Frobenius Error (SFE), measuring proximity to past weights, and plasticity through Deviation from Isometry (DfI), reflecting weight isotropy. The reinitialization point is obtained by solving a constrained optimization problem, minimizing SFE subject to DfI being zero, which is efficiently approximated by Newton-Schulz iteration. FIRE is evaluated on continual visual learning (CIFAR-10 with ResNet-18), language modeling (OpenWebText with GPT-0.1B), and reinforcement learning (HumanoidBench with SAC and Atari games with DQN). Across all domains, FIRE consistently outperforms both naive training without intervention and standard reinitialization methods, demonstrating effective balancing of the stability-plasticity tradeoff.
FIRE:用于平衡稳定性-可塑性权衡的Frobenius等距重初始化方法 / FIRE: Frobenius-Isometry Reinitialization for Balancing the Stability-Plasticity Tradeoff
这篇论文提出了一种名为FIRE的新方法,它通过一个数学优化问题来智能地调整神经网络权重,从而在持续学习任务中巧妙地平衡了‘记住旧知识’和‘学习新任务’之间的矛盾,并在图像、语言和强化学习等多个领域取得了更好的效果。
源自 arXiv: 2602.08040