📄
Abstract - On the Structural Limitations of Weight-Based Neural Adaptation and the Role of Reversible Behavioral Learning
Neural models are usually adapted through changes in parameters shared among model components via fine-tuning, alignment-based training, and reinforcement learning. These changes have been found effective in short-term optimization. However, they result in long-term alterations in the model's base behavior. In this study, we introduce the concept of structural irreversibility as a characteristic of shared-parameter model adaptation. This concept refers to the intertwining of task-specific objectives with the representational identity of the model. We show that when parameters are directly mutated, the resulting model behaves divergently from the original model. This divergence cannot be reversed deterministically without an explicit parameter snapshot. We introduce reversible behavioral learning, in which model behaviors are structurally dissociated from identity parameters and can be deterministically unloaded through an explicit unload process. We also introduce the Recoverability Factor as a normalized measure of behavioral recoverability and provide additional diagnostics based on model divergence. Experiments show that reversible model adaptation achieves rollback within numerical precision, whereas shared-parameter mutation exhibits persistent post-reset divergence.
论基于权重的神经适应之结构局限与可逆行为学习的作用 /
On the Structural Limitations of Weight-Based Neural Adaptation and the Role of Reversible Behavioral Learning
1️⃣ 一句话总结
这篇论文指出,传统通过直接修改神经网络权重来适应新任务的方法会永久性地改变模型的基础行为,并提出了一个名为‘可逆行为学习’的新方法,让模型能够像安装卸载软件一样,在不改变核心身份的前提下,确定性地学习和移除特定行为。