NoRA:通过流形扩展打破低秩适应的线性天花板 / NoRA: Breaking the Linear Ceiling of Low-Rank Adaptation via Manifold Expansion
1️⃣ 一句话总结
这篇论文提出了一种名为NoRA的非线性低秩适应方法,通过引入SiLU门控和结构化丢弃来扩展模型表示能力,成功突破了传统LoRA方法在复杂推理任务中性能随参数增加而停滞的线性瓶颈,用更少的参数取得了更好的效果。
Low-Rank Adaptation (LoRA) dominates parameter-efficient fine-tuning (PEFT). However, it faces a critical ``linear ceiling'' in complex reasoning tasks: simply increasing the rank yields diminishing returns due to intrinsic linear constraints. We introduce NoRA (Non-linear Rank Adaptation), a weight-level parallel adapter that injects SiLU gating and structural dropout to induce manifold expansion. On the SlimOrca benchmark, NoRA breaks this linear barrier: NoRA remarkably at rank 64 (PPL 3.89) outperforms LoRA at rank 512 (PPL 3.90), demonstrating superior spectral efficiency. This advantage generalizes to mathematical reasoning, where NoRA achieves a perplexity of 1.97 on MathInstruct, significantly surpassing LoRA's saturation point of 2.07. Mechanism analysis via Singular Value Decomposition (SVD) confirms that NoRA activates the dormant tail of the singular value spectrum, effectively preventing the rank collapse observed in linear methods.
NoRA:通过流形扩展打破低秩适应的线性天花板 / NoRA: Breaking the Linear Ceiling of Low-Rank Adaptation via Manifold Expansion
这篇论文提出了一种名为NoRA的非线性低秩适应方法,通过引入SiLU门控和结构化丢弃来扩展模型表示能力,成功突破了传统LoRA方法在复杂推理任务中性能随参数增加而停滞的线性瓶颈,用更少的参数取得了更好的效果。
源自 arXiv: 2602.22911