并非所有方向都重要:迈向结构化与任务感知的低秩自适应 / Not All Directions Matter: Toward Structured and Task-Aware Low-Rank Adaptation
1️⃣ 一句话总结
这篇论文提出了一个名为StructLoRA的新框架,它通过智能筛选出对任务真正重要的参数更新方向并协调不同层之间的更新,从而在微调大模型时获得更好的性能,且不增加推理成本。
Low-Rank Adaptation (LoRA) has become a cornerstone of parameter-efficient fine-tuning (PEFT). Yet, its efficacy is hampered by two fundamental limitations: semantic drift, by treating all update directions with equal importance, and structural incoherence, from adapting layers independently, resulting in suboptimal, uncoordinated updates. To remedy these, we propose StructLoRA, a framework that addresses both limitations through a principled, dual-component design: (1) an Information Bottleneck-guided filter that prunes task-irrelevant directions to mitigate semantic drift, and (2) a lightweight, training-only graph-based coordinator that enforces inter-layer consistency to resolve structural incoherence. Extensive experiments across large language model , vision language model, and vision model (including LLaMA, LLaVA, and ViT) demonstrate that StructLoRA consistently establishes a new state-of-the-art, outperforming not only vanilla LoRA but also advanced dynamic rank allocation and sparsity-based methods. Notably, the benefits are particularly pronounced in challenging low-rank and low-data regimes. Crucially, since our proposed modules operate only during training, StructLoRA enhances performance with zero additional inference cost, advancing the focus of PEFT -- from mere parameter compression to a more holistic optimization of information quality and structural integrity.
并非所有方向都重要:迈向结构化与任务感知的低秩自适应 / Not All Directions Matter: Toward Structured and Task-Aware Low-Rank Adaptation
这篇论文提出了一个名为StructLoRA的新框架,它通过智能筛选出对任务真正重要的参数更新方向并协调不同层之间的更新,从而在微调大模型时获得更好的性能,且不增加推理成本。
源自 arXiv: 2603.14228