菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-05-07
📄 Abstract - Rethinking Adapter Placement: A Dominant Adaptation Module Perspective

Low-rank adaptation (LoRA) is a widely used parameter-efficient fine-tuning method that places trainable low-rank adapters into frozen pre-trained models. Recent studies show that using fewer LoRA adapters may still maintain or even improve performance, but existing methods still distribute adapters broadly, leaving where to place a limited number of adapters to maximize performance largely open. To investigate this, we introduce PAGE (Projected Adapter Gradient Energy), a gradient-based sensitivity probe that estimates the initial trainable gradient energy available to each candidate LoRA adapter. Surprisingly, we find that PAGE is highly concentrated on a single shallow FFN down-projection across two model families and four downstream tasks. We term this module the dominant adaptation module and show that its layer index is architecture-dependent but task-stable. Motivated by this finding, we propose DomLoRA, a placement method that places a single adapter at the dominant adaptation module. With only ~0.7% of vanilla LoRA's trainable parameters, DomLoRA outperforms it on average across various downstream tasks, including instruction following, mathematical reasoning, code generation, and multi-turn conversation. This method also improves other LoRA variants, supporting the dominant adaptation module perspective as a practical placement guideline.

顶级标签: llm model training model evaluation
详细标签: lora adapter placement parameter-efficient fine-tuning gradient sensitivity domlora 或 搜索:

重新思考适配器放置:一种主导适配模块视角 / Rethinking Adapter Placement: A Dominant Adaptation Module Perspective


1️⃣ 一句话总结

本文发现,在微调预训练模型时,只需在一个关键的单层前馈网络降维投影模块中放置一个极小的适配器,就能超越传统方法在全模型多处放置大量适配器的效果,从而提出了一种更简单高效的适配器放置策略。

源自 arXiv: 2605.06183