DR-LoRA:面向专家混合模型调优的动态秩LoRA方法 / DR-LoRA: Dynamic Rank LoRA for Mixture-of-Experts Adaptation
1️⃣ 一句话总结
这篇论文提出了一种名为DR-LoRA的新方法,它能够根据任务需求,动态地为大语言模型中的不同专家模块分配不同的学习能力,从而在相同参数预算下实现更高效、性能更好的模型调优。
Mixture-of-Experts (MoE) has become a prominent paradigm for scaling Large Language Models (LLMs). Parameter-efficient fine-tuning (PEFT), such as LoRA, is widely adopted to adapt pretrained MoE LLMs to downstream tasks. However, existing approaches assign identical LoRA ranks to all experts, overlooking the intrinsic functional specialization within MoE LLMs. This uniform allocation leads to resource mismatch, task-relevant experts are under-provisioned while less relevant ones receive redundant parameters. We propose a Dynamic Rank LoRA framework named DR-LoRA, which dynamically grows expert LoRA ranks during fine-tuning based on task-specific demands. DR-LoRA employs an Expert Saliency Scoring mechanism that integrates expert routing frequency and LoRA rank importance to quantify each expert's demand for additional capacity. Experts with higher saliency scores are prioritized for rank expansion, enabling the automatic formation of a heterogeneous rank distribution tailored to the target task. Experiments on multiple benchmarks demonstrate that DR-LoRA consistently outperforms standard LoRA and static allocation strategies under the same parameter budget, achieving superior task performance with more efficient parameter utilization.
DR-LoRA:面向专家混合模型调优的动态秩LoRA方法 / DR-LoRA: Dynamic Rank LoRA for Mixture-of-Experts Adaptation
这篇论文提出了一种名为DR-LoRA的新方法,它能够根据任务需求,动态地为大语言模型中的不同专家模块分配不同的学习能力,从而在相同参数预算下实现更高效、性能更好的模型调优。
源自 arXiv: 2601.04823