菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-04-21
📄 Abstract - ShadowPEFT: Shadow Network for Parameter-Efficient Fine-Tuning

Parameter-efficient fine-tuning (PEFT) reduces the training cost of full-parameter fine-tuning for large language models (LLMs) by training only a small set of task-specific parameters while freezing the pretrained backbone. However, existing approaches, such as Low-Rank Adaptation (LoRA), achieve adaptation by inserting independent low-rank perturbations directly to individual weights, resulting in a local parameterization of adaptation. We propose ShadowPEFT, a centralized PEFT framework that instead performs layer-level refinement through a depth-shared shadow module. At each transformer layer, ShadowPEFT maintains a parallel shadow state and evolves it repeatedly for progressively richer hidden states. This design shifts adaptation from distributed weight-space perturbations to a shared layer-space refinement process. Since the shadow module is decoupled from the backbone, it can be reused across depth, independently pretrained, and optionally deployed in a detached mode, benefiting edge computing scenarios. Experiments on generation and understanding benchmarks show that ShadowPEFT matches or outperforms LoRA and DoRA under comparable trainable-parameter budgets. Additional analyses on shadow pretraining, cross-dataset transfer, parameter scaling, inference latency, and system-level evaluation suggest that centralized layer-space adaptation is a competitive and flexible alternative to conventional low-rank PEFT.

顶级标签: llm model training systems
详细标签: parameter-efficient fine-tuning low-rank adaptation shadow network efficiency layer-space adaptation 或 搜索:

ShadowPEFT:基于影子网络的参数高效微调方法 / ShadowPEFT: Shadow Network for Parameter-Efficient Fine-Tuning


1️⃣ 一句话总结

ShadowPEFT提出了一种新的参数高效微调框架,通过一个可跨层共享的“影子”模块来逐步优化每层的隐藏状态,从而取代传统方法中对每个权重单独添加低秩扰动的方式,在保持模型大部分参数冻结的同时,以更灵活、更集中的方式实现高效微调,并在多项任务上取得了与LoRA相当或更优的性能。

源自 arXiv: 2604.19254