SOLAR:通过子空间导向的潜在适配器重参数化实现通信高效的模型适配 / SOLAR: Communication-Efficient Model Adaptation via Subspace-Oriented Latent Adapter Reparametrization
1️⃣ 一句话总结
这篇论文提出了一种名为SOLAR的压缩方法,它能将大模型微调时产生的适配器参数大幅压缩,从而显著降低在分布式系统或边缘设备上部署时的通信和存储开销,同时保持模型性能。
Parameter-efficient fine-tuning (PEFT) methods, such as LoRA, enable scalable adaptation of foundation models by injecting low-rank adapters. However, their communication and storage costs remain a major bottleneck in resource-constrained settings. We propose SOLAR (Subspace-Oriented Latent Adapter Reparameterization), a post-training compression framework that substantially reduces the communication cost (i.e., the number of parameters to transmit or store) of PEFT adapters. SOLAR expresses each PEFT update as a linear combination of basis vectors formed from the foundation model's singular vectors with controlled random perturbations. By exploiting the subspace similarity (the alignment of principal directions) between the foundation model and task-specific fine-tuned updates, SOLAR decouples the adapter size from PEFT structure and ensures compact yet expressive representations. It is model-agnostic and compatible with existing PEFT methods, including LoRA, AdaLoRA, and other adapter modules. We theoretically establish a bound on the reconstruction error. Experiments on language and vision tasks using LLaMA, GPT, and ViT models demonstrate that SOLAR preserves task performance while significantly reducing model representation sizes, offering an effective and communication-efficient solution for deployment in distributed systems and edge devices.
SOLAR:通过子空间导向的潜在适配器重参数化实现通信高效的模型适配 / SOLAR: Communication-Efficient Model Adaptation via Subspace-Oriented Latent Adapter Reparametrization
这篇论文提出了一种名为SOLAR的压缩方法,它能将大模型微调时产生的适配器参数大幅压缩,从而显著降低在分布式系统或边缘设备上部署时的通信和存储开销,同时保持模型性能。
源自 arXiv: 2604.08368