菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-03-04
📄 Abstract - Spectral Surgery: Training-Free Refinement of LoRA via Gradient-Guided Singular Value Reweighting

Low-Rank Adaptation (LoRA) improves downstream performance by restricting task updates to a low-rank parameter subspace, yet how this limited capacity is allocated within a trained adapter remains unclear. Through a geometric and empirical study across multiple tasks and backbones, we find that trained LoRA updates often exhibit an inefficient spectrum: task effects concentrate in a small subset of singular directions, while many remaining components are neutral or detrimental, motivating post-hoc refinement within the learned subspace. We propose Spectral Surgery, a training-free refinement that decomposes a LoRA update with SVD, estimates per-component sensitivity using gradients on a small calibration set, and reweights singular values under a magnitude constraint while keeping the learned directions fixed. Across Llama-3.1-8B and Qwen3-8B on four benchmarks, Spectral Surgery yields consistent gains (up to +4.4 points on CommonsenseQA and +2.4 pass@1 on HumanEval) by adjusting only $\approx 1{,}000$ scalar coefficients. These results demonstrate that SVD-structured, low-cost parameter editing can serve as a practical route to improving trained LoRA adapters in a purely post-hoc manner.

顶级标签: model training llm machine learning
详细标签: lora fine-tuning singular value decomposition parameter efficiency post-hoc refinement 或 搜索:

谱手术:通过梯度引导的奇异值重加权实现无需训练的LoRA微调 / Spectral Surgery: Training-Free Refinement of LoRA via Gradient-Guided Singular Value Reweighting


1️⃣ 一句话总结

这篇论文提出了一种名为‘谱手术’的新方法,它能在不重新训练的情况下,通过分析并调整现有LoRA适配器中各个成分的重要性,来显著提升大语言模型在特定任务上的表现。

源自 arXiv: 2603.03995