面向模型合并的稀疏感知进化框架 / Sparsity-Aware Evolution for Model Merging
1️⃣ 一句话总结
这篇论文提出了一种新的模型合并方法,它像生物进化一样,通过不断修剪和合并模型,并特别鼓励生成结构更精简(参数更少)的模型,从而在各种大型语言模型测试中更可靠地提升合并效果。
We propose a sparsity-aware evolutionary (SAE) framework for model merging that involves iterative pruning-merging cycles to act as a novel mutation operator. We incorporate the sparsity constraints into the score function, which steers the evolutionary process to favor more sparse models, in addition to other conventional performance scores. Interestingly, the by-product of \textit{competition} for sparsity introduces an extra local \textit{attraction} and interplay into the evolutionary process: if one competitor has more zero elements, the other competitor's non-zero elements will occupy those positions, even though the less sparse competitor loses to the more sparse competitor in other positions. The proposed pipeline is evaluated on a variety of large-scale LLM benchmarks. Experiments demonstrate that our approach can improve model merging reliability across multiple benchmarks, and is easy to incorporate due to its simplicity and being orthogonal to most existing approaches.
面向模型合并的稀疏感知进化框架 / Sparsity-Aware Evolution for Model Merging
这篇论文提出了一种新的模型合并方法,它像生物进化一样,通过不断修剪和合并模型,并特别鼓励生成结构更精简(参数更少)的模型,从而在各种大型语言模型测试中更可靠地提升合并效果。
源自 arXiv: 2602.08218