TuneShift-KD:面向微调模型的知识蒸馏与迁移 / TuneShift-KD: Knowledge Distillation and Transfer for Fine-tuned Models
1️⃣ 一句话总结
这篇论文提出了一种名为TuneShift-KD的新方法,它能在不依赖原始专业数据的情况下,通过对比微调模型和基础模型的困惑度差异,自动识别并蒸馏出专业领域的知识,然后利用少量代表性提示生成合成数据集,从而将知识高效迁移到新的预训练模型中。
To embed domain-specific or specialized knowledge into pre-trained foundation models, fine-tuning using techniques such as parameter efficient fine-tuning (e.g. LoRA) is a common practice. However, as new LLM architectures and pre-trained models emerge, transferring this specialized knowledge to newer models becomes an important task. In many scenarios, the original specialized data may be unavailable due to privacy or commercial restrictions, necessitating distillation and transfer of this specialized knowledge from the fine-tuned base model to a different pre-trained model. We present TuneShift-KD, a novel approach that automatically distills specialized knowledge from a fine-tuned model to a target model using only a few examples representative of the specialized information. Our key insight is that specialized knowledge can be identified through perplexity differences between base and fine-tuned models: prompts where the fine-tuned model responds confidently (low perplexity), but the base model struggles (high perplexity), indicate queries corresponding to the specialized knowledge learned by the fine-tuned model. TuneShift-KD leverages this insight to create a synthetic training dataset to transfer the specialized knowledge. Using an iterative process, TuneShift-KD generates more prompts similar to those that generated responses with specialized knowledge. TuneShift-KD does not require training discriminators or access to training datasets. It is an automated approach that only requires the initial fine-tuned and base models and a few representative prompts. Our experiments demonstrate that models fine-tuned using TuneShift-KD achieve higher accuracy than prior approaches, enabling ease of deployment and more effective transfer of the specialized knowledge.
TuneShift-KD:面向微调模型的知识蒸馏与迁移 / TuneShift-KD: Knowledge Distillation and Transfer for Fine-tuned Models
这篇论文提出了一种名为TuneShift-KD的新方法,它能在不依赖原始专业数据的情况下,通过对比微调模型和基础模型的困惑度差异,自动识别并蒸馏出专业领域的知识,然后利用少量代表性提示生成合成数据集,从而将知识高效迁移到新的预训练模型中。
源自 arXiv: 2603.24518