菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-02-11
📄 Abstract - The Magic Correlations: Understanding Knowledge Transfer from Pretraining to Supervised Fine-Tuning

Understanding how language model capabilities transfer from pretraining to supervised fine-tuning (SFT) is fundamental to efficient model development and data curation. In this work, we investigate four core questions: RQ1. To what extent do accuracy and confidence rankings established during pretraining persist after SFT? RQ2. Which benchmarks serve as robust cross-stage predictors and which are unreliable? RQ3. How do transfer dynamics shift with model scale? RQ4. How well does model confidence align with accuracy, as a measure of calibration quality? Does this alignment pattern transfer across training stages? We address these questions through a suite of correlation protocols applied to accuracy and confidence metrics across diverse data mixtures and model scales. Our experiments reveal that transfer reliability varies dramatically across capability categories, benchmarks, and scales -- with accuracy and confidence exhibiting distinct, sometimes opposing, scaling dynamics. These findings shed light on the complex interplay between pretraining decisions and downstream outcomes, providing actionable guidance for benchmark selection, data curation, and efficient model development.

顶级标签: llm model training model evaluation
详细标签: knowledge transfer fine-tuning pretraining scaling laws calibration 或 搜索:

魔法相关性:理解从预训练到监督微调的知识迁移 / The Magic Correlations: Understanding Knowledge Transfer from Pretraining to Supervised Fine-Tuning


1️⃣ 一句话总结

这篇论文通过系统研究发现,大语言模型从预训练到监督微调的知识迁移效果并不稳定,其可靠程度会因任务类型、评估基准和模型规模的不同而产生巨大差异,为如何高效选择和利用数据来训练模型提供了实用指导。

源自 arXiv: 2602.11217