菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-04-28
📄 Abstract - Knowledge Distillation Must Account for What It Loses

This position paper argues that knowledge distillation must account for what it loses: student models should be judged not only by retained task scores, but by whether they preserve the teacher capabilities that make those scores reliable. This matters because distillation is increasingly used to turn large, often frontier models into deployable systems, yet headline metrics can hide losses in uncertainty, boundary behavior, process reliability, on-policy stability, grounding, privacy, safety, and diversity. We identify the retention assumption behind current evaluation and reframe distillation as a lossy projection of teacher behavior rather than a faithful copy. We then synthesize existing evidence into a taxonomy of off-metric distillation losses, showing that these losses are concrete, recurring, and measurable. To make the position actionable, we propose scenario-specific preservation targets and a Distillation Loss Statement that reports what was preserved, what was lost, and why the remaining losses are acceptable. The goal is not lossless distillation, but accountable distillation.

顶级标签: machine learning model training
详细标签: knowledge distillation model evaluation reliability safety accountability 或 搜索:

知识蒸馏必须考虑其损失了什么 / Knowledge Distillation Must Account for What It Loses


1️⃣ 一句话总结

本论文指出,在知识蒸馏过程中,不仅要关注学生模型在主要任务上的表现,还必须系统评估其是否保留了教师模型的可靠性关键能力(如不确定性判断、边界行为、过程可靠性等),并提出通过“蒸馏损失报告”机制,让蒸馏过程变得可问责、可评估。

源自 arXiv: 2604.25110