反蒸馏指纹识别 / Antidistillation Fingerprinting
1️⃣ 一句话总结
这篇论文提出了一种名为‘反蒸馏指纹识别’的新方法,它通过优化指纹植入过程,使得大语言模型在知识蒸馏给第三方学生模型后,既能保持高质量的文本生成能力,又能被高效、可靠地追踪检测,解决了现有技术中检测效果与模型性能难以兼顾的问题。
Model distillation enables efficient emulation of frontier large language models (LLMs), creating a need for robust mechanisms to detect when a third-party student model has trained on a teacher model's outputs. However, existing fingerprinting techniques that could be used to detect such distillation rely on heuristic perturbations that impose a steep trade-off between generation quality and fingerprinting strength, often requiring significant degradation of utility to ensure the fingerprint is effectively internalized by the student. We introduce antidistillation fingerprinting (ADFP), a principled approach that aligns the fingerprinting objective with the student's learning dynamics. Building upon the gradient-based framework of antidistillation sampling, ADFP utilizes a proxy model to identify and sample tokens that directly maximize the expected detectability of the fingerprint in the student after fine-tuning, rather than relying on the incidental absorption of the un-targeted biases of a more naive watermark. Experiments on GSM8K and OASST1 benchmarks demonstrate that ADFP achieves a significant Pareto improvement over state-of-the-art baselines, yielding stronger detection confidence with minimal impact on utility, even when the student model's architecture is unknown.
反蒸馏指纹识别 / Antidistillation Fingerprinting
这篇论文提出了一种名为‘反蒸馏指纹识别’的新方法,它通过优化指纹植入过程,使得大语言模型在知识蒸馏给第三方学生模型后,既能保持高质量的文本生成能力,又能被高效、可靠地追踪检测,解决了现有技术中检测效果与模型性能难以兼顾的问题。
源自 arXiv: 2602.03812