菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-02-03
📄 Abstract - Antidistillation Fingerprinting

Model distillation enables efficient emulation of frontier large language models (LLMs), creating a need for robust mechanisms to detect when a third-party student model has trained on a teacher model's outputs. However, existing fingerprinting techniques that could be used to detect such distillation rely on heuristic perturbations that impose a steep trade-off between generation quality and fingerprinting strength, often requiring significant degradation of utility to ensure the fingerprint is effectively internalized by the student. We introduce antidistillation fingerprinting (ADFP), a principled approach that aligns the fingerprinting objective with the student's learning dynamics. Building upon the gradient-based framework of antidistillation sampling, ADFP utilizes a proxy model to identify and sample tokens that directly maximize the expected detectability of the fingerprint in the student after fine-tuning, rather than relying on the incidental absorption of the un-targeted biases of a more naive watermark. Experiments on GSM8K and OASST1 benchmarks demonstrate that ADFP achieves a significant Pareto improvement over state-of-the-art baselines, yielding stronger detection confidence with minimal impact on utility, even when the student model's architecture is unknown.

顶级标签: llm model training model evaluation
详细标签: model distillation fingerprinting detection watermarking fine-tuning 或 搜索:

反蒸馏指纹识别 / Antidistillation Fingerprinting


1️⃣ 一句话总结

这篇论文提出了一种名为‘反蒸馏指纹识别’的新方法,它通过优化指纹植入过程,使得大语言模型在知识蒸馏给第三方学生模型后,既能保持高质量的文本生成能力,又能被高效、可靠地追踪检测,解决了现有技术中检测效果与模型性能难以兼顾的问题。

源自 arXiv: 2602.03812