重新思考机器遗忘:通过关键删除实现遗忘的模型设计 / Rethinking Machine Unlearning: Models Designed to Forget via Key Deletion
1️⃣ 一句话总结
这篇论文提出了一种名为‘设计遗忘’的新方法,通过训练模型使其天生具备遗忘能力,只需删除特定数据的关键标识即可实现即时、无需数据访问的遗忘,相比传统事后修正方法更高效实用。
Machine unlearning is rapidly becoming a practical requirement, driven by privacy regulations, data errors, and the need to remove harmful or corrupted training samples. Despite this, most existing methods tackle the problem purely from a post-hoc perspective. They attempt to erase the influence of targeted training samples through parameter updates that typically require access to the full training data. This creates a mismatch with real deployment scenarios where unlearning requests can be anticipated, revealing a fundamental limitation of post-hoc approaches. We propose \textit{unlearning by design}, a novel paradigm in which models are directly trained to support forgetting as an inherent capability. We instantiate this idea with Machine UNlearning via KEY deletion (MUNKEY), a memory augmented transformer that decouples instance-specific memorization from model weights. Here, unlearning corresponds to removing the instance-identifying key, enabling direct zero-shot forgetting without weight updates or access to the original samples or labels. Across natural image benchmarks, fine-grained recognition, and medical datasets, MUNKEY outperforms all post-hoc baselines. Our results establish that unlearning by design enables fast, deployment-oriented unlearning while preserving predictive performance.
重新思考机器遗忘:通过关键删除实现遗忘的模型设计 / Rethinking Machine Unlearning: Models Designed to Forget via Key Deletion
这篇论文提出了一种名为‘设计遗忘’的新方法,通过训练模型使其天生具备遗忘能力,只需删除特定数据的关键标识即可实现即时、无需数据访问的遗忘,相比传统事后修正方法更高效实用。
源自 arXiv: 2603.15033