为何不可学习样本有效:一种互信息的新视角 / Why Do Unlearnable Examples Work: A Novel Perspective of Mutual Information
1️⃣ 一句话总结
这篇论文从互信息减少的新视角,解释了通过在数据中添加特定扰动来制造‘不可学习样本’以保护隐私的原理,并提出了一种通过最大化同类样本特征相似度来更有效阻止模型学习的新方法。
The volume of freely scraped data on the Internet has driven the tremendous success of deep learning. Along with this comes the growing concern about data privacy and security. Numerous methods for generating unlearnable examples have been proposed to prevent data from being illicitly learned by unauthorized deep models by impeding generalization. However, the existing approaches primarily rely on empirical heuristics, making it challenging to enhance unlearnable examples with solid explanations. In this paper, we analyze and improve unlearnable examples from a novel perspective: mutual information reduction. We demonstrate that effective unlearnable examples always decrease mutual information between clean features and poisoned features, and when the network gets deeper, the unlearnability goes better together with lower mutual information. Further, we prove from a covariance reduction perspective that minimizing the conditional covariance of intra-class poisoned features reduces the mutual information between distributions. Based on the theoretical results, we propose a novel unlearnable method called Mutual Information Unlearnable Examples (MI-UE) that reduces covariance by maximizing the cosine similarity among intra-class features, thus impeding the generalization effectively. Extensive experiments demonstrate that our approach significantly outperforms the previous methods, even under defense mechanisms.
为何不可学习样本有效:一种互信息的新视角 / Why Do Unlearnable Examples Work: A Novel Perspective of Mutual Information
这篇论文从互信息减少的新视角,解释了通过在数据中添加特定扰动来制造‘不可学习样本’以保护隐私的原理,并提出了一种通过最大化同类样本特征相似度来更有效阻止模型学习的新方法。
源自 arXiv: 2603.03725