菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-04-14
📄 Abstract - Operationalising the Right to be Forgotten in LLMs: A Lightweight Sequential Unlearning Framework for Privacy-Aligned Deployment in Politically Sensitive Environments

Large Language Models (LLMs) are increasingly deployed in politically sensitive environments, where memorisation of personal data or confidential content raises regulatory concerns under frameworks such as the GDPR and its Right to be Forgotten. Translating such legal principles into large-scale generative systems presents significant technical challenges. We introduce a lightweight sequential unlearning framework that explicitly separates retention and suppression objectives. The method first stabilises benign capabilities through positive fine-tuning, then applies layer-restricted negative fine-tuning to suppress designated sensitive patterns while preserving general language competence. Experiments on the SemEval-2025 LLM Unlearning benchmark demonstrate effective behavioural suppression with minimal impact on factual accuracy and fluency. GPT-2 exhibits greater robustness than DistilGPT-2, highlighting the role of model capacity in privacy-aligned adaptation. We position sequential unlearning as a practical and reproducible mechanism for operationalising data erasure requirements in politically deployed LLMs.

顶级标签: llm model training systems
详细标签: machine unlearning privacy gdpr right to be forgotten sequential fine-tuning 或 搜索:

在LLMs中实现被遗忘权:一种适用于政治敏感环境隐私对齐部署的轻量级顺序遗忘框架 / Operationalising the Right to be Forgotten in LLMs: A Lightweight Sequential Unlearning Framework for Privacy-Aligned Deployment in Politically Sensitive Environments


1️⃣ 一句话总结

这篇论文提出了一种轻量级的顺序遗忘方法,让大型语言模型能有效‘忘记’指定的敏感信息,同时保持原有的语言能力,从而帮助模型在政治敏感环境中更好地满足数据隐私法规的要求。

源自 arXiv: 2604.12459