菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-02-03
📄 Abstract - UnHype: CLIP-Guided Hypernetworks for Dynamic LoRA Unlearning

Recent advances in large-scale diffusion models have intensified concerns about their potential misuse, particularly in generating realistic yet harmful or socially disruptive content. This challenge has spurred growing interest in effective machine unlearning, the process of selectively removing specific knowledge or concepts from a model without compromising its overall generative capabilities. Among various approaches, Low-Rank Adaptation (LoRA) has emerged as an effective and efficient method for fine-tuning models toward targeted unlearning. However, LoRA-based methods often exhibit limited adaptability to concept semantics and struggle to balance removing closely related concepts with maintaining generalization across broader meanings. Moreover, these methods face scalability challenges when multiple concepts must be erased simultaneously. To address these limitations, we introduce UnHype, a framework that incorporates hypernetworks into single- and multi-concept LoRA training. The proposed architecture can be directly plugged into Stable Diffusion as well as modern flow-based text-to-image models, where it demonstrates stable training behavior and effective concept control. During inference, the hypernetwork dynamically generates adaptive LoRA weights based on the CLIP embedding, enabling more context-aware, scalable unlearning. We evaluate UnHype across several challenging tasks, including object erasure, celebrity erasure, and explicit content removal, demonstrating its effectiveness and versatility. Repository: this https URL.

顶级标签: model training multi-modal machine learning
详细标签: machine unlearning stable diffusion lora clip text-to-image 或 搜索:

UnHype:用于动态LoRA遗忘的CLIP引导超网络 / UnHype: CLIP-Guided Hypernetworks for Dynamic LoRA Unlearning


1️⃣ 一句话总结

这篇论文提出了一个名为UnHype的新方法,它通过一个能根据输入内容动态调整参数的智能网络,帮助AI图像生成模型更精准、灵活地“忘记”特定事物(如名人或有害内容),同时不影响模型生成其他正常图片的能力。

源自 arXiv: 2602.03410