菜单

关于 🐙 GitHub
arXiv 提交日期: 2025-12-18
📄 Abstract - DeContext as Defense: Safe Image Editing in Diffusion Transformers

In-context diffusion models allow users to modify images with remarkable ease and realism. However, the same power raises serious privacy concerns: personal images can be easily manipulated for identity impersonation, misinformation, or other malicious uses, all without the owner's consent. While prior work has explored input perturbations to protect against misuse in personalized text-to-image generation, the robustness of modern, large-scale in-context DiT-based models remains largely unexamined. In this paper, we propose DeContext, a new method to safeguard input images from unauthorized in-context editing. Our key insight is that contextual information from the source image propagates to the output primarily through multimodal attention layers. By injecting small, targeted perturbations that weaken these cross-attention pathways, DeContext breaks this flow, effectively decouples the link between input and output. This simple defense is both efficient and robust. We further show that early denoising steps and specific transformer blocks dominate context propagation, which allows us to concentrate perturbations where they matter most. Experiments on Flux Kontext and Step1X-Edit show that DeContext consistently blocks unwanted image edits while preserving visual quality. These results highlight the effectiveness of attention-based perturbations as a powerful defense against image manipulation.

顶级标签: computer vision model training systems
详细标签: diffusion transformers image editing privacy protection attention perturbation adversarial defense 或 搜索:

去上下文作为防御:扩散变换器中的安全图像编辑 / DeContext as Defense: Safe Image Editing in Diffusion Transformers


1️⃣ 一句话总结

这篇论文提出了一种名为DeContext的新方法,通过向图像添加微小的针对性扰动来干扰扩散模型中的跨注意力机制,从而有效阻止未经授权的图像编辑,保护个人照片不被恶意篡改,同时保持图像质量。


源自 arXiv: 2512.16625