通过对数概率的各向异性检测和缓解扩散模型中的记忆效应 / Detecting and Mitigating Memorization in Diffusion Models through Anisotropy of the Log-Probability
1️⃣ 一句话总结
这篇论文提出了一种新的方法来检测和缓解扩散模型中的记忆效应,该方法通过分析模型在生成过程中的内部信号(特别是对数概率分布的方向性),能够更快速、更准确地识别出模型是否在简单地复制训练数据,而不是进行真正的创造性生成。
Diffusion-based image generative models produce high-fidelity images through iterative denoising but remain vulnerable to memorization, where they unintentionally reproduce exact copies or parts of training images. Recent memorization detection methods are primarily based on the norm of score difference as indicators of memorization. We prove that such norm-based metrics are mainly effective under the assumption of isotropic log-probability distributions, which generally holds at high or medium noise levels. In contrast, analyzing the anisotropic regime reveals that memorized samples exhibit strong angular alignment between the guidance vector and unconditional scores in the low-noise setting. Through these insights, we develop a memorization detection metric by integrating isotropic norm and anisotropic alignment. Our detection metric can be computed directly on pure noise inputs via two conditional and unconditional forward passes, eliminating the need for costly denoising steps. Detection experiments on Stable Diffusion v1.4 and v2 show that our metric outperforms existing denoising-free detection methods while being at least approximately 5x faster than the previous best approach. Finally, we demonstrate the effectiveness of our approach by utilizing a mitigation strategy that adapts memorized prompts based on our developed metric.
通过对数概率的各向异性检测和缓解扩散模型中的记忆效应 / Detecting and Mitigating Memorization in Diffusion Models through Anisotropy of the Log-Probability
这篇论文提出了一种新的方法来检测和缓解扩散模型中的记忆效应,该方法通过分析模型在生成过程中的内部信号(特别是对数概率分布的方向性),能够更快速、更准确地识别出模型是否在简单地复制训练数据,而不是进行真正的创造性生成。
源自 arXiv: 2601.20642