📄 论文总结
扩散-SDPO:扩散模型的安全直接偏好优化 / Diffusion-SDPO: Safeguarded Direct Preference Optimization for Diffusion Models
1️⃣ 一句话总结
本文提出了一种名为Diffusion-SDPO的新方法,通过自适应调整优化过程中的梯度更新,解决了现有扩散模型在偏好学习时可能导致图像质量下降的问题,从而在保持简单高效的同时,显著提升了生成图像与人类偏好的对齐效果。
Text-to-image diffusion models deliver high-quality images, yet aligning them with human preferences remains challenging. We revisit diffusion-based Direct Preference Optimization (DPO) for these models and identify a critical pathology: enlarging the preference margin does not necessarily improve generation quality. In particular, the standard Diffusion-DPO objective can increase the reconstruction error of both winner and loser branches. Consequently, degradation of the less-preferred outputs can become sufficiently severe that the preferred branch is also adversely affected even as the margin grows. To address this, we introduce Diffusion-SDPO, a safeguarded update rule that preserves the winner by adaptively scaling the loser gradient according to its alignment with the winner gradient. A first-order analysis yields a closed-form scaling coefficient that guarantees the error of the preferred output is non-increasing at each optimization step. Our method is simple, model-agnostic, broadly compatible with existing DPO-style alignment frameworks and adds only marginal computational overhead. Across standard text-to-image benchmarks, Diffusion-SDPO delivers consistent gains over preference-learning baselines on automated preference, aesthetic, and prompt alignment metrics. Code is publicly available at this https URL.
扩散-SDPO:扩散模型的安全直接偏好优化 / Diffusion-SDPO: Safeguarded Direct Preference Optimization for Diffusion Models
本文提出了一种名为Diffusion-SDPO的新方法,通过自适应调整优化过程中的梯度更新,解决了现有扩散模型在偏好学习时可能导致图像质量下降的问题,从而在保持简单高效的同时,显著提升了生成图像与人类偏好的对齐效果。