DirectEdit:基于流的图像编辑的逐步精确反演方法 / DirectEdit: Step-Level Accurate Inversion for Flow-Based Image Editing
1️⃣ 一句话总结
本文提出了一种无需额外训练的图像编辑方法DirectEdit,通过直接对齐前向路径而非修正反演路径,消除了传统方法中时间步不匹配导致的累积误差,从而在保持高保真度重建的同时实现高效的图像编辑。
With recent advancements in large-scale pre-trained text-to-image (T2I) models, training-free image editing methods have demonstrated remarkable success. Typically, these methods involve adding noise to a clean image via an inversion process, followed by separate denoising steps for the reconstruction and editing paths during the forward process. However, since the reconstruction path is approximated using noisy latents from mismatched timesteps, existing methods inevitably suffer from accumulated drift, which fundamentally limits reconstruction fidelity. To address this challenge, we systematically analyze the inversion process within the flow transformer and propose DirectEdit, a simple yet effective editing method that eliminates the inherent reconstruction error without introducing additional neural function evaluations (NFEs). Unlike most prior works that attempt to rectify the inversion path, DirectEdit focuses on directly aligning the forward paths, enabling precise reconstruction and reliable feature sharing. Furthermore, we introduce a preservation mechanism based on attention feature injection and multi-branch mask-guided noise blending, which effectively balances fidelity and editability. Extensive experiments across diverse scenarios demonstrate that DirectEdit achieves efficient and accurate image editing, delivering superior performance that outperforms state-of-the-art methods. Code and examples are available at this https URL.
DirectEdit:基于流的图像编辑的逐步精确反演方法 / DirectEdit: Step-Level Accurate Inversion for Flow-Based Image Editing
本文提出了一种无需额外训练的图像编辑方法DirectEdit,通过直接对齐前向路径而非修正反演路径,消除了传统方法中时间步不匹配导致的累积误差,从而在保持高保真度重建的同时实现高效的图像编辑。
源自 arXiv: 2605.02417