菜单

关于 🐙 GitHub
arXiv 提交日期: 2025-12-26
📄 Abstract - SpotEdit: Selective Region Editing in Diffusion Transformers

Diffusion Transformer models have significantly advanced image editing by encoding conditional images and integrating them into transformer layers. However, most edits involve modifying only small regions, while current methods uniformly process and denoise all tokens at every timestep, causing redundant computation and potentially degrading unchanged areas. This raises a fundamental question: Is it truly necessary to regenerate every region during editing? To address this, we propose SpotEdit, a training-free diffusion editing framework that selectively updates only the modified regions. SpotEdit comprises two key components: SpotSelector identifies stable regions via perceptual similarity and skips their computation by reusing conditional image features; SpotFusion adaptively blends these features with edited tokens through a dynamic fusion mechanism, preserving contextual coherence and editing quality. By reducing unnecessary computation and maintaining high fidelity in unmodified areas, SpotEdit achieves efficient and precise image editing.

顶级标签: computer vision model training aigc
详细标签: diffusion transformers selective editing image editing computational efficiency feature fusion 或 搜索:

SpotEdit:扩散变换器中的选择性区域编辑 / SpotEdit: Selective Region Editing in Diffusion Transformers


1️⃣ 一句话总结

这篇论文提出了一个名为SpotEdit的无训练图像编辑框架,它通过智能识别并跳过图像中未修改区域的冗余计算,只对需要编辑的部分进行更新,从而在保持高质量编辑效果的同时,大幅提升了编辑效率。

源自 arXiv: 2512.22323