MiM-DiT:基于扩散Transformer的双层专家混合网络用于一体化图像修复 / MiM-DiT: MoE in MoE with Diffusion Transformers for All-in-One Image Restoration
1️⃣ 一句话总结
这篇论文提出了一个创新的图像修复模型,它通过一个‘专家中的专家’双层结构,让一个模型能像多个专家一样,智能地处理雾霾、模糊、噪声、低光等多种不同的图像退化问题。
All-in-one image restoration is challenging because different degradation types, such as haze, blur, noise, and low-light, impose diverse requirements on restoration strategies, making it difficult for a single model to handle them effectively. In this paper, we propose a unified image restoration framework that integrates a dual-level Mixture-of-Experts (MoE) architecture with a pretrained diffusion model. The framework operates at two levels: the Inter-MoE layer adaptively combines expert groups to handle major degradation types, while the Intra-MoE layer further selects specialized sub-experts to address fine-grained variations within each type. This design enables the model to achieve coarse-grained adaptation across diverse degradation categories while performing fine-grained modulation for specific intra-class variations, ensuring both high specialization in handling complex, real-world corruptions. Extensive experiments demonstrate that the proposed method performs favorably against the state-of-the-art approaches on multiple image restoration task.
MiM-DiT:基于扩散Transformer的双层专家混合网络用于一体化图像修复 / MiM-DiT: MoE in MoE with Diffusion Transformers for All-in-One Image Restoration
这篇论文提出了一个创新的图像修复模型,它通过一个‘专家中的专家’双层结构,让一个模型能像多个专家一样,智能地处理雾霾、模糊、噪声、低光等多种不同的图像退化问题。
源自 arXiv: 2603.02710