菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-03-03
📄 Abstract - MiM-DiT: MoE in MoE with Diffusion Transformers for All-in-One Image Restoration

All-in-one image restoration is challenging because different degradation types, such as haze, blur, noise, and low-light, impose diverse requirements on restoration strategies, making it difficult for a single model to handle them effectively. In this paper, we propose a unified image restoration framework that integrates a dual-level Mixture-of-Experts (MoE) architecture with a pretrained diffusion model. The framework operates at two levels: the Inter-MoE layer adaptively combines expert groups to handle major degradation types, while the Intra-MoE layer further selects specialized sub-experts to address fine-grained variations within each type. This design enables the model to achieve coarse-grained adaptation across diverse degradation categories while performing fine-grained modulation for specific intra-class variations, ensuring both high specialization in handling complex, real-world corruptions. Extensive experiments demonstrate that the proposed method performs favorably against the state-of-the-art approaches on multiple image restoration task.

顶级标签: computer vision model training multi-modal
详细标签: image restoration mixture of experts diffusion transformers all-in-one model multi-degradation 或 搜索:

MiM-DiT:基于扩散Transformer的双层专家混合网络用于一体化图像修复 / MiM-DiT: MoE in MoE with Diffusion Transformers for All-in-One Image Restoration


1️⃣ 一句话总结

这篇论文提出了一个创新的图像修复模型,它通过一个‘专家中的专家’双层结构,让一个模型能像多个专家一样,智能地处理雾霾、模糊、噪声、低光等多种不同的图像退化问题。

源自 arXiv: 2603.02710