菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-04-06
📄 Abstract - HI-MoE: Hierarchical Instance-Conditioned Mixture-of-Experts for Object Detection

Mixture-of-Experts (MoE) architectures enable conditional computation by activating only a subset of model parameters for each input. Although sparse routing has been highly effective in language models and has also shown promise in vision, most vision MoE methods operate at the image or patch level. This granularity is poorly aligned with object detection, where the fundamental unit of reasoning is an object query corresponding to a candidate instance. We propose Hierarchical Instance-Conditioned Mixture-of-Experts (HI-MoE), a DETR-style detection architecture that performs routing in two stages: a lightweight scene router first selects a scene-consistent expert subset, and an instance router then assigns each object query to a small number of experts within that subset. This design aims to preserve sparse computation while better matching the heterogeneous, instance-centric structure of detection. In the current draft, experiments are concentrated on COCO with preliminary specialization analysis on LVIS. Under these settings, HI-MoE improves over a dense DINO baseline and over simpler token-level or instance-only routing variants, with especially strong gains on small objects. We also provide an initial visualization of expert specialization patterns. We present the method, ablations, and current limitations in a form intended to support further experimental validation.

顶级标签: computer vision model training machine learning
详细标签: object detection mixture-of-experts sparse routing conditional computation detr 或 搜索:

HI-MoE:用于目标检测的分层实例条件专家混合模型 / HI-MoE: Hierarchical Instance-Conditioned Mixture-of-Experts for Object Detection


1️⃣ 一句话总结

这篇论文提出了一种名为HI-MoE的新型目标检测架构,它通过一个两阶段的路由机制(先选场景专家,再为每个候选目标分配专家)来高效激活模型的不同部分,从而在保持计算稀疏性的同时,显著提升了对小物体等目标的检测性能。

源自 arXiv: 2604.04908