菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-03-09
📄 Abstract - Information Maximization for Long-Tailed Semi-Supervised Domain Generalization

Semi-supervised domain generalization (SSDG) has recently emerged as an appealing alternative to tackle domain generalization when labeled data is scarce but unlabeled samples across domains are abundant. In this work, we identify an important limitation that hampers the deployment of state-of-the-art methods on more challenging but practical scenarios. In particular, state-of-the-art SSDG severely suffers in the presence of long-tailed class distributions, an arguably common situation in real-world settings. To alleviate this limitation, we propose IMaX, a simple yet effective objective based on the well-known InfoMax principle adapted to the SSDG scenario, where the Mutual Information (MI) between the learned features and latent labels is maximized, constrained by the supervision from the labeled samples. Our formulation integrates an {\alpha}-entropic objective, which mitigates the class-balance bias encoded in the standard marginal entropy term of the MI, thereby better handling arbitrary class distributions. IMaX can be seamlessly plugged into recent state-of-the-art SSDG, consistently enhancing their performance, as demonstrated empirically across two different image modalities.

顶级标签: machine learning model training data
详细标签: semi-supervised learning domain generalization long-tailed distribution mutual information entropic regularization 或 搜索:

面向长尾分布的半监督领域泛化的信息最大化方法 / Information Maximization for Long-Tailed Semi-Supervised Domain Generalization


1️⃣ 一句话总结

本文提出了一种名为IMaX的新方法,通过改进信息最大化原则,有效解决了在半监督领域泛化任务中,当数据类别分布不均衡(长尾分布)时现有方法性能严重下降的问题,并能灵活地与现有先进模型结合使用。

源自 arXiv: 2603.08434