菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-02-17
📄 Abstract - Effective and Robust Multimodal Medical Image Analysis

Multimodal Fusion Learning (MFL), leveraging disparate data from various imaging modalities (e.g., MRI, CT, SPECT), has shown great potential for addressing medical problems such as skin cancer and brain tumor prediction. However, existing MFL methods face three key limitations: a) they often specialize in specific modalities, and overlook effective shared complementary information across diverse modalities, hence limiting their generalizability for multi-disease analysis; b) they rely on computationally expensive models, restricting their applicability in resource-limited settings; and c) they lack robustness against adversarial attacks, compromising reliability in medical AI applications. To address these limitations, we propose a novel Multi-Attention Integration Learning (MAIL) network, incorporating two key components: a) an efficient residual learning attention block for capturing refined modality-specific multi-scale patterns and b) an efficient multimodal cross-attention module for learning enriched complementary shared representations across diverse modalities. Furthermore, to ensure adversarial robustness, we extend MAIL network to design Robust-MAIL by incorporating random projection filters and modulated attention noise. Extensive evaluations on 20 public datasets show that both MAIL and Robust-MAIL outperform existing methods, achieving performance gains of up to 9.34% while reducing computational costs by up to 78.3%. These results highlight the superiority of our approaches, ensuring more reliable predictions than top competitors. Code: this https URL.

顶级标签: medical multi-modal model training
详细标签: multimodal fusion medical imaging adversarial robustness attention mechanisms computational efficiency 或 搜索:

高效且鲁棒的多模态医学图像分析 / Effective and Robust Multimodal Medical Image Analysis


1️⃣ 一句话总结

本文提出了一种名为MAIL的新型多模态融合学习网络,它通过高效的注意力机制整合不同医学影像数据,不仅显著提升了疾病预测的准确性和泛化能力,还大幅降低了计算成本并增强了模型对抗攻击的鲁棒性。

源自 arXiv: 2602.15346