超越实例级自监督:三维多模态医学影像中的拓扑一致性学习 / Beyond Instance-Level Self-Supervision in 3D Multi-Modal Medical Imaging
1️⃣ 一句话总结
本文提出一种利用不同患者间器官空间关系相似性(拓扑一致性)作为监督信号的新颖自监督预训练方法,通过跨模态与跨样本的局部结构对齐,在多种三维医学影像任务中显著提升了分割、分类精度及对缺失模态的鲁棒性。
Self-supervised pre-training methods in medical imaging typically treat each individual as an isolated instance, learning representations through augmentation-based objectives or masked reconstruction. They often do not adequately capitalize on a key characteristic of physiological features: anatomical structures maintain consistent spatial relationships across individuals (instances), such as the thalamus being medial to the basal ganglia, regardless of variations in brain size, shape, or pathology. We propose leveraging this cross-instance topological consistency as a supervisory signal. The challenge arises from the inherent variability in medical imaging, which can differ significantly across instances and modalities. To tackle this, we focus on two alignment regimes. (i) Intra-instance: with pixel-level correspondences available, a cross-modal triplet objective explicitly preserves local neighborhood topology. (ii) Inter-instance: without such supervision, we derive pseudo-correspondences to control partial neighborhood alignment and prevent topology collapse across modalities. We validate our approach across 7 downstream multi-modal tasks, achieving average improvements of 1.1% and 5.94% in segmentation and classification tasks, respectively, and demonstrating significantly better robustness when modalities are missing at test time.
超越实例级自监督:三维多模态医学影像中的拓扑一致性学习 / Beyond Instance-Level Self-Supervision in 3D Multi-Modal Medical Imaging
本文提出一种利用不同患者间器官空间关系相似性(拓扑一致性)作为监督信号的新颖自监督预训练方法,通过跨模态与跨样本的局部结构对齐,在多种三维医学影像任务中显著提升了分割、分类精度及对缺失模态的鲁棒性。
源自 arXiv: 2605.14654