MedKCO:通过知识驱动的认知编排进行医学视觉-语言预训练 / MedKCO: Medical Vision-Language Pretraining via Knowledge-Driven Cognitive Orchestration
1️⃣ 一句话总结
这篇论文提出了一种名为MedKCO的新方法,通过模仿人类由易到难的认知过程来改进医学AI模型的训练,使其在应对不同医疗任务时表现更稳定、更准确。
Medical vision-language pretraining (VLP) models have recently been investigated for their generalization to diverse downstream tasks. However, current medical VLP methods typically force the model to learn simple and complex concepts simultaneously. This anti-cognitive process leads to suboptimal feature representations, especially under distribution shift. To address this limitation, we propose a Knowledge-driven Cognitive Orchestration for Medical VLP (MedKCO) that involves both the ordering of the pretraining data and the learning objective of vision-language contrast. Specifically, we design a two level curriculum by incorporating diagnostic sensitivity and intra-class sample representativeness for the ordering of the pretraining data. Moreover, considering the inter-class similarity of medical images, we introduce a self-paced asymmetric contrastive loss to dynamically adjust the participation of the pretraining objective. We evaluate the proposed pretraining method on three medical imaging scenarios in multiple vision-language downstream tasks, and compare it with several curriculum learning methods. Extensive experiments show that our method significantly surpasses all baselines. this https URL.
MedKCO:通过知识驱动的认知编排进行医学视觉-语言预训练 / MedKCO: Medical Vision-Language Pretraining via Knowledge-Driven Cognitive Orchestration
这篇论文提出了一种名为MedKCO的新方法,通过模仿人类由易到难的认知过程来改进医学AI模型的训练,使其在应对不同医疗任务时表现更稳定、更准确。
源自 arXiv: 2603.09101