菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-04-22
📄 Abstract - Weighting What Matters: Boosting Sample Efficiency in Medical Report Generation via Token Reweighting

Training vision-language models (VLMs) for medical report generation is often hindered by the scarcity of high-quality annotated data. This work evaluates the use of a weighted loss function to improve data efficiency. Compared to standard cross-entropy loss, which treats all token prediction errors equally, the reweighted loss shifts the focus to semantically salient tokens with outsized clinical importance. In experiments on ophthalmological report generation, we show that this simple method improves efficiency across multiple data scales, achieving similar report quality with up to ten times less training data.

顶级标签: medical multi-modal model training
详细标签: report generation vision-language model token reweighting data efficiency loss function 或 搜索:

权衡重要内容:通过令牌重加权提升医学报告生成的样本效率 / Weighting What Matters: Boosting Sample Efficiency in Medical Report Generation via Token Reweighting


1️⃣ 一句话总结

本文提出了一种简单的加权损失函数方法,在医学报告生成中通过给关键临床词汇更高权重,使模型仅用十分之一的训练数据就能达到与标准方法相当的报告质量,大幅提升了数据利用效率。

源自 arXiv: 2604.21082