评估大型视觉语言模型在外科手术工具检测中的应用 / Evaluating Large Vision-language Models for Surgical Tool Detection
1️⃣ 一句话总结
这篇论文通过评估几种先进的大型视觉语言模型在手术工具检测任务上的表现,发现Qwen2.5模型在零样本学习和微调后均展现出优异的检测与识别能力,为开发更全面的手术AI辅助系统提供了有力支持。
Surgery is a highly complex process, and artificial intelligence has emerged as a transformative force in supporting surgical guidance and decision-making. However, the unimodal nature of most current AI systems limits their ability to achieve a holistic understanding of surgical workflows. This highlights the need for general-purpose surgical AI systems capable of comprehensively modeling the interrelated components of surgical scenes. Recent advances in large vision-language models that integrate multimodal data processing offer strong potential for modeling surgical tasks and providing human-like scene reasoning and understanding. Despite their promise, systematic investigations of VLMs in surgical applications remain limited. In this study, we evaluate the effectiveness of large VLMs for the fundamental surgical vision task of detecting surgical tools. Specifically, we investigate three state-of-the-art VLMs, Qwen2.5, LLaVA1.5, and InternVL3.5, on the GraSP robotic surgery dataset under both zero-shot and parameter-efficient LoRA fine-tuning settings. Our results demonstrate that Qwen2.5 consistently achieves superior detection performance in both configurations among the evaluated VLMs. Furthermore, compared with the open-set detection baseline Grounding DINO, Qwen2.5 exhibits stronger zero-shot generalization and comparable fine-tuned performance. Notably, Qwen2.5 shows superior instrument recognition, while Grounding DINO demonstrates stronger localization.
评估大型视觉语言模型在外科手术工具检测中的应用 / Evaluating Large Vision-language Models for Surgical Tool Detection
这篇论文通过评估几种先进的大型视觉语言模型在手术工具检测任务上的表现,发现Qwen2.5模型在零样本学习和微调后均展现出优异的检测与识别能力,为开发更全面的手术AI辅助系统提供了有力支持。
源自 arXiv: 2601.16895