菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-05-06
📄 Abstract - MRI-Eval: A Tiered Benchmark for Evaluating LLM Performance on MRI Physics and GE Scanner Operations Knowledge

Background: Existing MRI LLM benchmarks rely mainly on review-book multiple-choice questions, where top proprietary models already score highly, limiting discrimination. No systematic benchmark has evaluated vendor-specific scanner operational knowledge central to research MRI practice. Purpose: We developed MRI-Eval, a tiered benchmark for relative model comparison on MRI physics and GE scanner operations knowledge using primary multiple-choice questions (MCQ), with stem-only and primed diagnostic conditions as complementary analyses. Methods: MRI-Eval includes 1365 scored items across nine categories and three difficulty tiers from textbooks, GE scanner manuals, programming course materials, and expert-generated questions. Five model families were evaluated (GPT-5.4, Claude Opus 4.6, Claude Sonnet 4.6, Gemini 2.5 Pro, Llama 3.3 70B). MCQ was primary; stem-only removed options and used an independent LLM judge; primed stem-only tested responses to incorrect user claims. Results: Overall MCQ accuracy was 93.2% to 97.1%. GE scanner operations was the lowest category for every model (88.2% to 94.6%). In stem-only, frontier-model accuracy fell to 58.4% to 61.1%, and Llama 3.3 70B fell to 37.1%; GE scanner operations stem-only accuracy was 13.8% to 29.8%. Conclusion: High MCQ performance can mask weak free-text recall, especially for vendor-specific operational knowledge. MRI-Eval is most informative as a relative comparison benchmark rather than an absolute competency measure and supports caution in using raw LLM outputs for GE-specific protocol guidance.

顶级标签: llm medical benchmark
详细标签: mri physics ge scanner operations multiple-choice questions free-text recall 或 搜索:

MRI-Eval:用于评估大语言模型在磁共振物理和GE扫描仪操作知识上的分层基准 / MRI-Eval: A Tiered Benchmark for Evaluating LLM Performance on MRI Physics and GE Scanner Operations Knowledge


1️⃣ 一句话总结

该研究开发了一个包含三个难度等级、1365道题目的分层基准测试MRI-Eval,用于评估大语言模型在MRI物理和GE扫描仪操作知识上的表现,结果发现虽然模型在选择题上准确率很高(超过93%),但在无选项自由回答和面对错误用户假设时表现大幅下降,特别是对厂商特定的操作知识理解薄弱,因此该基准更适合用于模型间的相对比较而非绝对能力衡量。

源自 arXiv: 2605.05175