菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-03-11
📄 Abstract - VERI-DPO: Evidence-Aware Alignment for Clinical Summarization via Claim Verification and Direct Preference Optimization

Brief Hospital Course (BHC) narratives must be clinically useful yet faithful to fragmented EHR evidence. LLM-based clinical summarizers still introduce unsupported statements, and alignment can encourage omissions ("say-less" degeneration). We introduce VERI-DPO, which uses claim verification to mine preferences and distill them into the summarizer with Direct Preference Optimization (DPO). On MIMIC-III-Ext-VeriFact-BHC (100 ICU patients; patient-level splits), we train a retrieval-augmented verifier to label claim-evidence pairs as Supported, Not Supported, or Not Addressed via a single-token format. The verifier scores sentence-level claims from sampled BHC candidates and aggregates margins into a coverage-aware utility to mine length-controlled, contradiction-anchored preference pairs. On held-out patients, verifier-mined preferences separate candidates by contradiction density, and VERI-DPO reduces Not Supported claim rates from 10.7% to 1.9% (local verifier judge) and from 11.6% to 6.4% (GPT-4o judge), while improving validity from 76.7% to 82.5% and maintaining informative length.

顶级标签: medical llm natural language processing
详细标签: clinical summarization faithfulness direct preference optimization claim verification evidence alignment 或 搜索:

VERI-DPO:通过声明验证与直接偏好优化实现证据感知的临床摘要对齐 / VERI-DPO: Evidence-Aware Alignment for Clinical Summarization via Claim Verification and Direct Preference Optimization


1️⃣ 一句话总结

这篇论文提出了一种名为VERI-DPO的新方法,它通过一个验证器来检查临床摘要中的陈述是否有医疗记录支持,并利用这些检查结果来训练摘要模型,从而在保持信息量的同时,显著减少了摘要中无依据或错误的陈述。

源自 arXiv: 2603.10494