菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-02-10
📄 Abstract - MEVER: Multi-Modal and Explainable Claim Verification with Graph-based Evidence Retrieval

Verifying the truthfulness of claims usually requires joint multi-modal reasoning over both textual and visual evidence, such as analyzing both textual caption and chart image for claim verification. In addition, to make the reasoning process transparent, a textual explanation is necessary to justify the verification result. However, most claim verification works mainly focus on the reasoning over textual evidence only or ignore the explainability, resulting in inaccurate and unconvincing verification. To address this problem, we propose a novel model that jointly achieves evidence retrieval, multi-modal claim verification, and explanation generation. For evidence retrieval, we construct a two-layer multi-modal graph for claims and evidence, where we design image-to-text and text-to-image reasoning for multi-modal retrieval. For claim verification, we propose token- and evidence-level fusion to integrate claim and evidence embeddings for multi-modal verification. For explanation generation, we introduce multi-modal Fusion-in-Decoder for explainability. Finally, since almost all the datasets are in general domain, we create a scientific dataset, AIChartClaim, in AI domain to complement claim verification community. Experiments show the strength of our model.

顶级标签: multi-modal natural language processing model evaluation
详细标签: claim verification evidence retrieval explainable ai graph-based reasoning scientific dataset 或 搜索:

MEVER:基于图证据检索的多模态可解释声明验证 / MEVER: Multi-Modal and Explainable Claim Verification with Graph-based Evidence Retrieval


1️⃣ 一句话总结

这篇论文提出了一种名为MEVER的新模型,它能够同时从文本和图像中检索证据、验证声明的真伪,并生成解释性的文字说明,从而让AI的验证过程更准确、更透明,特别是在处理科学图表等复杂信息时效果显著。

源自 arXiv: 2602.10023