H2VLR:用于少样本异常检测的异质超图视觉语言推理 / H2VLR: Heterogeneous Hypergraph Vision-Language Reasoning for Few-Shot Anomaly Detection
1️⃣ 一句话总结
这篇论文提出了一种名为H2VLR的新方法,它通过构建一个结合图像区域和文本概念的异质超图,将少样本异常检测问题转化为对视觉与语义关系的高阶推理,从而超越了传统基于简单特征匹配的方法,在工业和医疗图像检测任务中取得了领先的性能。
As a classic vision task, anomaly detection has been widely applied in industrial inspection and medical imaging. In this task, data scarcity is often a frequently-faced issue. To solve it, the few-shot anomaly detection (FSAD) scheme is attracting increasing attention. In recent years, beyond traditional visual paradigm, Vision-Language Model (VLM) has been extensively explored to boost this field. However, in currently-existing VLM-based FSAD schemes, almost all perform anomaly inference only by pairwise feature matching, ignoring structural dependencies and global consistency. To further redound to FSAD via VLM, we propose a Heterogeneous Hypergraph Vision-Language Reasoning (H2VLR) framework. It reformulates the FSAD as a high-order inference problem of visual-semantic relations, by jointly modeling visual regions and semantic concepts in a unified hypergraph. Experimental comparisons verify the effectiveness and advantages of H2VLR. It could often achieve state-of-the-art (SOTA) performance on representative industrial and medical benchmarks. Our code will be released upon acceptance.
H2VLR:用于少样本异常检测的异质超图视觉语言推理 / H2VLR: Heterogeneous Hypergraph Vision-Language Reasoning for Few-Shot Anomaly Detection
这篇论文提出了一种名为H2VLR的新方法,它通过构建一个结合图像区域和文本概念的异质超图,将少样本异常检测问题转化为对视觉与语义关系的高阶推理,从而超越了传统基于简单特征匹配的方法,在工业和医疗图像检测任务中取得了领先的性能。
源自 arXiv: 2604.14507