AtomEval:事实核查中对抗性主张的原子化评估 / AtomEval: Atomic Evaluation of Adversarial Claims in Fact Verification
1️⃣ 一句话总结
这篇论文提出了一个名为AtomEval的新评估框架,它通过将事实主张拆解成原子成分并检查其真实性是否被破坏,从而更可靠地评估对抗性改写对事实核查系统的攻击效果,研究发现更强的AI模型未必能生成更有效的对抗性主张。
Adversarial claim rewriting is widely used to test fact-checking systems, but standard metrics fail to capture truth-conditional consistency and often label semantically corrupted rewrites as successful. We introduce AtomEval, a validity-aware evaluation framework that decomposes claims into subject-relation-object-modifier (SROM) atoms and scores adversarial rewrites with Atomic Validity Scoring (AVS), enabling detection of factual corruption beyond surface similarity. Experiments on the FEVER dataset across representative attack strategies and LLM generators show that AtomEval provides more reliable evaluation signals in our experiments. Using AtomEval, we further analyze LLM-based adversarial generators and observe that stronger models do not necessarily produce more effective adversarial claims under validity-aware evaluation, highlighting previously overlooked limitations in current adversarial evaluation practices.
AtomEval:事实核查中对抗性主张的原子化评估 / AtomEval: Atomic Evaluation of Adversarial Claims in Fact Verification
这篇论文提出了一个名为AtomEval的新评估框架,它通过将事实主张拆解成原子成分并检查其真实性是否被破坏,从而更可靠地评估对抗性改写对事实核查系统的攻击效果,研究发现更强的AI模型未必能生成更有效的对抗性主张。
源自 arXiv: 2604.07967