差异至关重要:用于能力差距发现与修正的模型审计框架 / Differences That Matter: Auditing Models for Capability Gap Discovery and Rectification
1️⃣ 一句话总结
这篇论文提出了一个名为AuditDM的自动化审计框架,它通过生成能最大化不同模型之间分歧的挑战性问题和图像,来主动发现并修正多模态大语言模型的能力缺陷,从而在无需人工标注的情况下有效提升模型性能。
Conventional evaluation methods for multimodal LLMs (MLLMs) lack interpretability and are often insufficient to fully disclose significant capability gaps across models. To address this, we introduce AuditDM, an automated framework that actively discovers and rectifies MLLM failure modes by auditing their divergence. AuditDM fine-tunes an MLLM as an auditor via reinforcement learning to generate challenging questions and counterfactual images that maximize disagreement among target models. Once trained, the auditor uncovers diverse, interpretable exemplars that reveal model weaknesses and serve as annotation-free data for rectification. When applied to SoTA models like Gemma-3 and PaliGemma-2, AuditDM discovers more than 20 distinct failure types. Fine-tuning on these discoveries consistently improves all models across 16 benchmarks, and enables a 3B model to surpass its 28B counterpart. Our results suggest that as data scaling hits diminishing returns, targeted model auditing offers an effective path to model diagnosis and improvement.
差异至关重要:用于能力差距发现与修正的模型审计框架 / Differences That Matter: Auditing Models for Capability Gap Discovery and Rectification
这篇论文提出了一个名为AuditDM的自动化审计框架,它通过生成能最大化不同模型之间分歧的挑战性问题和图像,来主动发现并修正多模态大语言模型的能力缺陷,从而在无需人工标注的情况下有效提升模型性能。
源自 arXiv: 2512.16921