菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-01-27
📄 Abstract - Benchmarks Saturate When The Model Gets Smarter Than The Judge

Benchmarks are important tools to track progress in the development of Large Language Models (LLMs), yet inaccuracies in datasets and evaluation methods consistently undermine their effectiveness. Here, we present Omni-MATH-2, a manually revised version of the Omni-MATH dataset comprising a clean, exact-answer subset ($n{=}4181$) and a tagged, non-standard subset ($n{=}247$). Each problem was audited to ensure LaTeX compilability, solvability and verifiability, which involved adding missing figures or information, labeling problems requiring a proof, estimation or image, and removing clutter. This process significantly reduces dataset-induced noise, thereby providing a more precise assessment of model performance. The annotated dataset also allows us to evaluate judge-induced noise by comparing GPT-5 mini with the original Omni-Judge, revealing substantial discrepancies between judges on both the clean and tagged problem subsets. Expert annotations reveal that Omni-Judge is wrong in $96.4\%$ of the judge disagreements, indicating its inability to differentiate between models' abilities, even well before saturation of the benchmark occurs. As problems become more challenging, we find that increasingly competent judges become essential in order to prevent judge errors from masking genuine differences between models. Finally, neither judge identifies the present failure modes for the subset of tagged problems, demonstrating that dataset quality and judge reliability are both critical to develop accurate benchmarks of model performance.

顶级标签: llm benchmark model evaluation
详细标签: benchmark saturation evaluation noise dataset auditing judge reliability mathematical reasoning 或 搜索:

当模型比评估者更聪明时,基准测试会趋于饱和 / Benchmarks Saturate When The Model Gets Smarter Than The Judge


1️⃣ 一句话总结

这篇论文通过构建一个高质量、经过人工审核的数学数据集(Omni-MATH-2),揭示了当前大语言模型基准测试中的一个关键问题:当模型能力超过评估工具(Judge)的理解水平时,评估工具本身的错误会掩盖模型间的真实性能差异,导致基准测试过早失效。

源自 arXiv: 2601.19532