菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-04-12
📄 Abstract - DiningBench: A Hierarchical Multi-view Benchmark for Perception and Reasoning in the Dietary Domain

Recent advancements in Vision-Language Models (VLMs) have revolutionized general visual understanding. However, their application in the food domain remains constrained by benchmarks that rely on coarse-grained categories, single-view imagery, and inaccurate metadata. To bridge this gap, we introduce DiningBench, a hierarchical, multi-view benchmark designed to evaluate VLMs across three levels of cognitive complexity: Fine-Grained Classification, Nutrition Estimation, and Visual Question Answering. Unlike previous datasets, DiningBench comprises 3,021 distinct dishes with an average of 5.27 images per entry, incorporating fine-grained "hard" negatives from identical menus and rigorous, verification-based nutritional data. We conduct an extensive evaluation of 29 state-of-the-art open-source and proprietary models. Our experiments reveal that while current VLMs excel at general reasoning, they struggle significantly with fine-grained visual discrimination and precise nutritional reasoning. Furthermore, we systematically investigate the impact of multi-view inputs and Chain-of-Thought reasoning, identifying five primary failure modes. DiningBench serves as a challenging testbed to drive the next generation of food-centric VLM research. All codes are released in this https URL.

顶级标签: multi-modal benchmark model evaluation
详细标签: vision-language models food domain fine-grained classification nutrition estimation visual question answering 或 搜索:

DiningBench:一个用于饮食领域感知与推理的分层次多视角基准 / DiningBench: A Hierarchical Multi-view Benchmark for Perception and Reasoning in the Dietary Domain


1️⃣ 一句话总结

这篇论文提出了一个名为DiningBench的新基准测试,它通过包含多角度图片和精细分类的饮食数据,来全面评估AI模型在识别菜品、估算营养和回答食物相关问题上的能力,发现现有模型在细节分辨和精确营养推理方面仍有明显不足。

源自 arXiv: 2604.10425