菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-03-03
📄 Abstract - From Solver to Tutor: Evaluating the Pedagogical Intelligence of LLMs with KMP-Bench

Large Language Models (LLMs) show significant potential in AI mathematical tutoring, yet current evaluations often rely on simplistic metrics or narrow pedagogical scenarios, failing to assess comprehensive, multi-turn teaching effectiveness. In this paper, we introduce KMP-Bench, a comprehensive K-8 Mathematical Pedagogical Benchmark designed to assess LLMs from two complementary perspectives. The first module, KMP-Dialogue, evaluates holistic pedagogical capabilities against six core principles (e.g., Challenge, Explanation, Feedback), leveraging a novel multi-turn dialogue dataset constructed by weaving together diverse pedagogical components. The second module, KMP-Skills, provides a granular assessment of foundational tutoring abilities, including multi-turn problem-solving, error detection and correction, and problem generation. Our evaluations on KMP-Bench reveal a key disparity: while leading LLMs excel at tasks with verifiable solutions, they struggle with the nuanced application of pedagogical principles. Additionally, we present KMP-Pile, a large-scale (150K) dialogue dataset. Models fine-tuned on KMP-Pile show substantial improvement on KMP-Bench, underscoring the value of pedagogically-rich training data for developing more effective AI math tutors.

顶级标签: llm benchmark model evaluation
详细标签: mathematical tutoring pedagogical evaluation multi-turn dialogue fine-tuning educational ai 或 搜索:

从解题者到辅导者:使用KMP-Bench评估大语言模型的教学智能 / From Solver to Tutor: Evaluating the Pedagogical Intelligence of LLMs with KMP-Bench


1️⃣ 一句话总结

这篇论文提出了一个名为KMP-Bench的综合性评估基准,专门用于测试大语言模型在K-8年级数学辅导中的教学能力,发现当前模型虽然擅长解题,但在遵循教学原则(如引导、解释、反馈)方面仍有不足,并证明使用高质量教学对话数据训练可以显著提升模型的辅导效果。

源自 arXiv: 2603.02775