大型语言模型在教学中会进行心理推断吗? / Do Large Language Models Mentalize When They Teach?
1️⃣ 一句话总结
这篇论文通过模拟教学实验发现,大多数大型语言模型在决定教学内容时,其行为模式与人类相似,能够像最优贝叶斯教师一样推断学习者的知识状态,而不仅仅是依赖简单的启发式规则,但外部提示引导并不总能提升其教学决策质量。
How do LLMs decide what to teach next: by reasoning about a learner's knowledge, or by using simpler rules of thumb? We test this in a controlled task previously used to study human teaching strategies. On each trial, a teacher LLM sees a hypothetical learner's trajectory through a reward-annotated directed graph and must reveal a single edge so the learner would choose a better path if they replanned. We run a range of LLMs as simulated teachers and fit their trial-by-trial choices with the same cognitive models used for humans: a Bayes-Optimal teacher that infers which transitions the learner is missing (inverse planning), weaker Bayesian variants, heuristic baselines (e.g., reward based), and non-mentalizing utility models. In a baseline experiment matched to the stimuli presented to human subjects, most LLMs perform well, show little change in strategy over trials, and their graph-by-graph performance is similar to that of humans. Model comparison (BIC) shows that Bayes-Optimal teaching best explains most models' choices. When given a scaffolding intervention, models follow auxiliary inference- or reward-focused prompts, but these scaffolds do not reliably improve later teaching on heuristic-incongruent test graphs and can sometimes reduce performance. Overall, cognitive model fits provide insight into LLM tutoring policies and show that prompt compliance does not guarantee better teaching decisions.
大型语言模型在教学中会进行心理推断吗? / Do Large Language Models Mentalize When They Teach?
这篇论文通过模拟教学实验发现,大多数大型语言模型在决定教学内容时,其行为模式与人类相似,能够像最优贝叶斯教师一样推断学习者的知识状态,而不仅仅是依赖简单的启发式规则,但外部提示引导并不总能提升其教学决策质量。
源自 arXiv: 2604.01594