菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-03-23
📄 Abstract - Learning to Trust: How Humans Mentally Recalibrate AI Confidence Signals

Productive human-AI collaboration requires appropriate reliance, yet contemporary AI systems are often miscalibrated, exhibiting systematic overconfidence or underconfidence. We investigate whether humans can learn to mentally recalibrate AI confidence signals through repeated experience. In a behavioral experiment (N = 200), participants predicted the AI's correctness across four AI calibration conditions: standard, overconfidence, underconfidence, and a counterintuitive "reverse confidence" mapping. Results demonstrate robust learning across all conditions, with participants significantly improving their accuracy, discrimination, and calibration alignment over 50 trials. We present a computational model utilizing a linear-in-log-odds (LLO) transformation and a Rescorla-Wagner learning rule to explain these dynamics. The model reveals that humans adapt by updating their baseline trust and confidence sensitivity, using asymmetric learning rates to prioritize the most informative errors. While humans can compensate for monotonic miscalibration, we identify a significant boundary in the reverse confidence scenario, where a substantial proportion of participants struggled to override initial inductive biases. These findings provide a mechanistic account of how humans adapt their trust in AI confidence signals through experience.

顶级标签: agents human-ai interaction model evaluation
详细标签: trust calibration human-ai collaboration confidence learning behavioral experiment computational modeling 或 搜索:

学会信任:人类如何通过心理重新校准AI置信度信号 / Learning to Trust: How Humans Mentally Recalibrate AI Confidence Signals


1️⃣ 一句话总结

这项研究发现,人类通过与AI的反复互动,能够学习并调整自己对AI给出的置信度信号的信任程度,从而更准确地判断AI的对错,但对于违反直觉的‘反向置信度’模式,许多人难以克服固有的思维定式。

源自 arXiv: 2603.22634