菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-03-10
📄 Abstract - Common Sense vs. Morality: The Curious Case of Narrative Focus Bias in LLMs

Large Language Models (LLMs) are increasingly deployed across diverse real-world applications and user communities. As such, it is crucial that these models remain both morally grounded and knowledge-aware. In this work, we uncover a critical limitation of current LLMs -- their tendency to prioritize moral reasoning over commonsense understanding. To investigate this phenomenon, we introduce CoMoral, a novel benchmark dataset containing commonsense contradictions embedded within moral dilemmas. Through extensive evaluation of ten LLMs across different model sizes, we find that existing models consistently struggle to identify such contradictions without prior signal. Furthermore, we observe a pervasive narrative focus bias, wherein LLMs more readily detect commonsense contradictions when they are attributed to a secondary character rather than the primary (narrator) character. Our comprehensive analysis underscores the need for enhanced reasoning-aware training to improve the commonsense robustness of large language models.

顶级标签: llm model evaluation natural language processing
详细标签: moral reasoning commonsense reasoning benchmark bias detection robustness 或 搜索:

常识与道德:大语言模型中叙事焦点偏差的奇特案例 / Common Sense vs. Morality: The Curious Case of Narrative Focus Bias in LLMs


1️⃣ 一句话总结

这篇论文发现当前的大语言模型在处理道德困境时,存在一种‘叙事焦点偏差’,即模型会过度依赖道德推理而忽视常识矛盾,尤其当矛盾出现在故事主角身上时更难以察觉,这凸显了提升模型常识推理鲁棒性的必要性。

源自 arXiv: 2603.09434