语言模型智能体中的时间、身份与意识 / Time, Identity and Consciousness in Language Model Agents
1️⃣ 一句话总结
这篇论文提出了一种评估语言模型智能体身份稳定性的新方法,通过分析其在时间跨度内的行为一致性,来区分它仅仅是‘说’得像一个稳定的自我,还是真正在决策时‘组织’得像一个稳定的自我。
Machine consciousness evaluations mostly see behavior. For language model agents that behavior is language and tool use. That lets an agent say the right things about itself even when the constraints that should make those statements matter are not jointly present at decision time. We apply Stack Theory's temporal gap to scaffold trajectories. This separates ingredient-wise occurrence within an evaluation window from co-instantiation at a single objective step. We then instantiate Stack Theory's Arpeggio and Chord postulates on grounded identity statements. This yields two persistence scores that can be computed from instrumented scaffold traces. We connect these scores to five operational identity metrics and map common scaffolds into an identity morphospace that exposes predictable tradeoffs. The result is a conservative toolkit for identity evaluation. It separates talking like a stable self from being organized like one.
语言模型智能体中的时间、身份与意识 / Time, Identity and Consciousness in Language Model Agents
这篇论文提出了一种评估语言模型智能体身份稳定性的新方法,通过分析其在时间跨度内的行为一致性,来区分它仅仅是‘说’得像一个稳定的自我,还是真正在决策时‘组织’得像一个稳定的自我。
源自 arXiv: 2603.09043