从词语到世界:大型语言模型能成为隐式的基于文本的世界模型吗? / From Word to World: Can Large Language Models be Implicit Text-based World Models?
1️⃣ 一句话总结
这篇论文通过一个三层评估框架,在五个文本环境中研究发现,经过充分训练的大型语言模型能够作为有效的世界模型来预测环境状态,从而通过多种方式提升智能体的学习效率,但其效果好坏关键取决于训练数据的行为覆盖度和环境本身的复杂程度。
Agentic reinforcement learning increasingly relies on experience-driven scaling, yet real-world environments remain non-adaptive, limited in coverage, and difficult to scale. World models offer a potential way to improve learning efficiency through simulated experience, but it remains unclear whether large language models can reliably serve this role and under what conditions they meaningfully benefit agents. We study these questions in text-based environments, which provide a controlled setting to reinterpret language modeling as next-state prediction under interaction. We introduce a three-level framework for evaluating LLM-based world models: (i) fidelity and consistency, (ii) scalability and robustness, and (iii) agent utility. Across five representative environments, we find that sufficiently trained world models maintain coherent latent state, scale predictably with data and model size, and improve agent performance via action verification, synthetic trajectory generation, and warm-starting reinforcement learning. Meanwhile, these gains depend critically on behavioral coverage and environment complexity, delineating clear boundry on when world modeling effectively supports agent learning.
从词语到世界:大型语言模型能成为隐式的基于文本的世界模型吗? / From Word to World: Can Large Language Models be Implicit Text-based World Models?
这篇论文通过一个三层评估框架,在五个文本环境中研究发现,经过充分训练的大型语言模型能够作为有效的世界模型来预测环境状态,从而通过多种方式提升智能体的学习效率,但其效果好坏关键取决于训练数据的行为覆盖度和环境本身的复杂程度。
源自 arXiv: 2512.18832