菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-02-10
📄 Abstract - Agent World Model: Infinity Synthetic Environments for Agentic Reinforcement Learning

Recent advances in large language model (LLM) have empowered autonomous agents to perform complex tasks that require multi-turn interactions with tools and environments. However, scaling such agent training is limited by the lack of diverse and reliable environments. In this paper, we propose Agent World Model (AWM), a fully synthetic environment generation pipeline. Using this pipeline, we scale to 1,000 environments covering everyday scenarios, in which agents can interact with rich toolsets (35 tools per environment on average) and obtain high-quality observations. Notably, these environments are code-driven and backed by databases, providing more reliable and consistent state transitions than environments simulated by LLMs. Moreover, they enable more efficient agent interaction compared with collecting trajectories from realistic environments. To demonstrate the effectiveness of this resource, we perform large-scale reinforcement learning for multi-turn tool-use agents. Thanks to the fully executable environments and accessible database states, we can also design reliable reward functions. Experiments on three benchmarks show that training exclusively in synthetic environments, rather than benchmark-specific ones, yields strong out-of-distribution generalization. The code is available at this https URL.

顶级标签: agents reinforcement learning systems
详细标签: synthetic environments world model tool-use agents generalization multi-turn interaction 或 搜索:

智能体世界模型:用于智能体强化学习的无限合成环境 / Agent World Model: Infinity Synthetic Environments for Agentic Reinforcement Learning


1️⃣ 一句话总结

这篇论文提出了一个名为‘智能体世界模型’的自动化合成环境生成方法,它能大规模创建多样、可靠且可执行的虚拟场景,用于高效训练AI智能体使用工具完成任务,并显著提升其在新环境中的泛化能力。

源自 arXiv: 2602.10090