大语言模型的智能体推理 / Agentic Reasoning for Large Language Models
1️⃣ 一句话总结
这篇论文提出了一种将大语言模型转变为能够自主规划、行动和学习的智能体的新范式,以解决其在开放动态环境中的推理难题,并系统梳理了从单智能体基础能力到多智能体协作的完整技术路线图。
Reasoning is a fundamental cognitive process underlying inference, problem-solving, and decision-making. While large language models (LLMs) demonstrate strong reasoning capabilities in closed-world settings, they struggle in open-ended and dynamic environments. Agentic reasoning marks a paradigm shift by reframing LLMs as autonomous agents that plan, act, and learn through continual interaction. In this survey, we organize agentic reasoning along three complementary dimensions. First, we characterize environmental dynamics through three layers: foundational agentic reasoning, which establishes core single-agent capabilities including planning, tool use, and search in stable environments; self-evolving agentic reasoning, which studies how agents refine these capabilities through feedback, memory, and adaptation; and collective multi-agent reasoning, which extends intelligence to collaborative settings involving coordination, knowledge sharing, and shared goals. Across these layers, we distinguish in-context reasoning, which scales test-time interaction through structured orchestration, from post-training reasoning, which optimizes behaviors via reinforcement learning and supervised fine-tuning. We further review representative agentic reasoning frameworks across real-world applications and benchmarks, including science, robotics, healthcare, autonomous research, and mathematics. This survey synthesizes agentic reasoning methods into a unified roadmap bridging thought and action, and outlines open challenges and future directions, including personalization, long-horizon interaction, world modeling, scalable multi-agent training, and governance for real-world deployment.
大语言模型的智能体推理 / Agentic Reasoning for Large Language Models
这篇论文提出了一种将大语言模型转变为能够自主规划、行动和学习的智能体的新范式,以解决其在开放动态环境中的推理难题,并系统梳理了从单智能体基础能力到多智能体协作的完整技术路线图。
源自 arXiv: 2601.12538