菜单

关于 🐙 GitHub
arXiv 提交日期: 2025-12-18
📄 Abstract - Turn-PPO: Turn-Level Advantage Estimation with PPO for Improved Multi-Turn RL in Agentic LLMs

Reinforcement learning (RL) has re-emerged as a natural approach for training interactive LLM agents in real-world environments. However, directly applying the widely used Group Relative Policy Optimization (GRPO) algorithm to multi-turn tasks exposes notable limitations, particularly in scenarios requiring long-horizon reasoning. To address these challenges, we investigate more stable and effective advantage estimation strategies, especially for multi-turn settings. We first explore Proximal Policy Optimization (PPO) as an alternative and find it to be more robust than GRPO. To further enhance PPO in multi-turn scenarios, we introduce turn-PPO, a variant that operates on a turn-level MDP formulation, as opposed to the commonly used token-level MDP. Our results on the WebShop and Sokoban datasets demonstrate the effectiveness of turn-PPO, both with and without long reasoning components.

顶级标签: llm agents reinforcement learning
详细标签: policy optimization multi-turn reasoning advantage estimation ppo interactive agents 或 搜索:

Turn-PPO:基于回合级优势估计与PPO的改进多轮强化学习,用于提升智能大语言模型 / Turn-PPO: Turn-Level Advantage Estimation with PPO for Improved Multi-Turn RL in Agentic LLMs


1️⃣ 一句话总结

这篇论文提出了一种名为Turn-PPO的新强化学习方法,通过将优化过程从传统的“词级”提升到“回合级”,有效解决了智能大语言模型在需要多轮交互和长远规划任务中训练不稳定的问题,从而提升了模型的整体表现。

源自 arXiv: 2512.17008