菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-01-20
📄 Abstract - Jet-RL: Enabling On-Policy FP8 Reinforcement Learning with Unified Training and Rollout Precision Flow

Reinforcement learning (RL) is essential for enhancing the complex reasoning capabilities of large language models (LLMs). However, existing RL training pipelines are computationally inefficient and resource-intensive, with the rollout phase accounting for over 70% of total training time. Quantized RL training, particularly using FP8 precision, offers a promising approach to mitigating this bottleneck. A commonly adopted strategy applies FP8 precision during rollout while retaining BF16 precision for training. In this work, we present the first comprehensive study of FP8 RL training and demonstrate that the widely used BF16-training + FP8-rollout strategy suffers from severe training instability and catastrophic accuracy collapse under long-horizon rollouts and challenging tasks. Our analysis shows that these failures stem from the off-policy nature of the approach, which introduces substantial numerical mismatch between training and inference. Motivated by these observations, we propose Jet-RL, an FP8 RL training framework that enables robust and stable RL optimization. The key idea is to adopt a unified FP8 precision flow for both training and rollout, thereby minimizing numerical discrepancies and eliminating the need for inefficient inter-step calibration. Extensive experiments validate the effectiveness of Jet-RL: our method achieves up to 33% speedup in the rollout phase, up to 41% speedup in the training phase, and a 16% end-to-end speedup over BF16 training, while maintaining stable convergence across all settings and incurring negligible accuracy degradation.

顶级标签: reinforcement learning model training systems
详细标签: fp8 quantization rl training efficiency numerical stability on-policy learning precision flow 或 搜索:

Jet-RL:通过统一的训练与执行精度流实现基于策略的FP8强化学习 / Jet-RL: Enabling On-Policy FP8 Reinforcement Learning with Unified Training and Rollout Precision Flow


1️⃣ 一句话总结

这篇论文提出了一种名为Jet-RL的新框架,它通过让强化学习的训练和执行阶段都使用统一的低精度(FP8)计算格式,解决了现有混合精度方法导致的训练不稳定和性能崩溃问题,从而在显著提升训练速度的同时保证了模型的稳定收敛和最终性能。

源自 arXiv: 2601.14243