菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-04-09
📄 Abstract - QaRL: Rollout-Aligned Quantization-Aware RL for Fast and Stable Training under Training--Inference Mismatch

Large language model (LLM) reinforcement learning (RL) pipelines are often bottlenecked by rollout generation, making end-to-end training slow. Recent work mitigates this by running rollouts with quantization to accelerate decoding, which is the most expensive stage of the RL loop. However, these setups destabilize optimization by amplifying the training-inference gap: rollouts are operated at low precision, while learning updates are computed at full precision. To address this challenge, we propose QaRL (Rollout Alignment Quantization-Aware RL), which aligns training-side forward with the quantized rollout to minimize mismatch. We further identify a failure mode in quantized rollouts: long-form responses tend to produce repetitive, garbled tokens (error tokens). To mitigate these problems, we introduce TBPO (Trust-Band Policy Optimization), a sequence-level objective with dual clipping for negative samples, aimed at keeping updates within the trust region. On Qwen3-30B-A3B MoE for math problems, QaRL outperforms quantized-rollout training by +5.5 while improving stability and preserving low-bit throughput benefits.

顶级标签: llm reinforcement learning model training
详细标签: quantization training-inference mismatch rollout generation policy optimization llm efficiency 或 搜索:

QaRL:一种面向快速稳定训练、解决训练-推理不匹配问题的量化感知强化学习方法 / QaRL: Rollout-Aligned Quantization-Aware RL for Fast and Stable Training under Training--Inference Mismatch


1️⃣ 一句话总结

这篇论文提出了一种名为QaRL的新方法,它通过让训练过程与低精度推理过程保持一致,并引入新的优化目标来防止生成错误文本,从而在利用量化技术加速大语言模型强化学习训练的同时,有效解决了由此引发的训练不稳定和性能下降问题。

源自 arXiv: 2604.07853