菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-04-09
📄 Abstract - Mitigating Distribution Sharpening in Math RLVR via Distribution-Aligned Hint Synthesis and Backward Hint Annealing

Reinforcement learning with verifiable rewards (RLVR) can improve low-$k$ reasoning accuracy while narrowing solution coverage on challenging math questions, and pass@1 gains do not necessarily translate into better large-$k$ performance. Existing hint-based approaches can make challenging questions trainable, but they leave two issues underexplored: teacher-student distribution mismatch and the need to reduce hint exposure to match no-hint evaluation. We address these issues through two components. Distribution-Aligned Hint Synthesis (DAHS) constructs verified teacher hints conditioned on student-style responses. Backward Hint Annealing (BHA) anneals hint exposure across difficulty buckets and uses per-question hint dropout to preserve no-hint updates throughout RL training. We evaluate the method in math RLVR under the DAPO training framework across AIME24, AIME25, and AIME26 using $\texttt{Qwen3-1.7B-Base}$ and $\texttt{Llama-3.2-1B-Instruct}$. On $\texttt{Qwen3-1.7B-Base}$, our method improves both pass@1 and pass@2048 relative to DAPO across the three AIME benchmarks. On $\texttt{Llama-3.2-1B-Instruct}$, the gains are concentrated in the large-$k$ regime. These results suggest that, in math RLVR, hint scaffolding is effective when it restores learnable updates on challenging questions early in training and is then gradually removed before no-hint evaluation.

顶级标签: reinforcement learning llm model training
详细标签: math reasoning hint synthesis distribution alignment hint annealing verifiable rewards 或 搜索:

通过分布对齐提示合成与后向提示退火缓解数学RLVR中的分布锐化问题 / Mitigating Distribution Sharpening in Math RLVR via Distribution-Aligned Hint Synthesis and Backward Hint Annealing


1️⃣ 一句话总结

这篇论文提出了一种结合分布对齐提示合成与后向提示退火的新方法,旨在解决数学推理强化学习中提示教学与无提示评估之间的分布不匹配问题,从而在提升模型简单问题准确率的同时,也显著改善了其在复杂问题上的整体推理能力。

源自 arXiv: 2604.07747