菜单

关于 🐙 GitHub
arXiv 提交日期: 2025-12-02
📄 Abstract - From Imitation to Discrimination: Toward A Generalized Curriculum Advantage Mechanism Enhancing Cross-Domain Reasoning Tasks

Reinforcement learning has emerged as a paradigm for post-training large language models, boosting their reasoning capabilities. Such approaches compute an advantage value for each sample, reflecting better or worse performance than expected, thereby yielding both positive and negative signals for training. However, the indiscriminate mixing of the two signals in existing methods, especially from the early stages, may lead to ambiguous guidance and limited gains. To address this issue, we propose **CAPO** (**C**urriculum **A**dvantage **P**olicy **O**ptimization), an adaptive curriculum mechanism based on advantage signals. The proposed mechanism bootstraps imitation learning with positive-only advantage samples to establish robust foundations, and subsequently introduces negative signals to cultivate discriminative capabilities, thereby improving generalization across complex scenarios. Compatible with diverse optimization methods including GRPO, PPO, RLOO, and Reinforce++, our method consistently achieves stable and significant improvements in mathematical reasoning tasks, and further generalizes effectively to multimodal Graphical User Interface (GUI) reasoning scenarios, establishing itself as a versatile and robust optimization framework.

顶级标签: llm model training reinforcement learning
详细标签: policy optimization curriculum learning reasoning advantage function generalization 或 搜索:

从模仿到判别:一种增强跨领域推理任务的通用课程优势机制 / From Imitation to Discrimination: Toward A Generalized Curriculum Advantage Mechanism Enhancing Cross-Domain Reasoning Tasks


1️⃣ 一句话总结

这篇论文提出了一个名为CAPO的智能训练方法,它像老师教学生一样,先让大语言模型模仿好的例子打好基础,再逐步学习区分好坏,从而在数学和图形界面等多种复杂推理任务上取得更稳定、更出色的表现。


源自 arXiv: 2512.02580