菜单

关于 🐙 GitHub
arXiv 提交日期: 2025-12-18
📄 Abstract - INTELLECT-3: Technical Report

We present INTELLECT-3, a 106B-parameter Mixture-of-Experts model (12B active) trained with large-scale reinforcement learning on our end-to-end RL infrastructure stack. INTELLECT-3 achieves state of the art performance for its size across math, code, science and reasoning benchmarks, outperforming many larger frontier models. We open-source the model together with the full infrastructure stack used to create it, including RL frameworks, complete recipe, and a wide collection of environments, built with the verifiers library, for training and evaluation from our Environments Hub community platform. Built for this effort, we introduce prime-rl, an open framework for large-scale asynchronous reinforcement learning, which scales seamlessly from a single node to thousands of GPUs, and is tailored for agentic RL with first-class support for multi-turn interactions and tool use. Using this stack, we run both SFT and RL training on top of the GLM-4.5-Air-Base model, scaling RL training up to 512 H200s with high training efficiency.

顶级标签: llm model training systems
详细标签: mixture-of-experts reinforcement learning open-source infrastructure large-scale training benchmark evaluation 或 搜索:

INTELLECT-3:技术报告 / INTELLECT-3: Technical Report


1️⃣ 一句话总结

这篇论文介绍了INTELLECT-3,一个拥有1060亿参数的高效混合专家模型,它通过大规模强化学习训练,在数学、代码、科学和推理等多个基准测试中取得了同规模模型的最佳性能,并且作者开源了完整的模型及配套的训练基础设施。

源自 arXiv: 2512.16144