菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-03-17
📄 Abstract - IQuest-Coder-V1 Technical Report

In this report, we introduce the IQuest-Coder-V1 series-(7B/14B/40B/40B-Loop), a new family of code large language models (LLMs). Moving beyond static code representations, we propose the code-flow multi-stage training paradigm, which captures the dynamic evolution of software logic through different phases of the pipeline. Our models are developed through the evolutionary pipeline, starting with the initial pre-training consisting of code facts, repository, and completion data. Following that, we implement a specialized mid-training stage that integrates reasoning and agentic trajectories in 32k-context and repository-scale in 128k-context to forge deep logical foundations. The models are then finalized with post-training of specialized coding capabilities, which is bifurcated into two specialized paths: the thinking path (utilizing reasoning-driven RL) and the instruct path (optimized for general assistance). IQuest-Coder-V1 achieves state-of-the-art performance among competitive models across critical dimensions of code intelligence: agentic software engineering, competitive programming, and complex tool use. To address deployment constraints, the IQuest-Coder-V1-Loop variant introduces a recurrent mechanism designed to optimize the trade-off between model capacity and deployment footprint, offering an architecturally enhanced path for efficacy-efficiency trade-off. We believe the release of the IQuest-Coder-V1 series, including the complete white-box chain of checkpoints from pre-training bases to the final thinking and instruction models, will advance research in autonomous code intelligence and real-world agentic systems.

顶级标签: llm model training systems
详细标签: code llm multi-stage training software engineering agentic systems model efficiency 或 搜索:

IQuest-Coder-V1 技术报告 / IQuest-Coder-V1 Technical Report


1️⃣ 一句话总结

这篇论文介绍了一个名为IQuest-Coder-V1的新型代码大语言模型系列,它通过创新的‘代码流多阶段训练’方法,模拟软件逻辑的动态演变,从而在编程智能、竞争性编程和复杂工具使用等关键领域实现了顶尖性能,并提供了在模型能力与部署成本之间取得平衡的优化版本。

源自 arXiv: 2603.16733