菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-03-09
📄 Abstract - CODA: Difficulty-Aware Compute Allocation for Adaptive Reasoning

The emergence of large reasoning models demonstrates that scaling inference-time compute significantly enhances performance on complex tasks. However, it often falls into another trap: overthinking simple problems, where repetitive rationales yield minimal accuracy gains at a disproportionately high cost. This motivates adaptive reasoning: dynamically aligning reasoning depth with instance difficulty. In this paper, we study adaptive reasoning from an optimality perspective, formalizing it as a utility maximization problem where tokens are allocated until the marginal accuracy gain falls below the incremental cost. Based on this, we propose CODA (Compute Allocation by Difficulty Awareness), a method that operationalizes this principle by allocating tokens via a policy-internal difficulty signal. Specifically, CODA estimates difficulty via group-based rollouts and maps it to two non-negative gates that modulate a length-dependent shaping term on top of the binary base reward. The easy-side gate penalizes verbosity on simple instances, whereas the hard-side gate encourages more deliberative rollouts on challenging ones. Across model scales and benchmarks, CODA achieves adaptive reasoning without external annotations or user-provided budgets: on easy tasks, CODA reduces token costs by over 60% while maintaining strong accuracy, whereas on hard tasks it incentivizes more deliberative rollouts to maximize performance.

顶级标签: llm model evaluation systems
详细标签: adaptive reasoning compute allocation difficulty estimation token efficiency utility maximization 或 搜索:

CODA:面向自适应推理的难度感知计算分配方法 / CODA: Difficulty-Aware Compute Allocation for Adaptive Reasoning


1️⃣ 一句话总结

这篇论文提出了一种名为CODA的智能方法,它能让大型推理模型根据问题的难易程度自动调整思考深度,从而在简单问题上节省大量计算资源,在复杂问题上则投入更多思考以提升性能。

源自 arXiv: 2603.08659