何时思考足够?基于充分性评估的高效推理早期退出方法 / When Is Thinking Enough? Early Exit via Sufficiency Assessment for Efficient Reasoning
1️⃣ 一句话总结
这篇论文提出了一种名为DTSR的新方法,它模仿人类的元认知能力,让大型推理模型能够自己判断思考过程是否已经充分,从而在合适的时机停止计算,有效解决了模型‘过度思考’导致的效率低下问题,在几乎不影响准确性的前提下显著缩短了推理长度。
Large reasoning models (LRMs) have achieved remarkable performance in complex reasoning tasks, driven by their powerful inference-time scaling capability. However, LRMs often suffer from overthinking, which results in substantial computational redundancy and significantly reduces efficiency. Early-exit methods aim to mitigate this issue by terminating reasoning once sufficient evidence has been generated, yet existing approaches mostly rely on handcrafted or empirical indicators that are unreliable and impractical. In this work, we introduce Dynamic Thought Sufficiency in Reasoning (DTSR), a novel framework for efficient reasoning that enables the model to dynamically assess the sufficiency of its chain-of-thought (CoT) and determine the optimal point for early exit. Inspired by human metacognition, DTSR operates in two stages: (1) Reflection Signal Monitoring, which identifies reflection signals as potential cues for early exit, and (2) Thought Sufficiency Check, which evaluates whether the current CoT is sufficient to derive the final answer. Experimental results on the Qwen3 models show that DTSR reduces reasoning length by 28.9%-34.9% with minimal performance loss, effectively mitigating overthinking. We further discuss overconfidence in LRMs and self-evaluation paradigms, providing valuable insights for early-exit reasoning.
何时思考足够?基于充分性评估的高效推理早期退出方法 / When Is Thinking Enough? Early Exit via Sufficiency Assessment for Efficient Reasoning
这篇论文提出了一种名为DTSR的新方法,它模仿人类的元认知能力,让大型推理模型能够自己判断思考过程是否已经充分,从而在合适的时机停止计算,有效解决了模型‘过度思考’导致的效率低下问题,在几乎不影响准确性的前提下显著缩短了推理长度。
源自 arXiv: 2604.06787