数学推理中的策略可执行性:利用人-模型差异实现有效引导 / Strategy Executability in Mathematical Reasoning: Leveraging Human-Model Differences for Effective Guidance
1️⃣ 一句话总结
这篇论文发现,在数学推理中,人类和AI模型各自擅长的解题策略不同,直接照搬成功案例中的策略作为引导往往效果不稳定,因此提出了一种能自动选择并组合高‘可执行性’策略的新方法,显著提升了模型的解题准确率。
Example-based guidance is widely used to improve mathematical reasoning at inference time, yet its effectiveness is highly unstable across problems and models-even when the guidance is correct and problem-relevant. We show that this instability arises from a previously underexplored gap between strategy usage-whether a reasoning strategy appears in successful solutions-and strategy executability-whether the strategy remains effective when instantiated as guidance for a target model. Through a controlled analysis of paired human-written and model-generated solutions, we identify a systematic dissociation between usage and executability: human- and model-derived strategies differ in structured, domain-dependent ways, leading to complementary strengths and consistent source-dependent reversals under guidance. Building on this diagnosis, we propose Selective Strategy Retrieval (SSR), a test-time framework that explicitly models executability by selectively retrieving and combining strategies using empirical, multi-route, source-aware signals. Across multiple mathematical reasoning benchmarks, SSR yields reliable and consistent improvements over direct solving, in-context learning, and single-source guidance, improving accuracy by up to $+13$ points on AIME25 and $+5$ points on Apex for compact reasoning models. Code and benchmark are publicly available at: this https URL.
数学推理中的策略可执行性:利用人-模型差异实现有效引导 / Strategy Executability in Mathematical Reasoning: Leveraging Human-Model Differences for Effective Guidance
这篇论文发现,在数学推理中,人类和AI模型各自擅长的解题策略不同,直接照搬成功案例中的策略作为引导往往效果不稳定,因此提出了一种能自动选择并组合高‘可执行性’策略的新方法,显著提升了模型的解题准确率。
源自 arXiv: 2602.22583