是什么造就了良好的多语言推理?用可度量特征解构推理轨迹 / What Makes Good Multilingual Reasoning? Disentangling Reasoning Traces with Measurable Features
1️⃣ 一句话总结
这篇论文挑战了“让所有语言的推理都模仿英语推理就能提升多语言模型性能”的假设,通过定义一套可度量的推理特征并进行分析,发现不同语言的有效推理模式存在显著差异,因此需要设计适应语言特定模式的评估标准和奖励机制。
Large Reasoning Models (LRMs) still exhibit large performance gaps between English and other languages, yet much current work assumes these gaps can be closed simply by making reasoning in every language resemble English reasoning. This work challenges this assumption by asking instead: what actually characterizes effective reasoning in multilingual settings, and to what extent do English-derived reasoning features genuinely help in other languages? We first define a suite of measurable reasoning features spanning multilingual alignment, reasoning step, and reasoning flow aspects of reasoning traces, and use logistic regression to quantify how each feature associates with final answer accuracy. We further train sparse autoencoders over multilingual traces to automatically discover latent reasoning concepts that instantiate or extend these features. Finally, we use the features as test-time selection policies to examine whether they can steer models toward stronger multilingual reasoning. Across two mathematical reasoning benchmarks, four LRMs, and 10 languages, we find that most features are positively associated with accuracy, but the strength of association varies considerably across languages and can even reverse in some. Our findings challenge English-centric reward designs and point toward adaptive objectives that accommodate language-specific reasoning patterns, with concrete implications for multilingual benchmark and reward design.
是什么造就了良好的多语言推理?用可度量特征解构推理轨迹 / What Makes Good Multilingual Reasoning? Disentangling Reasoning Traces with Measurable Features
这篇论文挑战了“让所有语言的推理都模仿英语推理就能提升多语言模型性能”的假设,通过定义一套可度量的推理特征并进行分析,发现不同语言的有效推理模式存在显著差异,因此需要设计适应语言特定模式的评估标准和奖励机制。
源自 arXiv: 2604.04720