T-pro 2.0:一个高效的俄语混合推理模型与实验平台 / T-pro 2.0: An Efficient Russian Hybrid-Reasoning Model and Playground
1️⃣ 一句话总结
这篇论文发布了一个名为T-pro 2.0的高效开源俄语大语言模型,它不仅支持直接回答和生成推理步骤,还通过优化技术降低了响应延迟,并配套发布了训练数据、评测基准和推理工具,旨在为构建和评估实用的俄语AI应用提供一个可访问的开放系统。
We introduce T-pro 2.0, an open-weight Russian LLM for hybrid reasoning and efficient inference. The model supports direct answering and reasoning-trace generation, using a Cyrillic-dense tokenizer and an adapted EAGLE speculative-decoding pipeline to reduce latency. To enable reproducible and extensible research, we release the model weights, the T-Wix 500k instruction corpus, the T-Math reasoning benchmark, and the EAGLE weights on Hugging Face. These resources allow users to study Russian-language reasoning and to extend or adapt both the model and the inference pipeline. A public web demo exposes reasoning and non-reasoning modes and illustrates the speedups achieved by our inference stack across domains. T-pro 2.0 thus serves as an accessible open system for building and evaluating efficient, practical Russian LLM applications.
T-pro 2.0:一个高效的俄语混合推理模型与实验平台 / T-pro 2.0: An Efficient Russian Hybrid-Reasoning Model and Playground
这篇论文发布了一个名为T-pro 2.0的高效开源俄语大语言模型,它不仅支持直接回答和生成推理步骤,还通过优化技术降低了响应延迟,并配套发布了训练数据、评测基准和推理工具,旨在为构建和评估实用的俄语AI应用提供一个可访问的开放系统。
源自 arXiv: 2512.10430