菜单

关于 🐙 GitHub
arXiv 提交日期: 2025-12-11
📄 Abstract - T-pro 2.0: An Efficient Russian Hybrid-Reasoning Model and Playground

We introduce T-pro 2.0, an open-weight Russian LLM for hybrid reasoning and efficient inference. The model supports direct answering and reasoning-trace generation, using a Cyrillic-dense tokenizer and an adapted EAGLE speculative-decoding pipeline to reduce latency. To enable reproducible and extensible research, we release the model weights, the T-Wix 500k instruction corpus, the T-Math reasoning benchmark, and the EAGLE weights on Hugging Face. These resources allow users to study Russian-language reasoning and to extend or adapt both the model and the inference pipeline. A public web demo exposes reasoning and non-reasoning modes and illustrates the speedups achieved by our inference stack across domains. T-pro 2.0 thus serves as an accessible open system for building and evaluating efficient, practical Russian LLM applications.

顶级标签: llm natural language processing model training
详细标签: russian language model speculative decoding reasoning benchmark efficient inference instruction tuning 或 搜索:

T-pro 2.0:一个高效的俄语混合推理模型与实验平台 / T-pro 2.0: An Efficient Russian Hybrid-Reasoning Model and Playground


1️⃣ 一句话总结

这篇论文发布了一个名为T-pro 2.0的高效开源俄语大语言模型,它不仅支持直接回答和生成推理步骤,还通过优化技术降低了响应延迟,并配套发布了训练数据、评测基准和推理工具,旨在为构建和评估实用的俄语AI应用提供一个可访问的开放系统。


源自 arXiv: 2512.10430