菜单

关于 🐙 GitHub
arXiv 提交日期: 2025-12-16
📄 Abstract - Universal Reasoning Model

Universal transformers (UTs) have been widely used for complex reasoning tasks such as ARC-AGI and Sudoku, yet the specific sources of their performance gains remain underexplored. In this work, we systematically analyze UTs variants and show that improvements on ARC-AGI primarily arise from the recurrent inductive bias and strong nonlinear components of Transformer, rather than from elaborate architectural designs. Motivated by this finding, we propose the Universal Reasoning Model (URM), which enhances the UT with short convolution and truncated backpropagation. Our approach substantially improves reasoning performance, achieving state-of-the-art 53.8% pass@1 on ARC-AGI 1 and 16.0% pass@1 on ARC-AGI 2. Our code is avaliable at this https URL.

顶级标签: llm model training theory
详细标签: universal transformer reasoning arc-agi inductive bias model analysis 或 搜索:

通用推理模型 / Universal Reasoning Model


1️⃣ 一句话总结

这篇论文发现通用Transformer在复杂推理任务上的性能提升主要源于其循环结构和强大的非线性能力,并据此提出了一个结合短卷积和截断反向传播的改进模型,在ARC-AGI基准测试上取得了当前最好的成绩。


源自 arXiv: 2512.14693