📄 论文总结
Orion-MSP:面向表格上下文学习的多尺度稀疏注意力机制 / Orion-MSP: Multi-Scale Sparse Attention for Tabular In-Context Learning
1️⃣ 一句话总结
这篇论文提出了一个名为Orion-MSP的创新模型,通过多尺度特征处理和高效稀疏注意力机制,解决了现有表格数据处理方法在捕捉层次依赖和计算效率上的不足,实现了无需专门训练即可达到领先水平的表格数据学习能力。
Tabular data remain the predominant format for real-world applications. Yet, developing effective neural models for tabular data remains challenging due to heterogeneous feature types and complex interactions occurring at multiple scales. Recent advances in tabular in-context learning (ICL), such as TabPFN and TabICL, have achieved state-of-the-art performance comparable to gradient-boosted trees (GBTs) without task-specific fine-tuning. However, current architectures exhibit key limitations: (1) single-scale feature processing that overlooks hierarchical dependencies, (2) dense attention with quadratic scaling in table width, and (3) strictly sequential component processing that prevents iterative representation refinement and cross-component communication. To address these challenges, we introduce Orion-MSP, a tabular ICL architecture featuring three key innovations: (1) multi-scale processing to capture hierarchical feature interactions; (2) block-sparse attention combining windowed, global, and random patterns for scalable efficiency and long-range connectivity; and (3) a Perceiver-style memory enabling safe bidirectional information flow across components. Across diverse benchmarks, Orion-MSP matches or surpasses state-of-the-art performance while scaling effectively to high-dimensional tables, establishing a new standard for efficient tabular in-context learning. The model is publicly available at this https URL .
Orion-MSP:面向表格上下文学习的多尺度稀疏注意力机制 / Orion-MSP: Multi-Scale Sparse Attention for Tabular In-Context Learning
这篇论文提出了一个名为Orion-MSP的创新模型,通过多尺度特征处理和高效稀疏注意力机制,解决了现有表格数据处理方法在捕捉层次依赖和计算效率上的不足,实现了无需专门训练即可达到领先水平的表格数据学习能力。