菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-02-24
📄 Abstract - Test-Time Training with KV Binding Is Secretly Linear Attention

Test-time training (TTT) with KV binding as sequence modeling layer is commonly interpreted as a form of online meta-learning that memorizes a key-value mapping at test time. However, our analysis reveals multiple phenomena that contradict this memorization-based interpretation. Motivated by these findings, we revisit the formulation of TTT and show that a broad class of TTT architectures can be expressed as a form of learned linear attention operator. Beyond explaining previously puzzling model behaviors, this perspective yields multiple practical benefits: it enables principled architectural simplifications, admits fully parallel formulations that preserve performance while improving efficiency, and provides a systematic reduction of diverse TTT variants to a standard linear attention form. Overall, our results reframe TTT not as test-time memorization, but as learned linear attention with enhanced representational capacity.

顶级标签: theory model training model evaluation
详细标签: test-time training linear attention kv binding online learning sequence modeling 或 搜索:

基于KV绑定的测试时训练本质上是线性注意力机制 / Test-Time Training with KV Binding Is Secretly Linear Attention


1️⃣ 一句话总结

这篇论文通过分析发现,基于键值对绑定的测试时训练并非传统认为的在线记忆学习,而本质上是一种学习到的线性注意力算子,这一新视角不仅解释了模型行为,还带来了架构简化和效率提升。

源自 arXiv: 2602.21204