注意力残差 / Attention Residuals
1️⃣ 一句话总结
这篇论文提出了一种名为‘注意力残差’的新方法,它用可学习的注意力机制取代了传统大语言模型中固定权重的残差连接,从而让模型能根据输入内容动态选择性地组合不同深度的信息,有效缓解了深层网络中的信息稀释问题,并在实际训练中提升了模型性能。
Residual connections with PreNorm are standard in modern LLMs, yet they accumulate all layer outputs with fixed unit weights. This uniform aggregation causes uncontrolled hidden-state growth with depth, progressively diluting each layer's contribution. We propose Attention Residuals (AttnRes), which replaces this fixed accumulation with softmax attention over preceding layer outputs, allowing each layer to selectively aggregate earlier representations with learned, input-dependent weights. To address the memory and communication overhead of attending over all preceding layer outputs for large-scale model training, we introduce Block AttnRes, which partitions layers into blocks and attends over block-level representations, reducing the memory footprint while preserving most of the gains of full AttnRes. Combined with cache-based pipeline communication and a two-phase computation strategy, Block AttnRes becomes a practical drop-in replacement for standard residual connections with minimal overhead. Scaling law experiments confirm that the improvement is consistent across model sizes, and ablations validate the benefit of content-dependent depth-wise selection. We further integrate AttnRes into the Kimi Linear architecture (48B total / 3B activated parameters) and pre-train on 1.4T tokens, where AttnRes mitigates PreNorm dilution, yielding more uniform output magnitudes and gradient distribution across depth, and improves downstream performance across all evaluated tasks.
注意力残差 / Attention Residuals
这篇论文提出了一种名为‘注意力残差’的新方法,它用可学习的注意力机制取代了传统大语言模型中固定权重的残差连接,从而让模型能根据输入内容动态选择性地组合不同深度的信息,有效缓解了深层网络中的信息稀释问题,并在实际训练中提升了模型性能。
源自 arXiv: 2603.15031