TensorLens:通过高阶注意力张量进行端到端的Transformer分析 / TensorLens: End-to-End Transformer Analysis via High-Order Attention Tensors
1️⃣ 一句话总结
这篇论文提出了一个名为TensorLens的新方法,它用一个统一的高阶注意力张量来完整表示整个Transformer模型的计算过程,为模型可解释性研究提供了更强大的分析基础。
Attention matrices are fundamental to transformer research, supporting a broad range of applications including interpretability, visualization, manipulation, and distillation. Yet, most existing analyses focus on individual attention heads or layers, failing to account for the model's global behavior. While prior efforts have extended attention formulations across multiple heads via averaging and matrix multiplications or incorporated components such as normalization and FFNs, a unified and complete representation that encapsulates all transformer blocks is still lacking. We address this gap by introducing TensorLens, a novel formulation that captures the entire transformer as a single, input-dependent linear operator expressed through a high-order attention-interaction tensor. This tensor jointly encodes attention, FFNs, activations, normalizations, and residual connections, offering a theoretically coherent and expressive linear representation of the model's computation. TensorLens is theoretically grounded and our empirical validation shows that it yields richer representations than previous attention-aggregation methods. Our experiments demonstrate that the attention tensor can serve as a powerful foundation for developing tools aimed at interpretability and model understanding. Our code is attached as a supplementary.
TensorLens:通过高阶注意力张量进行端到端的Transformer分析 / TensorLens: End-to-End Transformer Analysis via High-Order Attention Tensors
这篇论文提出了一个名为TensorLens的新方法,它用一个统一的高阶注意力张量来完整表示整个Transformer模型的计算过程,为模型可解释性研究提供了更强大的分析基础。
源自 arXiv: 2601.17958