Tucker Attention: A generalization of approximate attention mechanisms
AI 摘要
提出了Tucker Attention,一种广义的近似注意力机制,在降低参数量同时保持性能。
主要贡献
- 提出了Tucker Attention,一种更参数高效的注意力机制
- Tucker Attention包含了GQA、MLA、MHA等作为特例
- 揭示了MHA、GQA、MLA的实际秩
方法论
通过对自注意力层中的权重对象进行广义分解,构建参数高效的Tucker Attention。
原文摘要
The pursuit of reducing the memory footprint of the self-attention mechanism in multi-headed self attention (MHA) spawned a rich portfolio of methods, e.g., group-query attention (GQA) and multi-head latent attention (MLA). The methods leverage specialized low-rank factorizations across embedding dimensions or attention heads. From the point of view of classical low-rank approximation, these methods are unconventional and raise questions of which objects they really approximate and how to interpret the low-rank behavior of the resulting representations. To answer these questions, this work proposes a generalized view on the weight objects in the self-attention layer and a factorization strategy, which allows us to construct a parameter efficient scheme, called Tucker Attention. Tucker Attention requires an order of magnitude fewer parameters for comparable validation metrics, compared to GQA and MLA, as evaluated in LLM and ViT test cases. Additionally, Tucker Attention~encompasses GQA, MLA, MHA as special cases and is fully compatible with flash-attention and rotary position embeddings (RoPE). This generalization strategy yields insights of the actual ranks achieved by MHA, GQA, and MLA, and further enables simplifications for MLA.