Agent Tuning & Optimization 相关度: 7/10

Optimal Brain Decomposition for Accurate LLM Low-Rank Approximation

Yuhang Li, Donghyun Lee, Ruokai Yin, Priyadarshini Panda
arXiv: 2604.00821v1 发布: 2026-04-01 更新: 2026-04-01

AI 摘要

提出OBD-LLM,利用二阶 Hessian 信息进行LLM的低秩分解,显著提升分解效果。

主要贡献

  • 提出基于二阶 Hessian 信息的 OBD-LLM 分解方法
  • 理论证明了分解需要考虑输入和输出信息
  • 实现了比 SVD-LLM 更好的低秩分解效果

方法论

通过 Hessian 矩阵的 Kronecker 分解,推导出考虑输入和输出信息的双向白化权重矩阵分解方案,得到闭式解。

原文摘要

Low-rank decomposition has emerged as an important problem in Large Language Model (LLM) fine-tuning and inference. Through Singular Value Decomposition (SVD), the weight matrix can be factorized into low-rank spaces optimally. Previously, a common practice was to decompose the weight in the activation-whitened space, and then achieve satisfying results. In this work, we propose Optimal Brain Decomposition LLM (OBD-LLM), which studies the decomposition problem in the model space by utilizing second-order Hessian information. Through a rigorous Kronecker-factorization of the Hessian, we show that the decomposition needs to consider both input and output information of the layer, and achieves much better decomposition results compared to input only method. Our loss-aware decomposition method involves a bi-directional whitening on the weight matrix. As a result, OBD-LLM is a closed-form solution for the optimal decomposition of weights in the language model. Remarkably, we achieve ~20-40\% better results than previous state-of-the-art decomposition methods, the SVD-LLM.

标签

LLM 低秩分解 模型压缩 Hessian 优化

arXiv 分类

cs.LG