LLM Reasoning 相关度: 7/10

Sparse-BitNet: 1.58-bit LLMs are Naturally Friendly to Semi-Structured Sparsity

Di Zhang, Xun Wu, Shaohan Huang, Yudong Wang, Hanyong Shao, Yingbo Hao, Zewen Chi, Li Dong, Ting Song, Yan Xia, Zhifang Sui, Furu Wei
arXiv: 2603.05168v1 发布: 2026-03-05 更新: 2026-03-05

AI 摘要

论文提出Sparse-BitNet,结合1.58-bit量化与N:M稀疏化,提升LLM效率并加速训练和推理。

主要贡献

  • 提出Sparse-BitNet框架
  • 证明1.58-bit量化与N:M稀疏化的兼容性
  • 实现训练和推理加速

方法论

联合应用1.58-bit量化和动态N:M稀疏化,定制稀疏张量核心,并在不同模型规模和训练方式下进行实验。

原文摘要

Semi-structured N:M sparsity and low-bit quantization (e.g., 1.58-bit BitNet) are two promising approaches for improving the efficiency of large language models (LLMs), yet they have largely been studied in isolation. In this work, we investigate their interaction and show that 1.58-bit BitNet is naturally more compatible with N:M sparsity than full-precision models. To study this effect, we propose Sparse-BitNet, a unified framework that jointly applies 1.58-bit quantization and dynamic N:M sparsification while ensuring stable training for the first time. Across multiple model scales and training regimes (sparse pretraining and dense-to-sparse schedules), 1.58-bit BitNet consistently exhibits smaller performance degradation than full-precision baselines at the same sparsity levels and can tolerate higher structured sparsity before accuracy collapse. Moreover, using our custom sparse tensor core, Sparse-BitNet achieves substantial speedups in both training and inference, reaching up to 1.30X. These results highlight that combining extremely low-bit quantization with semi-structured N:M sparsity is a promising direction for efficient LLMs. Code available at https://github.com/AAzdi/Sparse-BitNet

标签

量化 稀疏化 LLM BitNet 效率

arXiv 分类

cs.CL