Stable-LoRA: Stabilizing Feature Learning of Low-Rank Adaptation
AI 摘要
Stable-LoRA通过动态权重衰减优化LoRA,解决其特征学习不稳定的问题,提升模型性能。
主要贡献
- 揭示LoRA特征学习不稳定的根本原因
- 提出Stable-LoRA权重衰减优化策略
- 实验验证Stable-LoRA在多种模型和任务上的有效性
方法论
分析LoRA的特征学习稳定性,发现非零初始化导致不稳定。通过在训练初期逐步缩小A矩阵来增强稳定性。
原文摘要
Low-Rank Adaptation (LoRA) is a widely adopted parameter-efficient method for fine-tuning Large Langauge Models. It updates the weight matrix as $W=W_0+sBA$, where $W_0$ is the original frozen weight, $s$ is a scaling factor and $A$,$B$ are trainable low-rank matrices. Despite its robust empirical effectiveness, the theoretical foundations of LoRA remain insufficiently understood, particularly with respect to feature learning stability. In this paper, we first establish that, LoRA can, in principle, naturally achieve and sustain stable feature learning (i.e., be self-stabilized) under appropriate hyper-parameters and initializations of $A$ and $B$. However, we also uncover a fundamental limitation that the necessary non-zero initialization of $A$ compromises self-stability, leading to suboptimal performances. To address this challenge, we propose Stable-LoRA, a weight-shrinkage optimization strategy that dynamically enhances stability of LoRA feature learning. By progressively shrinking $A$ during the earliest training steps, Stable-LoRA is both theoretically and empirically validated to effectively eliminate instability of LoRA feature learning while preserving the benefits of the non-zero start. Experiments show that Stable-LoRA consistently outperforms other baselines across diverse models and tasks, with no additional memory usage and only negligible computation overheads. The code is available at https://github.com/Yize-Wu/Stable-LoRA.