LLM Memory & RAG 相关度: 6/10

Least but not Last: Fine-tuning Intermediate Principal Components for Better Performance-Forgetting Trade-Offs

Alessio Quercia, Arya Bangun, Ira Assent, Hanno Scharr
arXiv: 2602.03493v1 发布: 2026-02-03 更新: 2026-02-03

AI 摘要

通过微调中间主成分,LoRA方法在性能和遗忘之间实现了更好的权衡。

主要贡献

  • 分析了LoRA中性能-遗忘的权衡问题
  • 提出了一种基于中间主成分的LoRA初始化方法
  • 经验证明该方法在多个任务上提高了精度并减少了遗忘

方法论

通过分析主成分初始化LoRA,发现微调中间成分效果更好,并提出了一种更优的初始化方法,并在实验中验证其有效性。

原文摘要

Low-Rank Adaptation (LoRA) methods have emerged as crucial techniques for adapting large pre-trained models to downstream tasks under computational and memory constraints. However, they face a fundamental challenge in balancing task-specific performance gains against catastrophic forgetting of pre-trained knowledge, where existing methods provide inconsistent recommendations. This paper presents a comprehensive analysis of the performance-forgetting trade-offs inherent in low-rank adaptation using principal components as initialization. Our investigation reveals that fine-tuning intermediate components leads to better balance and show more robustness to high learning rates than first (PiSSA) and last (MiLoRA) components in existing work. Building on these findings, we provide a practical approach for initialization of LoRA that offers superior trade-offs. We demonstrate in a thorough empirical study on a variety of computer vision and NLP tasks that our approach improves accuracy and reduces forgetting, also in continual learning scenarios.

标签

LoRA 低秩适配 灾难性遗忘 微调 主成分分析 持续学习

arXiv 分类

cs.LG