Agent Tuning & Optimization 相关度: 6/10

Revisiting Weight Regularization for Low-Rank Continual Learning

Yaoyue Zheng, Yin Zhang, Joost van de Weijer, Gido M van de Ven, Shaoyi Du, Xuetao Zhang, Zhiqiang Tian
arXiv: 2602.17559v1 发布: 2026-02-19 更新: 2026-02-19

AI 摘要

该论文提出EWC-LoRA方法,通过正则化低秩更新缓解参数高效持续学习中的任务干扰。

主要贡献

  • 提出EWC-LoRA方法,将EWC应用于低秩持续学习。
  • 利用低秩表示估计全维度参数重要性。
  • 实验证明EWC-LoRA在稳定性-可塑性权衡方面优于现有方法。

方法论

使用低秩适配器进行参数高效的持续学习,并通过EWC正则化共享的低秩更新,缓解任务间的干扰。

原文摘要

Continual Learning (CL) with large-scale pre-trained models (PTMs) has recently gained wide attention, shifting the focus from training from scratch to continually adapting PTMs. This has given rise to a promising paradigm: parameter-efficient continual learning (PECL), where task interference is typically mitigated by assigning a task-specific module during training, such as low-rank adapters. However, weight regularization techniques, such as Elastic Weight Consolidation (EWC)-a key strategy in CL-remain underexplored in this new paradigm. In this paper, we revisit weight regularization in low-rank CL as a new perspective for mitigating task interference in PECL. Unlike existing low-rank CL methods, we mitigate task interference by regularizing a shared low-rank update through EWC, thereby keeping the storage requirement and inference costs constant regardless of the number of tasks. Our proposed method EWC-LoRA leverages a low-rank representation to estimate parameter importance over the full-dimensional space. This design offers a practical, computational- and memory-efficient solution for CL with PTMs, and provides insights that may inform the broader application of regularization techniques within PECL. Extensive experiments on various benchmarks demonstrate the effectiveness of EWC-LoRA, achieving a stability-plasticity trade-off superior to existing low-rank CL approaches. These results indicate that, even under low-rank parameterizations, weight regularization remains an effective mechanism for mitigating task interference. Code is available at: https://github.com/yaoyz96/low-rank-cl.

标签

Continual Learning Low-Rank Adaptation Elastic Weight Consolidation Parameter-Efficient Learning

arXiv 分类

cs.LG