Shared LoRA Subspaces for almost Strict Continual Learning
AI 摘要
Share提出一种共享LoRA子空间的方法,用于解决严格持续学习中的灾难性遗忘问题。
主要贡献
- 提出Share方法,学习并动态更新共享低秩子空间
- 实现了高达100倍的参数缩减和281倍的内存节省
- 在多个任务和模态上验证了Share的有效性
方法论
Share通过构建基础子空间提取核心知识,并增量集成新信息,最小化灾难性干扰。
原文摘要
Adapting large pretrained models to new tasks efficiently and continually is crucial for real-world deployment but remains challenging due to catastrophic forgetting and the high cost of retraining. While parameter-efficient tuning methods like low rank adaptation (LoRA) reduce computational demands, they lack mechanisms for strict continual learning and knowledge integration, without relying on data replay, or multiple adapters. We propose Share, a novel approach to parameter efficient continual finetuning that learns and dynamically updates a single, shared low-rank subspace, enabling seamless adaptation across multiple tasks and modalities. Share constructs a foundational subspace that extracts core knowledge from past tasks and incrementally integrates new information by identifying essential subspace directions. Knowledge from each new task is incorporated into this evolving subspace, facilitating forward knowledge transfer, while minimizing catastrophic interference. This approach achieves up to 100x parameter reduction and 281x memory savings over traditional LoRA methods, maintaining performance comparable to jointly trained models. A single Share model can replace hundreds of task-specific LoRA adapters, supporting scalable, asynchronous continual learning. Experiments across image classification, natural language understanding, 3D pose estimation, and text-to-image generation validate its effectiveness, making Share a practical and scalable solution for lifelong learning in large-scale AI systems.