MSSR: Memory-Aware Adaptive Replay for Continual LLM Fine-Tuning
AI 摘要
MSSR通过估计样本记忆强度自适应地进行经验回放,有效缓解了LLM持续微调中的灾难性遗忘。
主要贡献
- 提出了Memory-Inspired Sampler and Scheduler Replay (MSSR)框架
- MSSR通过记忆强度估计自适应地选择回放样本和调整回放频率
- 实验证明MSSR在多种任务上优于现有replay方法,尤其在推理任务上表现突出
方法论
MSSR估计样本级别的记忆强度,并根据记忆强度自适应地采样和调度回放,以减少灾难性遗忘。
原文摘要
Continual fine-tuning of large language models (LLMs) is becoming increasingly crucial as these models are deployed in dynamic environments where tasks and data distributions evolve over time. While strong adaptability enables rapid acquisition of new knowledge, it also exposes LLMs to catastrophic forgetting, where previously learned skills degrade during sequential training. Existing replay-based strategies, such as fixed interleaved replay, accuracy-supervised, and loss-driven scheduling, remain limited: some depend on heuristic rules and provide only partial mitigation of forgetting, while others improve performance but incur substantial computational overhead. Motivated by retention dynamics under sequential fine-tuning, we propose Memory-Inspired Sampler and Scheduler Replay (MSSR), an experience replay framework that estimates sample-level memory strength and schedules rehearsal at adaptive intervals to mitigate catastrophic forgetting while maintaining fast adaptation. Extensive experiments across three backbone models and 11 sequential tasks show that MSSR consistently outperforms state-of-the-art replay baselines, with particularly strong gains on reasoning-intensive and multiple-choice benchmarks.