Agent Tuning & Optimization 相关度: 6/10

Elastic Weight Consolidation Done Right for Continual Learning

Xuan Liu, Xiaobin Chang
arXiv: 2603.18596v1 发布: 2026-03-19 更新: 2026-03-19

AI 摘要

针对EWC在持续学习中的不足,提出Logits Reversal方法,显著提升性能。

主要贡献

  • 揭示EWC梯度消失和重要性估计不准确的问题
  • 发现MAS算法存在冗余保护问题
  • 提出Logits Reversal (LR) 操作校正EWC的重要性估计

方法论

通过分析EWC的梯度和信息矩阵,发现问题,提出LR操作来改进重要性估计,并通过实验验证。

原文摘要

Weight regularization methods in continual learning (CL) alleviate catastrophic forgetting by assessing and penalizing changes to important model weights. Elastic Weight Consolidation (EWC) is a foundational and widely used approach within this framework that estimates weight importance based on gradients. However, it has consistently shown suboptimal performance. In this paper, we conduct a systematic analysis of importance estimation in EWC from a gradient-based perspective. For the first time, we find that EWC's reliance on the Fisher Information Matrix (FIM) results in gradient vanishing and inaccurate importance estimation in certain scenarios. Our analysis also reveals that Memory Aware Synapses (MAS), a variant of EWC, imposes unnecessary constraints on parameters irrelevant to prior tasks, termed the redundant protection. Consequently, both EWC and its variants exhibit fundamental misalignments in estimating weight importance, leading to inferior performance. To tackle these issues, we propose the Logits Reversal (LR) operation, a simple yet effective modification that rectifies EWC's importance estimation. Specifically, reversing the logit values during the calculation of FIM can effectively prevent both gradient vanishing and redundant protection. Extensive experiments across various CL tasks and datasets show that the proposed method significantly outperforms existing EWC and its variants. Therefore, we refer to it as EWC Done Right (EWC-DR).

标签

持续学习 灾难性遗忘 弹性权重巩固 梯度分析

arXiv 分类

cs.LG cs.AI cs.CV