Multimodal Learning 相关度: 9/10

Model-Dowser: Data-Free Importance Probing to Mitigate Catastrophic Forgetting in Multimodal Large Language Models

Hyeontaek Hwang, Nguyen Dinh Son, Daeyoung Kim
arXiv: 2602.04509v1 发布: 2026-02-04 更新: 2026-02-04

AI 摘要

Model-Dowser通过参数重要性评估进行稀疏微调,有效缓解多模态大模型中的灾难性遗忘。

主要贡献

  • 提出Model-Dowser方法,通过评估参数重要性缓解灾难性遗忘
  • 该方法在不访问数据情况下选择性地保留重要参数
  • 实验证明该方法在多个MLLM上优于现有方法

方法论

Model-Dowser计算参数重要性,结合权重、激活和输出敏感度。微调时,保留高重要性参数,更新其余参数。

原文摘要

Fine-tuning Multimodal Large Language Models (MLLMs) on task-specific data is an effective way to improve performance on downstream applications. However, such adaptation often leads to a degradation in generalization on pretrained tasks, a phenomenon known as Catastrophic Forgetting. Existing methods that aim to mitigate this issue either become ineffective when fine-tuning deeper layers of the language decoder or scale poorly with increasing model size. To address these limitations, we propose Model-Dowser, a novel sparse fine-tuning approach for MLLMs. Model-Dowser measures a principled importance score for each model parameter with respect to pretrained generalization (prior to downstream adaptation) by jointly considering weight magnitudes, input activations, and output sensitivities. During fine-tuning, Model-Dowser selectively preserves high-importance parameters and updates the remaining. Comprehensive experiments on two representative MLLMs, LLaVA and NVILA, demonstrate that Model-Dowser effectively mitigates catastrophic forgetting and consistently outperforms prior methods, while remaining resource-efficient and scalable to multi-billion-parameter models.

标签

MLLM Catastrophic Forgetting Sparse Fine-tuning Importance Probing

arXiv 分类

cs.CL