Model-Dowser: Data-Free Importance Probing to Mitigate Catastrophic Forgetting in Multimodal Large Language Models
AI 摘要
Model-Dowser通过参数重要性评估进行稀疏微调,有效缓解多模态大模型中的灾难性遗忘。
主要贡献
- 提出Model-Dowser方法,通过评估参数重要性缓解灾难性遗忘
- 该方法在不访问数据情况下选择性地保留重要参数
- 实验证明该方法在多个MLLM上优于现有方法
方法论
Model-Dowser计算参数重要性,结合权重、激活和输出敏感度。微调时,保留高重要性参数,更新其余参数。
原文摘要
Fine-tuning Multimodal Large Language Models (MLLMs) on task-specific data is an effective way to improve performance on downstream applications. However, such adaptation often leads to a degradation in generalization on pretrained tasks, a phenomenon known as Catastrophic Forgetting. Existing methods that aim to mitigate this issue either become ineffective when fine-tuning deeper layers of the language decoder or scale poorly with increasing model size. To address these limitations, we propose Model-Dowser, a novel sparse fine-tuning approach for MLLMs. Model-Dowser measures a principled importance score for each model parameter with respect to pretrained generalization (prior to downstream adaptation) by jointly considering weight magnitudes, input activations, and output sensitivities. During fine-tuning, Model-Dowser selectively preserves high-importance parameters and updates the remaining. Comprehensive experiments on two representative MLLMs, LLaVA and NVILA, demonstrate that Model-Dowser effectively mitigates catastrophic forgetting and consistently outperforms prior methods, while remaining resource-efficient and scalable to multi-billion-parameter models.