Beyond Parameter Arithmetic: Sparse Complementary Fusion for Distribution-Aware Model Merging
AI 摘要
提出SCF-RKL模型融合框架,通过稀疏互补融合和分布感知更新,有效提升模型融合效果。
主要贡献
- 提出Sparse Complementary Fusion with reverse KL (SCF-RKL) 模型融合框架
- 利用reverse Kullback-Leibler divergence测量模型间的函数差异
- 通过稀疏更新选择性地整合互补参数,减少干扰
方法论
SCF-RKL通过RKL散度衡量模型差异,并进行稀疏互补融合,以分布感知的更新方式合并模型权重。
原文摘要
Model merging has emerged as a promising paradigm for composing the capabilities of large language models by directly operating in weight space, enabling the integration of specialized models without costly retraining. However, existing merging methods largely rely on parameter-space heuristics, which often introduce severe interference, leading to degraded generalization and unstable generation behaviors such as repetition and incoherent outputs. In this work, we propose Sparse Complementary Fusion with reverse KL (SCF-RKL), a novel model merging framework that explicitly controls functional interference through sparse, distribution-aware updates. Instead of assuming linear additivity in parameter space, SCF-RKL measures the functional divergence between models using reverse Kullback-Leibler divergence and selectively incorporates complementary parameters. This mode-seeking, sparsity-inducing design effectively preserves stable representations while integrating new capabilities. We evaluate SCF-RKL across a wide range of model scales and architectures, covering both reasoning-focused and instruction-tuned models. Extensive experiments on 24 benchmarks spanning advanced reasoning, general reasoning and knowledge, instruction following, and safety demonstrate, vision classification that SCF-RKL consistently outperforms existing model merging methods while maintaining strong generalization and generation stability.