Multimodal Learning 相关度: 9/10

BadCLIP++: Stealthy and Persistent Backdoors in Multimodal Contrastive Learning

Siyuan Liang, Yongcheng Jing, Yingjie Wang, Jiaxing Huang, Ee-chien Chang, Dacheng Tao
arXiv: 2602.17168v1 发布: 2026-02-19 更新: 2026-02-19

AI 摘要

BadCLIP++提出了一种隐蔽且持久的多模态对比学习后门攻击框架,有效抵抗检测和微调。

主要贡献

  • 提出语义融合QR微触发器,增强隐蔽性
  • 引入目标对齐子集选择,强化低注入率下的信号
  • 通过半径收缩、质心对齐、曲率控制和弹性权重巩固,提升后门持久性

方法论

设计隐蔽触发器并采用策略稳定触发器嵌入和模型参数,以抵抗检测和持续微调。

原文摘要

Research on backdoor attacks against multimodal contrastive learning models faces two key challenges: stealthiness and persistence. Existing methods often fail under strong detection or continuous fine-tuning, largely due to (1) cross-modal inconsistency that exposes trigger patterns and (2) gradient dilution at low poisoning rates that accelerates backdoor forgetting. These coupled causes remain insufficiently modeled and addressed. We propose BadCLIP++, a unified framework that tackles both challenges. For stealthiness, we introduce a semantic-fusion QR micro-trigger that embeds imperceptible patterns near task-relevant regions, preserving clean-data statistics while producing compact trigger distributions. We further apply target-aligned subset selection to strengthen signals at low injection rates. For persistence, we stabilize trigger embeddings via radius shrinkage and centroid alignment, and stabilize model parameters through curvature control and elastic weight consolidation, maintaining solutions within a low-curvature wide basin resistant to fine-tuning. We also provide the first theoretical analysis showing that, within a trust region, gradients from clean fine-tuning and backdoor objectives are co-directional, yielding a non-increasing upper bound on attack success degradation. Experiments demonstrate that with only 0.3% poisoning, BadCLIP++ achieves 99.99% attack success rate (ASR) in digital settings, surpassing baselines by 11.4 points. Across nineteen defenses, ASR remains above 99.90% with less than 0.8% drop in clean accuracy. The method further attains 65.03% success in physical attacks and shows robustness against watermark removal defenses.

标签

多模态学习 后门攻击 对比学习 安全性

arXiv 分类

cs.CV