Dual-Teacher Distillation with Subnetwork Rectification for Black-Box Domain Adaptation
AI 摘要
提出了一种双教师蒸馏方法,通过子网络校正解决黑盒域适应问题,提升模型性能。
主要贡献
- 提出了双教师蒸馏框架DDSR,结合黑盒模型和ViL的优势
- 引入子网络驱动的正则化策略,减轻噪声监督的影响
- 迭代优化伪标签和ViL提示,提高适应的准确性和语义一致性
方法论
利用黑盒模型的预测和视觉语言模型的语义信息,通过双教师蒸馏和子网络校正进行知识迁移。
原文摘要
Assuming that neither source data nor the source model is accessible, black box domain adaptation represents a highly practical yet extremely challenging setting, as transferable information is restricted to the predictions of the black box source model, which can only be queried using target samples. Existing approaches attempt to extract transferable knowledge through pseudo label refinement or by leveraging external vision language models (ViLs), but they often suffer from noisy supervision or insufficient utilization of the semantic priors provided by ViLs, which ultimately hinder adaptation performance. To overcome these limitations, we propose a dual teacher distillation with subnetwork rectification (DDSR) model that jointly exploits the specific knowledge embedded in black box source models and the general semantic information of a ViL. DDSR adaptively integrates their complementary predictions to generate reliable pseudo labels for the target domain and introduces a subnetwork driven regularization strategy to mitigate overfitting caused by noisy supervision. Furthermore, the refined target predictions iteratively enhance both the pseudo labels and ViL prompts, enabling more accurate and semantically consistent adaptation. Finally, the target model is further optimized through self training with classwise prototypes. Extensive experiments on multiple benchmark datasets validate the effectiveness of our approach, demonstrating consistent improvements over state of the art methods, including those using source data or models.