Multimodal Learning 相关度: 7/10

CD-FKD: Cross-Domain Feature Knowledge Distillation for Robust Single-Domain Generalization in Object Detection

Junseok Lee, Sungho Shin, Seongju Lee, Kyoobin Lee
arXiv: 2603.16439v1 发布: 2026-03-17 更新: 2026-03-17

AI 摘要

提出CD-FKD,通过跨域特征知识蒸馏提升目标检测模型在单领域泛化中的鲁棒性。

主要贡献

  • 提出Cross-Domain Feature Knowledge Distillation (CD-FKD)方法
  • 利用全局和实例级特征蒸馏增强学生网络的泛化能力
  • 使用数据增强和扰动训练学生网络

方法论

通过数据扰动训练学生网络,模仿教师网络特征,利用全局和实例级蒸馏提取对象中心特征。

原文摘要

Single-domain generalization is essential for object detection, particularly when training models on a single source domain and evaluating them on unseen target domains. Domain shifts, such as changes in weather, lighting, or scene conditions, pose significant challenges to the generalization ability of existing models. To address this, we propose Cross-Domain Feature Knowledge Distillation (CD-FKD), which enhances the generalization capability of the student network by leveraging both global and instance-wise feature distillation. The proposed method uses diversified data through downscaling and corruption to train the student network, whereas the teacher network receives the original source domain data. The student network mimics the features of the teacher through both global and instance-wise distillation, enabling it to extract object-centric features effectively, even for objects that are difficult to detect owing to corruption. Extensive experiments on challenging scenes demonstrate that CD-FKD outperforms state-of-the-art methods in both target domain generalization and source domain performance, validating its effectiveness in improving object detection robustness to domain shifts. This approach is valuable in real-world applications, like autonomous driving and surveillance, where robust object detection in diverse environments is crucial.

标签

目标检测 领域泛化 知识蒸馏 鲁棒性

arXiv 分类

cs.CV cs.AI