FedAFD: Multimodal Federated Learning via Adversarial Fusion and Distillation
AI 摘要
FedAFD提出了一种新的多模态联邦学习框架,通过对抗融合和蒸馏提升客户端和服务器端的学习效果。
主要贡献
- 提出了双层对抗对齐策略,缓解模态和任务差异
- 设计了细粒度融合模块,自适应整合全局知识
- 提出了基于相似性的集成蒸馏机制,处理模型异构性
方法论
采用双层对抗对齐、细粒度融合和基于相似性的集成蒸馏,提升多模态联邦学习性能。
原文摘要
Multimodal Federated Learning (MFL) enables clients with heterogeneous data modalities to collaboratively train models without sharing raw data, offering a privacy-preserving framework that leverages complementary cross-modal information. However, existing methods often overlook personalized client performance and struggle with modality/task discrepancies, as well as model heterogeneity. To address these challenges, we propose FedAFD, a unified MFL framework that enhances client and server learning. On the client side, we introduce a bi-level adversarial alignment strategy to align local and global representations within and across modalities, mitigating modality and task gaps. We further design a granularity-aware fusion module to integrate global knowledge into the personalized features adaptively. On the server side, to handle model heterogeneity, we propose a similarity-guided ensemble distillation mechanism that aggregates client representations on shared public data based on feature similarity and distills the fused knowledge into the global model. Extensive experiments conducted under both IID and non-IID settings demonstrate that FedAFD achieves superior performance and efficiency for both the client and the server.