Multimodal Learning 相关度: 6/10

GFPL: Generative Federated Prototype Learning for Resource-Constrained and Data-Imbalanced Vision Task

Shiwei Lu, Yuhang He, Jiashuo Li, Qiang Wang, Yihong Gong
arXiv: 2602.21873v1 发布: 2026-02-25 更新: 2026-02-25

AI 摘要

GFPL框架通过生成式联邦原型学习解决资源受限和数据不平衡的联邦学习问题。

主要贡献

  • 提出基于GMM的原型生成方法
  • 设计基于Bhattacharyya距离的原型聚合策略
  • 引入双分类器和混合损失函数

方法论

使用GMM捕获类别特征统计信息,用Bhattacharyya距离融合原型,并用双分类器和混合损失优化特征对齐。

原文摘要

Federated learning (FL) facilitates the secure utilization of decentralized images, advancing applications in medical image recognition and autonomous driving. However, conventional FL faces two critical challenges in real-world deployment: ineffective knowledge fusion caused by model updates biased toward majority-class features, and prohibitive communication overhead due to frequent transmissions of high-dimensional model parameters. Inspired by the human brain's efficiency in knowledge integration, we propose a novel Generative Federated Prototype Learning (GFPL) framework to address these issues. Within this framework, a prototype generation method based on Gaussian Mixture Model (GMM) captures the statistical information of class-wise features, while a prototype aggregation strategy using Bhattacharyya distance effectively fuses semantically similar knowledge across clients. In addition, these fused prototypes are leveraged to generate pseudo-features, thereby mitigating feature distribution imbalance across clients. To further enhance feature alignment during local training, we devise a dual-classifier architecture, optimized via a hybrid loss combining Dot Regression and Cross-Entropy. Extensive experiments on benchmarks show that GFPL improves model accuracy by 3.6% under imbalanced data settings while maintaining low communication cost.

标签

联邦学习 数据不平衡 原型学习 生成模型 图像识别

arXiv 分类

cs.CV cs.LG