Multimodal Learning 相关度: 9/10

Med-MMFL: A Multimodal Federated Learning Benchmark in Healthcare

Aavash Chhetri, Bibek Niroula, Pratik Shrestha, Yash Raj Shrestha, Lesley A Anderson, Prashnna K Gyawali, Loris Bazzani, Binod Bhattarai
arXiv: 2602.04416v1 发布: 2026-02-04 更新: 2026-02-04

AI 摘要

提出了首个综合性的医学多模态联邦学习(MMFL)基准Med-MMFL,促进该领域研究。

主要贡献

  • 提出了医学多模态联邦学习基准Med-MMFL
  • 涵盖多种模态、任务和联邦场景
  • 评估了六种代表性的联邦学习算法

方法论

构建包含多种医学模态的数据集,模拟不同联邦学习场景,评估现有算法在分割、分类等任务上的表现,并公开基准实现。

原文摘要

Federated learning (FL) enables collaborative model training across decentralized medical institutions while preserving data privacy. However, medical FL benchmarks remain scarce, with existing efforts focusing mainly on unimodal or bimodal modalities and a limited range of medical tasks. This gap underscores the need for standardized evaluation to advance systematic understanding in medical MultiModal FL (MMFL). To this end, we introduce Med-MMFL, the first comprehensive MMFL benchmark for the medical domain, encompassing diverse modalities, tasks, and federation scenarios. Our benchmark evaluates six representative state-of-the-art FL algorithms, covering different aggregation strategies, loss formulations, and regularization techniques. It spans datasets with 2 to 4 modalities, comprising a total of 10 unique medical modalities, including text, pathology images, ECG, X-ray, radiology reports, and multiple MRI sequences. Experiments are conducted across naturally federated, synthetic IID, and synthetic non-IID settings to simulate real-world heterogeneity. We assess segmentation, classification, modality alignment (retrieval), and VQA tasks. To support reproducibility and fair comparison of future multimodal federated learning (MMFL) methods under realistic medical settings, we release the complete benchmark implementation, including data processing and partitioning pipelines, at https://github.com/bhattarailab/Med-MMFL-Benchmark .

标签

联邦学习 多模态学习 医学图像 基准测试

arXiv 分类

cs.CV cs.AI