Federated Learning for Cross-Modality Medical Image Segmentation via Augmentation-Driven Generalization
AI 摘要
提出一种基于数据增强的联邦学习方法,解决跨模态医学图像分割泛化问题。
主要贡献
- 提出全局强度非线性增强方法(GIN)以模拟模态差异。
- 验证GIN在联邦学习框架下跨模态分割的有效性。
- 实现接近中心化训练的精度,保护数据隐私。
方法论
通过卷积空间增强、频域操作、领域特定归一化和全局强度非线性增强等策略,提高模型跨模态泛化能力,并在联邦学习框架下验证。
原文摘要
Artificial intelligence has emerged as a transformative tool in medical image analysis, yet developing robust and generalizable segmentation models remains difficult due to fragmented, privacy-constrained imaging data siloed across institutions. While federated learning (FL) enables collaborative model training without centralizing data, cross-modality domain shifts pose a critical challenge, particularly when models trained on one modality fail to generalize to another. Many existing solutions require paired multimodal data per patient or rely on complex architectures, both of which are impractical in real clinical settings. In this work, we consider a realistic FL scenario where each client holds single-modality data (CT or MRI), and systematically investigate augmentation strategies for cross-modality generalization. Using abdominal organ segmentation and whole-heart segmentation as representative multi-class and binary segmentation benchmarks, we evaluate convolution-based spatial augmentation, frequency-domain manipulation, domain-specific normalization, and global intensity nonlinear (GIN) augmentation. Our results show that GIN consistently outperforms alternatives in both centralized and federated settings by simulating cross-modality appearance variations while preserving anatomical structure. For the pancreas, Dice score improved from 0.073 to 0.437, a 498% gain. Our federated approach achieves 93-98% of centralized training accuracy, demonstrating strong cross-modality generalization without compromising data privacy, pointing toward feasible federated AI deployment across diverse healthcare systems.