Multi-Modal Federated Learning with Differential Privacy for Privacy-Preserving Healthcare AI
Discuss this preprint
Start a discussion What are Sciety discussions?Listed in
This article is not in any list yet, why not save it to one of your lists.Abstract
The growing adoption of artificial intelligence in healthcare highlights the need for models that can leverage heterogeneous patient data while preserving strict privacy requirements. This paper proposes a novel multi-modal federated learning framework with differential privacy for decentralized healthcare AI. The model integrates electronic health records and ECG time-series using modality-specific encoders and a shared latent fusion network, enabling comprehensive representation learning without centralizing sensitive data. Differential privacy is incorporated into local updates to provide formal guarantees against information leakage in federated aggregation. Extensive experiments on real-world healthcare datasets show that the proposed method achieves $94.12\%$ accuracy, $93.64\%$ precision, $93.21\%$ recall, $93.42\%$ F1-score, and $95.03\%$ AUC, outperforming centralized, single-modality, and non-private baselines. The framework also converges $32.4\%$ faster than single-modality federated learning, reaching $90\%$ accuracy in $35$ rounds. An ablation study confirms the contribution of multi-modal fusion and class balancing, while client variance analysis shows the lowest performance deviation ($\pm 1.2\%$) under heterogeneous distributions. These results indicate that combining federated optimization, differential privacy, and multi-modal learning provides an effective and scalable solution for privacy-preserving clinical AI.