Federated learning with knowledge distillation for multi-organ segmentation with partially labeled datasets

Research output: Contribution to journalArticlepeer-review

25 Scopus citations

Abstract

The state-of-the-art multi-organ CT segmentation relies on deep learning models, which only generalize when trained on large samples of carefully curated data. However, it is challenging to train a single model that can segment all organs and types of tumors since most large datasets are partially labeled or are acquired across multiple institutes that may differ in their acquisitions. A possible solution is Federated learning, which is often used to train models on multi-institutional datasets where the data is not shared across sites. However, predictions of federated learning can be unreliable after the model is locally updated at sites due to ‘catastrophic forgetting’. Here, we address this issue by using knowledge distillation (KD) so that the local training is regularized with the knowledge of a global model and pre-trained organ-specific segmentation models. We implement the models in a multi-head U-Net architecture that learns a shared embedding space for different organ segmentation, thereby obtaining multi-organ predictions without repeated processes. We evaluate the proposed method using 8 publicly available abdominal CT datasets of 7 different organs. Of those datasets, 889 CTs were used for training, 233 for internal testing, and 30 volumes for external testing. Experimental results verified that our proposed method substantially outperforms other state-of-the-art methods in terms of accuracy, inference time, and the number of parameters.

Original languageEnglish
Article number103156
JournalMedical Image Analysis
Volume95
DOIs
StatePublished - Jul 2024

Bibliographical note

Publisher Copyright:
© 2024 Elsevier B.V.

Keywords

  • Federated learning
  • Knowledge distillation
  • Multi-organ segmentation
  • Partially labeled datasets

Fingerprint

Dive into the research topics of 'Federated learning with knowledge distillation for multi-organ segmentation with partially labeled datasets'. Together they form a unique fingerprint.

Cite this