| Literature DB >> 33776607 |
Ines Feki1, Sourour Ammar1,2, Yousri Kessentini1,2, Khan Muhammad3.
Abstract
Today, the whole world is facing a great medical disaster that affects the health and lives of the people: the COVID-19 disease, colloquially known as the Corona virus. Deep learning is an effective means to assist radiologists to analyze the vast amount of chest X-ray images, which can potentially have a substantial role in streamlining and accelerating the diagnosis of COVID-19. Such techniques involve large datasets for training and all such data must be centralized in order to be processed. Due to medical data privacy regulations, it is often not possible to collect and share patient data in a centralized data server. In this work, we present a collaborative federated learning framework allowing multiple medical institutions screening COVID-19 from Chest X-ray images using deep learning without sharing patient data. We investigate several key properties and specificities of federated learning setting including the not independent and identically distributed (non-IID) and unbalanced data distributions that naturally arise. We experimentally demonstrate that the proposed federated learning framework provides competitive results to that of models trained by sharing data, considering two different model architectures. These findings would encourage medical institutions to adopt collaborative process and reap benefits of the rich private data in order to rapidly build a powerful model for COVID-19 screening.Entities:
Keywords: CNN; COVID-19 screening; Decentralized training; Deep learning; Federated learning; X-ray images
Year: 2021 PMID: 33776607 PMCID: PMC7979273 DOI: 10.1016/j.asoc.2021.107330
Source DB: PubMed Journal: Appl Soft Comput ISSN: 1568-4946 Impact factor: 6.725
Fig. 1Federated Learning architecture for COVID-19 detection from Chest X-ray images.
Fig. 2Sample Chest X-ray images from the used dataset. Left : sample images selected from the original dataset. Right : corresponding augmented images generated with random zoom and rotation augmentations.
Fig. 3Comparison of Federated Learning to data-sharing learning using original and augmented dataset for learning. Left: results using the VGG16 as the model backbone. Right: results using the ResNet50 as the model backbone. An epoch for centralized methods is defined as a single training pass over all of the centralized data. A round for FL methods is defined as a parallel training pass of every client over their local training data.
Fig. 4Comparison of Federated Learning to data-sharing learning using original and augmented dataset for learning. Curves represent average results obtained over the 5 simulations for each method. Left: results using the VGG16 as the model backbone. Right: results using the ResNet50 as the model backbone.
Accuracy, Sensitivity, and Specificity rates after the last FL round/Data sharing epoch. Reported results are given with respect to our experiments made with the 5-fold cross-validation method. Accuracy, Sensitivity, and Specificity rates provided in this table are average results over the 5 simulations.
| Method | Accuracy | Sensitivity | Specificity |
|---|---|---|---|
| FL-VGG16 | 93.57 | 95.03 | 92.12 |
| FL-VGG16 | 94.40 | 96.15 | 92.66 |
| Centralized-VGG16 | 93.75 | 95.20 | 92.3 |
| Centralized-VGG16 | 94.0 | 95.01 | 93.0 |
| FL-ResNet50 | 95.4 | 96.03 | 94.78 |
| FL-ResNet50 | 97.0 | 98.11 | 95.89 |
| Centralized-ResNet50 | 95.3 | 96.0 | 94.6 |
| Centralized-ResNet50 | 96.5 | 96.8 | 96.2 |
Fig. 5Effect of the client fraction on the test accuracy of our proposed method FL-VGG16. Note corresponds to all clients are selected at each round (4 clients in our case), corresponds to half clients (2 clients in our case) and corresponds to only one client per round. Left: results using the VGG16 as the model backbone. Right: results using the ResNet50 as the model backbone.
Fig. 6Comparison of Federated Learning results on IID data and non-IID data partitions with (all clients are considered at each round). Left: results using the VGG16 as the model backbone. Right: results using the ResNet50 as the model backbone.
Fig. 7Comparison of Federated Learning results on balanced data and unbalanced data partitions with (all clients are considered at each round). Left: results using the VGG16 as the model backbone. Right: results using the ResNet50 as the model backbone.