| Literature DB >> 33301073 |
Khaled Bayoudh1, Fayçal Hamdaoui2, Abdellatif Mtibaa3.
Abstract
The novel Coronavirus disease (COVID-19), which first appeared at the end of December 2019, continues to spread rapidly in most countries of the world. Respiratory infections occur primarily in the majority of patients treated with COVID-19. In light of the growing number of COVID-19 cases, the need for diagnostic tools to identify COVID-19 infection at early stages is of vital importance. For decades, chest X-ray (CXR) technologies have proven their ability to accurately detect respiratory diseases. More recently, with the availability of COVID-19 CXR scans, deep learning algorithms have played a critical role in the healthcare arena by allowing radiologists to recognize COVID-19 patients from their CXR images. However, the majority of screening methods for COVID-19 reported in recent studies are based on 2D convolutional neural networks (CNNs). Although 3D CNNs are capable of capturing contextual information compared to their 2D counterparts, their use is limited due to their increased computational cost (i.e. requires much extra memory and much more computing power). In this study, a transfer learning-based hybrid 2D/3D CNN architecture for COVID-19 screening using CXRs has been developed. The proposed architecture consists of the incorporation of a pre-trained deep model (VGG16) and a shallow 3D CNN, combined with a depth-wise separable convolution layer and a spatial pyramid pooling module (SPP). Specifically, the depth-wise separable convolution helps to preserve the useful features while reducing the computational burden of the model. The SPP module is designed to extract multi-level representations from intermediate ones. Experimental results show that the proposed framework can achieve reasonable performances when evaluated on a collected dataset (3 classes to be predicted: COVID-19, Pneumonia, and Normal). Notably, it achieved a sensitivity of 98.33%, a specificity of 98.68% and an overall accuracy of 96.91.Entities:
Keywords: COVID-19; Chest X-ray; Deep learning; Hybrid 2D/3D CNN; Pneumonia
Mesh:
Year: 2020 PMID: 33301073 PMCID: PMC7726306 DOI: 10.1007/s13246-020-00957-1
Source DB: PubMed Journal: Phys Eng Sci Med ISSN: 2662-4729
Fig. 1Temporal evolution of CXR outcomes in an older male patient with three consolidation changes in the lung periphery [13]
Fig. 2A selection of CXR samples taken from the collected dataset: a Normal case (a), a COVID-19 patient (b), and a Pneumonia case (c)
Distribution of collected samples per category
| Collection | COVID-19 | Normal | Pneumonia |
|---|---|---|---|
| C1 | 535 | – | – |
| C2 | 219 | 670 | – |
| C3 | – | 671 | 1345 |
| Total | 3440 |
Fig. 3Workflow diagram of the proposed framework
Output shapes of the proposed architecture
| Layer | Output shape |
|---|---|
| Input | |
| Conv2D | |
| Conv2D | |
| MaxPooling2D | |
| Conv2D | |
| Conv2D | |
| MaxPooling2D | |
| Conv2D | |
| Conv2D | |
| Conv2D | |
| MaxPooling2D | |
| Conv2D | |
| Conv2D | |
| Conv2D | |
| MaxPooling2D | |
| Conv2D | |
| Conv2D | |
| Conv2D | |
| MaxPooling2D | |
| UpSampling2D | |
| Depth-wiseConv2D | |
| Reshape | |
| Conv3D | |
| Conv3D | |
| Conv3D | |
| Reshape | |
| Spatial pyramid pooling | 35,840 |
| Dense | 64 |
| Dropout | 64 |
| Dense | 3 (Output) |
Fig. 4Depth-wise separable convolution steps
Fig. 5A schematic illustration of a 3D convolution layer
Fig. 6Proposed spatial pyramid pooling (SPP) structure. Here, 3-level pyramid pooling is applied: , , and , respectively
Comparison with state-of-the art methods. The best performances achieved are marked in bold
| Method | Architecture | Resolution | Number of CXR samples | SEN (%) | SPE (%) | ACC (%) |
|---|---|---|---|---|---|---|
| Hybrid-COVID | 2D CNN + 3D CNN | 3440 | ||||
| DarkCovidNet | 2D CNN | 3440 | 88.10 | 93.20 | 89.20 | |
| VGG19 + Transfer learning | 2D CNN | 3440 | 93.15 | 96.25 | 95.71 | |
| CoroNet | 2D CNN | 3440 | 94.17 | 96.20 | 93.89 | |
| CapsNet | 2D CNN | 3440 | 85.90 | 90.11 | 86.27 |
Characteristics of generated datasets
| Dataset | COVID-19 | Normal | Pneumonia |
|---|---|---|---|
| Dataset-1 | 535 | 670 | – |
| Dataset-2 | 535 | 671 | – |
| Dataset-3 | 535 | 1341 | – |
| Dataset-4 | 535 | 671 | 1345 |
| Dataset-5 | 535 | 1341 | 1345 |
| Dataset-6 | 219 | 670 | – |
| Dataset-7 | 219 | 671 | – |
| Dataset-8 | 219 | 1341 | – |
| Dataset-9 | 219 | 671 | 1345 |
| Dataset-10 | 219 | 1341 | 1345 |
| Dataset-11 | 754 | 670 | – |
| Dataset-12 | 754 | 671 | – |
| Dataset-13 | 754 | 1341 | – |
| Dataset-14 | 754 | 671 | 1345 |
| Dataset-15 | 754 | 1341 | 1345 |
Fig. 7Visualization of model performance per epoch for: training and validation accuracy (a) and training and validation loss (b)
Fig. 8Confusion matrix of the Hybrid-COVID model
Evaluation of classification performance
| Method | Resolution | Class | PREC (%) | REC (%) | F-SCORE (%) |
|---|---|---|---|---|---|
| Hybrid-COVID | COVID-19 | 99.77 | 99.67 | 99.13 | |
| Normal | 96.75 | 95.82 | 95.98 | ||
| Pneumonia | 95.83 | 96.88 | 96.24 |
Comparison of classification results for each generated dataset (Datasets-1 to -15)
| Dataset | SEN (%) | SPE (%) | ACC (%) |
|---|---|---|---|
| Dataset-1 | 98.21 | 100 | 99.10 |
| Dataset-2 | 96.98 | 100 | 98.87 |
| Dataset-3 | 98.39 | 99.85 | 99.33 |
| Dataset-4 | 97.08 | 100 | 98.90 |
| Dataset-5 | 97.23 | 100 | 98.94 |
| Dataset-6 | 97.40 | 99.67 | 98.76 |
| Dataset-7 | 96.01 | 100 | 97.21 |
| Dataset-8 | 97.97 | 100 | 99.52 |
| Dataset-9 | 97.39 | 99.84 | 99.04 |
| Dataset-10 | 98.03 | 100 | 99.32 |
| Dataset-11 | 96.76 | 100 | 98.45 |
| Dataset-12 | 97.39 | 100 | 98.60 |
| Dataset-13 | 97.11 | 98.93 | 98.04 |
| Dataset-14 | 96.04 | 98.83 | 97.01 |
| Dataset-15 | 96.68 | 98.91 | 97.47 |
Comparison of performance assessment between the baseline and three scenarios
| Method | Scenario | Total parameters | ACC (%) |
|---|---|---|---|
| Hybrid-COVID (2D CNN + 3D CNN) | Baseline | 17M | 96.91 |
| Hybrid-COVID (2D CNN + 3D CNN++) | 1 | 96.06 | |
| Hybrid-COVID (2D CNN + 2D CNN) | 2 | 14M | 94.78 |
| Hybrid-COVID (2D CNN only) | 3 | 92.81 |
Fig. 9Visualization of some of the feature maps produced by the output of the building block (3D CNN)
Fig. 10CXR instances and corresponding localization maps: localization maps (a) and CXR instances (b)