Literature DB >> 32397844

Using X-ray images and deep learning for automated detection of coronavirus disease.

Khalid El Asnaoui1, Youness Chawki2.   

Abstract

Coronavirus is still the leading cause of death worldwide. There are a set number of COVID-19 test units accessible in emergency clinics because of the expanding cases daily. Therefore, it is important to implement an automatic detection and classification system as a speedy elective finding choice to forestall COVID-19 spreading among individuals. Medical images analysis is one of the most promising research areas, it provides facilities for diagnosis and making decisions of a number of diseases such as Coronavirus. This paper conducts a comparative study of the use of the recent deep learning models (VGG16, VGG19, DenseNet201, Inception_ResNet_V2, Inception_V3, Resnet50, and MobileNet_V2) to deal with detection and classification of coronavirus pneumonia. The experiments were conducted using chest X-ray & CT dataset of 6087 images (2780 images of bacterial pneumonia, 1493 of coronavirus, 231 of Covid19, and 1583 normal) and confusion matrices are used to evaluate model performances. Results found out that the use of inception_Resnet_V2 and Densnet201 provide better results compared to other models used in this work (92.18% accuracy for Inception-ResNetV2 and 88.09% accuracy for Densnet201).Communicated by Ramaswamy H. Sarma.

Entities:  

Keywords:  CT and X-ray images; Computer-aided diagnosis; Covid-19; coronavirus automatic detection; deep learning; pneumonia

Mesh:

Year:  2020        PMID: 32397844      PMCID: PMC7256347          DOI: 10.1080/07391102.2020.1767212

Source DB:  PubMed          Journal:  J Biomol Struct Dyn        ISSN: 0739-1102


Introduction

Identified first time in Wuhan city of China in late December 2019 (Aanouz et al., 2020; Elfiky & Azzam, 2020; Elmezayen et al., 2020; Enayatkhani et al., 2020; Fausto et al., 2020; Ghosh et al., 2020; Muralidharan et al., 2020; Phulen et al., 2020; Rajib et al., 2020; Rothan & Byrareddy, 2020; Rowan & Laffey, 2020; Salman et al., 2020; Sourav et al., 2020; Umesh et al., 2020), Covid19 is a respiratory disease that is caused by the new coronavirus SARS-CoV-2 (Severe Acute Respiratory Syndrome Coronavirus 2 (SARS-Cov-2) 2019 (Abdelli et al., 2020; Rakesh et al., 2020; Rameez et al., 2020; World Health Organization, 2020). Covid19 is the name given by the World Health Organization (WHO) on 11 February 2020 (World Health Organization, 2020). It causes illnesses ranging from the common cold to more severe pathologies (Rowan & Laffey, 2020). First people to have contracted the virus went to Wuhan market in China’s Hubei Province. The disease would, therefore, to have originated from an animal (zoonosis) but the origin has not been confirmed (Boopathi et al., 2020; CNHC, 2020; El Zowalaty & Järhult, 2020; Kandel et al., 2020; Manoj et al., 2020; Pant et al., 2020; Roosa et al., 2020; Sun et al., 2020; Wilder-Smith et al., 2020). The period between contamination and appearance of the first Covid19 symptoms can extend to 15 days. Therefore, people carrying the virus without knowing can affect other people, which allows the spread of the virus in a large way. In fact, after a few weeks of confirmed cases in Wuhan, Covid19 was not only spread into China but crossed the border (212 countries) and the number of people affected increased and claimed many victims. Indeed, on 11 February 2020, the World Health Organization declared COVID19 a pandemic (Djalante et al., 2020; Landry et al., 2020; Lee & Morling, 2020; Rowan & Laffey, 2020; Saurabh et al., 2020; Wahedi et al., 2020; World Health Organization, 2020). While writing this paper, the number of confirmed cases has reached 3,641,205 including 251,943 deaths, and 1,192,948 recovred (Worldometers, 2020) (Last updated: May 04, 2020, 23:45 GMT). The main symptoms of Covid19 are fever (38 °C or higher), dry cough, the difficulty of breathing, tiredness, aches, and pains, sore throat, and diarrhea for some people. Sudden loss of smell, without nasal obstruction and total disappearance of taste, are also symptoms that have been observed in patients. In people developing more severe forms, respiratory difficulties are found, which can lead to hospitalization in intensive care and death (Landry et al., 2020; Sharifi-Razavi et al., 2020; Simcock et al., 2020; World Health Organization, 2020). The way with Covid19 is transmitted makes it a very dangerous disease. In fact, the disease can be transmitted by droplets (secretions projected invisible when talking, sneezing, or coughing). It is therefore considered that close contact with a sick person is necessary to transmit the disease: the same place of residence, direct contact within one meter when talking, coughing, sneezing, or in the absence of protective measures. One of the other preferred vectors of virus transmission is in contact with unwashed hands soiled with droplets (Anwarul et al., 2020; Elfiky, 2020a; Enmozhi et al., 2020; Landry et al., 2020; Li, Guan, et al., 2020; Liu, Han, et al., 2020; Lu et al., 2020; Rowan & Laffey, 2020; Yang et al., 2020). Waiting for a vaccine of Covid19, the World Health Organization is advising that certain precautions be taken (Elkbuli et al., 2020; Li et al., 2020; World Health Organization, 2020). We can cite the following precautions: frequent hand washing with soap or a hydroalcoholic solution; avoid close contact, such as kissing or shaking hands, with people who are coughing or sneezing; covering of mouth with the crease of the elbow, or a disposable handkerchief, when coughing or sneezing; no touching the eyes, nose or mouth; in case of respiratory symptoms and fever, wearing a mask. To limit the spreading of Covid19, some countries limited the movements, and the activities in the cities, and others are under lockdown (Elfiky, 2020; Ghosh et al., 2020; Li et al., 2020). The real-time polymerase chain reaction (RT-PCR) is the standard for detecting Covid19, but its problem is that it takes time to confirm patients as well as it’s expensive (Huang et al., 2020). Therefore, medical image processing can overcome this problem by affirming positive covid19 patients. Indeed, Chest X-ray and Computed Tomography (CT) are the most used image in medical image processing (Liu et al., 2020; Ng et al., 2020) and several researches used them to develop models that can help radiologists to predict the disease. On the other hand, in the last years, deep learning gave an excellent result in medical image analysis and this allows the specialists to make good decisions when diagnosing patients. Hence, various studies have demonstrated the capacity of neural systems, particularly convolutional neural systems to precisely recognize the presence of Pneumonia (Gozes et al., 2020; Xu et al., 2020). In this study, we are going to present a comparison of different Deep Convolutional Neural Network (DCNN) algorithms (VGG16, VGG19, DenseNet201, Inception_ResNet_V2, Inception_V3, Resnet50, and MobileNet_V2) to automatically classify X-ray images into Coronavirus, Bacteria, and Normal. The contributions of our paper are as follows: (1) We design fined tuned versions of (VGG16, VGG19, DenseNet201, Inception_ResNet_V2, Inception_V3, Resnet50, and MobileNet_V2), (2) To avoid over-fitting in different models, we used weight decay and L2-regularizers. (3) The models have been tested on the chest X-ray & CT dataset for multiclass classification. The structure of this paper is as follows. The paper starts with detailing the literature review in Section 2, Section 3 describes the proposed method, Section 4 presents some results obtained and their interpreting. The discussion is given in Section 5. Conclusion section ends the paper along with few upcoming tasks to be headed in the section” Conclusions and future directions”.

Related work

Since the vaccine is not yet developed, the right measure to reduce the epidemic is to isolate people who are positively affected. But the problem is making a quick diagnostic to distinguish positive patients from negative. In this scenario, several studies were presented allowing to identify abnormalities in Chest X-ray and CT images. Indeed, Gozes et al. (2020) proposed a model allowed to differentiate coronavirus patients from healthy patients. The proposed system produced a localization map of the lung abnormality as well as measurements. Indeed, it was split into two subsystems: Subsystem A: a 3D analysis was used to detect nodules and small opacities using commercial off-the-shelf software thereafter measurements and localization were provided. Subsystem B: the first step is the lung Crop stage where the lung region of interest (ROI) was extracted using a lung segmentation module (U-net architecture). The second step is the detection of coronavirus abnormalities using deep convolutional neural network model ResNet50. The third step was the abnormality localization step. If a new slice identified positive, the network-activation maps were extracted using the Grad-cam technique. Thereafter, after the combination of the output of subsystem A and subsystem B, the authors added a Corona score calculated by a volumetric summation of the network-activation maps. An automatic and deep learning-based method using X-ray images to predict Covid19 was proposed by Narin et al. (2020). The proposed method used three Deep Convolution Neural Network architectures. They have used a dataset containing 50 X-ray images of covid19 patients and 50 normal X-ray images and all the images were resized to 224 × 224. To overcome the problem of the limited number dataset, the authors used transfer learning models. The dataset was divided into two parts: 80% for training and 20% for testing. The developed DCNN was based on pre-trained models (ResNet50, Inception_V3, and Inceptio_ResNet_V2) allowed to identify Covid19 from normal X-ray images. They used also a transfer learning technique and the k-fold method was used as a cross-validation method with k = 5. The obtained results showed the pre-trained model ResNet50 gave good (the value of accuracy is equal to 98%). In Hemdan et al. (2020), a deep learning classifiers framework “COVIDX-Net” helping radiologists to automatically identify Covid19 was proposed. The developed framework allows classifying Covid19 X-ray images into positive and negative Covid19. Authors used seven DCNN architectures (VGG19, DenseNet121, ResNetV2, InceptionV3, InceptionResNetV2, Xception, and MobileNetV2). They also used a dataset including 50 X-ray images split into two categories normal and Covid19 positive cases (25 X-ray images for each). The images were resized to 224 × 224 pixels. 80% of images were used for the training stage and 20% for testing. The obtained results depicted that VGG19 and DenseNet201 architectures have good performances with an F1 score of 89% and 91% for normal and covid19. In order to identify Covid19 cases from other Pneumonia (Bacteria and virus) and normal cases, Farooq and Hafeez (2020) proposed a convolutional neural network (CNN) framework. They used the COVDIX dataset made by Wang and Wong (2020). The dataset contains 5941 chest radiography images collected from 2839 patients. In this work, they used a portion of the COVIDX dataset, and it was divided into four sets: Covid19 (48 images), Bacterial (660), Viral (931), and Normal (1203 images). In the training step, which was performed in 3 steps, the Cyclical Learning Rate was used for helping to select the optimal learning rate and that for each step. The obtained results depict that the proposed Covid-ResNes gave good identification accuracy of 96.23% compared to Covid-Net 83.5%. Bhandary et al. (2020) reported a deep learning framework to classify lung abnormalities like pneumonia using chest X-ray images and cancer using lung CT images. The proposed model was based on a Modified AlexNet model (MAN). Hence, they proposed two models: A) a MAN model combined with Support Vector Machine (SVM) used to identify pneumonia images from normal images. For the results, the proposed model showed good results (accuracy 96.8%) compared to other models AlexNet, VGG16, VGG19, ResNet50, and MAN_Softmax. B) For this examination, the lung CT images were used. Authors merged MAN with Ensemble-Feature-Technique (EFT) to improve the performance of classification. After extracting features from images, the Principal Component Analysis (PCA) was implemented. Finally, to classify CT images into Malignant and Benign, the model was combined with SVM, k-Nearest Neighbors (k-NN), and Random Forest (RF). The obtained results depicted that MAN combined with SVM achieved good accuracy with and without EFT 97.27% and 86.47 respectively. In Zhang et al. (2020), the authors presented a deep learning model allowed to detect Covid19 from healthy people using Chest X-ray images. The model was based on three components: The first one is the backbone network which is composed of 18 layers residual convolutional neural network. Its rule is to extract the high-level features from the chest X-ray image. The second one is the classification head intended to generate a classification score Pcls. It was powered by the extracted features by the backbone network. The third component is the anomaly detection head allows generating a scalar anomaly score Pano. After calculating the classification score and scalar anomaly score, the decision was made according to a threshold T. The obtained results showed that the sensitivity decreased as long as the value of threshold T decreased (sensitivity of 96% for T = 0.15). The work (Xu et al., 2020) reported a method to distinguish COVID-19 from Influenza-A viral pneumonia and healthy images using deep learning techniques. They used multiple CNN to classify Computed Tomography (CT) images. The presented process can be summarized as 4 steps: 1) the images were pre-processed to extract effective pulmonary regions 2) a 3D CNN was used to segment multiple candidate image cubes 3) a model of image classification was used to distinguish the images patch into Covid19, Influenza-A and normal 4) by using the noisy-or Bayesian function an overall analysis report for one CT sample was calculated. The VNET-IR-RPN model was used for the segmentation while ResNet-18 model and ResNet-18 with the location-attention mechanism model were used for the classification step. The experimental results show that the ResNet-18 model with the location-attention mechanism gave the overall accuracy rate of 86.7%. A new deep learning model, that allows segmenting and quantifying infection regions in CT scans of COVID-19 patients, was reported by Shan et al. (2020). Authors used VB-Net Neural Network and a human-in-the-loop (HITL) approach in the order to help radiologists to clarify automatic annotation of each case. Then, they used evaluation metrics to assess the effectiveness of the model (volumes and percentage of infection in the whole lung). They divided the CT images into a set of collections. These CT images that were contoured manually by the radiologists will feed the segmentation network for training. Then, the segmentation results were manually corrected by radiologists and were considered as new data to feed the model. This process was repeated to iteratively build the model. El Asnaoui et al. (2020) presented a comparison of recent DCNN architectures for automatic binary classification of pneumonia images based on fined tuned versions of VGG16 (Simonyan & Zisserman, 2014; Zhang et al., 2019), VGG19 (Simonyan & Zisserman, 2014; Zhang et al., 2019), DenseNet201 (Huang et al., 2017), Inception_ResNet_V2 (Szegedy et al., 2016), Inception_V3 (Szegedy et al., 2015), Resnet50 (He et al., 2016) and MobileNet_V2 (Sandler et al., 2018). The proposed work has been tested using chest X-ray & CT dataset. The study selection is designed for high sensitivity over precision, to guarantee that no relevant studies were leaved out. At this time, all works done in this field focus on binary classification except few studies. For this purpose, the main goal of this work is going to present a comparison of recent deep convolutional neural network architectures for automatic multiclass classification of X-ray and CT images between normal, bacteria, and coronavirus in order to answer to the following research questions (RQ): RQ1). Is there any DL technique that distinctly outperforms other DL techniques? RQ2). Can DL use to early screen coronavirus from CT and X-ray images? RQ3). What is the diagnostic accuracy that DL can be attained based on CT and X-ray images? RQ4). Can DL assist in the efforts to accurately detect and track the progression or resolution of the coronavirus?

Materials and methodology

In this study, we built our contribution for automatic multiclass classification on two new publicly available image datasets (chest X-ray & CT dataset) (Cohen et al., 2020; Kermany et al., 2018). Figure 1 depicts the diagram of the main proposed methodology. As it is shown, the entire contribution is mainly divided into four steps: dataset, data pre-processing, pre-trained transfer models, and finally training and classification. The following sections provide in detail the steps of the present contribution.
Figure 1.

Block diagram of the proposed methodology.

Block diagram of the proposed methodology.

Dataset

This present work introduces two publicly available image datasets that contain X-ray and computed tomography (CT) images. The first dataset (Kermany et al., 2018) is named chest X-ray & CT dataset and composed of 5856 images and has two categories (4273 pneumonia and 1583 normal) whereas the second one is named Covid Chest X-ray Dataset (Cohen et al., 2020). It contains 231 Covid19 Chest X-ray images. We added images of the second dataset to the first one in order to constitute our dataset which finally composed of 6087 images (jpeg format) and has three classes (2780 bacterial pneumonia, 1724 coronavirus (1493 viral pneumonia, 231 covid19) and 1583 normal). Figure 2 depicts an example of chest X-rays in patients with pneumonia, the normal chest X-ray (Figure 2(a)) shows clear lungs with no zones of abnormal opacification. Moreover, Figure 2(b) shows a focal lobar consolidation (white arrows). In addition, Figure 2(c) shows with a more diffuse “interstitial” pattern in both lungs (Kermany et al., 2018) while Figure 2(d) presents an image of patient infected by covid19 (Cohen et al., 2020).
Figure 2.

Examples of Chest X-rays in patients with pneumonia.

Examples of Chest X-rays in patients with pneumonia.

Data preprocessing

The next stage is to pre-process input images using different pre-processing techniques. The motivation behind image pre-processing is to improve the quality of visual information of each input image (eliminate or decrease noise present in the original input image, improve image quality through increased contrast, delete the low or high frequencies, etc). In this study, we used intensity normalization and Contrast Limited Adaptive Histogram Equalization (CLAHE) (El Asnaoui et al., 2020). For data splitting, the dataset was randomly split in this experiment with 80% of the images for training and 20% of the images for validation. We ensure that the images chosen for validation are not used during training in order to perform successfully the classification task.

Pre-trained transfer models

In this study, we implemented the present contribution for automatic multiclass classification based VGG16, VGG19, DenseNet201, Inception_ResNet_V2, Inception_V3, Resnet50, and MobileNet_V2 models for the classification of Chest X-ray images to normal, bacteria and coronavirus classes. These different models are explained in (El Asnaoui et al., 2020). Moreover, these deep learning models require a large amount of training data, which is yet not available in this field of applications (El Asnaoui et al., 2020). Following the context of no availability of medical imaging dataset and motivated by the success of deep learning and medical image processing, the present work is going to apply transfer learning technique that was utilized by using ImageNet data to overcome the training time and insufficient data. Data augmentation is used for the training process after dataset pre-processing and splitting and has the goal to avoid the risk of over-fitting. Moreover, the strategies we used include geometric transforms such as rescaling, rotations, shifts, shears, zooms, and flips (El Asnaoui et al., 2020).

Training and classification

After data pre-processing, splitting, and data augmentation techniques used, the training dataset size is increased and ready to be passed to the feature extraction step with the proposed models in order to extract the appropriate and pertinent features. The extracted features from each proposed model are flattened together to create the vectorized feature maps. The generated feature vector is passed to a multilayer perceptron to classify each image into corresponding classes. The Cyclical Learning Rate was used for helping to select the optimal learning rate (Smith, 2017).

Experiments

Experimental parameters

The present experimentations were performed based on the following experimental parameters: For simulation, Python programming language is used, and Keras/tensorflow as deep learning backend. The training and validation steps were performed on NVIDIA Tesla P40 with 24 Go RAM. Moreover, all the images of the dataset were resized to 224 × 224 pixels except those of Inception_V3 and Inception_Resnet_V2 models that were resized to 299 × 299. To train the models, we set batch size, number of epochs, and learning rate to 32, 300, and 0.00001 respectively. The learning rate used is based on Cyclical Learning Rates (Smith, 2017) with these parameters: base_lr = 0.00001, max_lr = 0.001, step_size = 2000, mode = exp_range and gamma = 0.99994. Adam with β1 = 0.9, β2 = 0.999 is used for optimization. Besides, we employed weight decay to reduce the over-fitting of the models. A fully connected layer was trained with the Rectified Linear Unit (ReLU). For fine-tuning, we modified the last dense layer in all models to output three classes corresponding to normal, bacteria and coronavirus instead of 1000 classes as was used for ImageNet. Categorical_crossentropy was used in this work as a classical loss function. The implementation of the proposed deep transfer learning models is done using a computer with Processor: Intel (R) Core (TM) i7- 7700 CPU @ 3.60 GHZ and 8 Go in RAM running on a Microsoft Windows 10 Professional (64-bit).

Performance metrics

The performance of the proposed classification model was evaluated based on accuracy, sensitivity, specificity, precision, and F1 score (Bhandary et al., 2020; Blum & Chawla, 2001). Given the number of false positives (FP), true positives (TP), false negatives (FN) and true negatives (TN), the parameters are mathematically defined as follows: Moreover, the present study supports the use of confusion matrix analysis in validation (Ruuska et al., 2018) since it is strong to type of relationship and any data distribution, it makes a stringent evaluation of validity, and it provides extra information on the type and sources of errors. Before starting the analysis of the confusion matrix of each model, let’s first see how it is structured and define all the parameters and variables that can be extracted (Table 1).
Table 1.

Confusion matrix structure.

 Predicted
BacteriaCoronavirusNormal
ActualBacteriaPbbPcbPnb
CoronavirusPbcPccPnc
NormalPbnPcnPnn

where:

Pbb :Bacteria class were correctly classified as Bacteria.

Pcb :Bacteria class were incorrectly classified as Coronavirus.

Pnb :Bacteria class were incorrectly classified as Normal.

Pbc :Coronavirus class were incorrectly classified as Bacteria.

Pcc :Coronavirus class were correctly classified as Coronavirus.

Pnc :Coronavirus class were incorrectly classified as Normal.

Pbn :Normal class were incorrectly classified as Bacteria.

Pcn :Normal class were incorrectly classified as Coronavirus.

Pnn :Normal class were correctly classified as Normal.

Confusion matrix structure. where: Pbb :Bacteria class were correctly classified as Bacteria. Pcb :Bacteria class were incorrectly classified as Coronavirus. Pnb :Bacteria class were incorrectly classified as Normal. Pbc :Coronavirus class were incorrectly classified as Bacteria. Pcc :Coronavirus class were correctly classified as Coronavirus. Pnc :Coronavirus class were incorrectly classified as Normal. Pbn :Normal class were incorrectly classified as Bacteria. Pcn :Normal class were incorrectly classified as Coronavirus. Pnn :Normal class were correctly classified as Normal. Using these parameters, we can define other variables: True Positives TP: True Negatives TN: False Positives FP: False Negatives FN:

Multi-classification results

In this section, we present the multi-classification results followed by a brief discussion of the results given by each model.

DensNet201

The table (Table 2) of Densnet201 reports that Normal class was identified with good precision, sensitivity and specificity (92.59%, 95.88%, and 96.28% respectively), that means that the sum of false positives was low, the sum of false negatives was low and the sum of true negatives was high respectively. The accuracy value is equal to 31.31% which is a third of the model’s accuracy (Figure 3).
Table 2.

Evaluation metric for DenNet201.

ClassTPTNFNFPAccuracy (%)Sensitivity (%)Specificity (%)Precision (%)F1 Score (%)
Normal5131062224131.3195.8896.2892.5994.21
Bacteria5289594011132.2392.9589.6282.6287.48
Coronavirus40210601334324.5475.1496.1090.3382.04
Figure 3.

Confusion matrix of DensNet201.

Confusion matrix of DensNet201. Evaluation metric for DenNet201. For Bacteria class (Table 2), it was identified with good sensitivity of 92.95% because the sum of false negatives was low. About specificity and precision, their values were reasonable 89.62% and 82.62%, and this means that sum of true negatives was relatively high, and the sum of false positives was relatively low respectively. For accuracy, it is equal to 32.23%. Regarding Coronavirus class (Table 2), it was distinguished well since precision and specificity were good (90.33%, and 96.10%) and reasonable sensitivity (75.14%). These values can be explained by the fact that the sum of false positives was low, the sum of true negatives was high, and the sum of false negatives was low respectively. The accuracy value is equal to 24.54% which is a third of the model’s accuracy.

Inception_resnet_V2

For Inception_Resnet_V2 table (Table 3), we see that Normal class was classified with good precision, sensitivity, and specificity. The value of precision is 94.40% means that sum of false positives was low. For sensitivity, the value is 97.75% because the sum of false negatives was low while the specificity is of value 97.18% which is by reason of the sum of true negatives was high. Finally, the value of accuracy is 31.92% (Figure 4).
Table 3.

Evaluation metric for Inception_Resnet_V2.

ClassTPTNFNFPAccuracy (%)Sensitivity (%)Specificity (%)Precision (%)F1 Score (%)
Normal5231072123131.9297.7597.1894.4096.05
Bacteria5441002246833.2195.7793.6488.8892.20
Coronavirus4431074922927.0482.8097.3793.8587.98
Figure 4.

Confusion matrix of Inception_ResNet_V2.

Confusion matrix of Inception_ResNet_V2. Evaluation metric for Inception_Resnet_V2. Bacteria class (Table 3) was detected with good specificity and sensitivity (93.64% and 95.77%) and with reasonable precision of 88.88%. The values obtained are due to the fact that the sum of true negatives was high, the sum of false negatives was low, and the sum of false positives was relatively low. Whereas the accuracy is equivalent to 33.21%. For the Coronavirus class (Table 3), we observe that it was identified with good specificity and precision (97.37% and 93.85%) and likewise with a moderate value of sensitivity 82.80%. We can explain those values by the fact the sum of true negatives was high, and the sum of false positives was low (specificity and precision) and also the sum of false negatives was practical low (sensitivity). We notice that the value of accuracy is 27.04%.

Inception_V3

The table of Inception_V3 model (Table 4) depicts that the Normal class was distinguished well since precision, sensitivity, and specificity reached good value (93.76%, 95.51%, and 96.91%). This can be explained by the sum of false positives and false negatives were low (precision and sensitivity), and the sum of true negatives was high. Furthermore, accuracy has a value of 31.19%.
Table 4.

Evaluation metric for Inception_V3.

ClassTPTNFNFPAccuracy (%)Sensitivity (%)Specificity (%)Precision (%)F1 Score (%)
Normal5111069243431.1995.5196.9193.7694.62
Bacteria5259634310732.0592.4290.0083.0687.50
Coronavirus40610481295524.7875.8895.0188.0681.52
Evaluation metric for Inception_V3. Concerning Bacteria class (Table 4), it was detected well since sensitivity and specificity were equivalent to 92.42% and 90.00% and also with tolerable precision (83.06%). These values can be interpreted by the sum of false negatives was low and the sum of true negatives was high (sensitivity and specificity). On the other hand, the value of precision (83.06%) is due to the sum of true negatives was low. Moreover, it can be observed that the value of accuracy is 32.05%. Coronavirus class (Table 4) was identified relatively well because precision and sensitivity were reasonable and also with good specificity. Their values were 88.06%, 75.88% and 95.01% respectively. We can explain these values like this: the sum of false positives was relatively low for the precision; the sum of false negatives was practically low for the sensitivity and the sum of true negatives was high for the specificity. While the value of accuracy is equal to 24.78% (Figure 5).
Figure 5.

Confusion matrix of Inception_V3.

Confusion matrix of Inception_V3.

Mobilenet_V2

The obtained results by Mobilenet_V2 (Table 5) tell us that Normal class was detected with good values of precision, sensitivity, and specificity (92.41%, 95.70%, and 96.19%). These achieved values are due to the fact that the false positives and sum of false positives were low for precision and sensitivity respectively and also the sum of true negatives was high. We can see that the value of the accuracy is 31.25%.
Table 5.

Evaluation metric for Mobilenet_V2.

ClassTPTNFNFPAccuracy (%)Sensitivity (%)Specificity (%)Precision (%)F1 Score (%)
Normal5121061234231.2595.7096.1992.4194.03
Bacteria4639871058328.2681.5192.2484.7983.12
Coronavirus42599011011325.9479.4389.7578.9979.21
Evaluation metric for Mobilenet_V2. Acceptable precision and sensitivity where the values are 84.79% and 81.51% and also good specificity of value 92.24% were identified Bacteria class (Table 5). This means that sum of false positives and the sum of false negatives were low (precision and sensitivity). Likewise, the value of specificity is explained by the sum of true negatives was high. As we can see the value of accuracy is 28.26% (Figure 6).
Figure 6.

Confusion matrix of Mobilenet_V2.

Confusion matrix of Mobilenet_V2. About Coronavirus class (Table 5), it was detected relatively well since precision, sensitivity, and specificity were reasonable. The values are 78.99%, 79.43% and 89.75% respectively. The obtained results can be explained by the sum of false positives and false negatives were practically low for the precision and sensitivity as well as the sum of true negatives was acceptable high for the specificity. While the value of the accuracy is 25.94%.

Resnet50

Concerning the Resnet50 results (Table 6), we may notice that Normal class was detected with good precision, sensitivity, and specificity (95.22%, 93.27%, and 97.73%). This can be explained by the sum of false positives and false negatives were low (precision and sensitivity). Besides, the value of specificity is due to the fact of the sum of true negatives that was high. As noticed, the value of accuracy is 30.46%.
Table 6.

Evaluation metric for Resnet50.

ClassTPTNFNFPAccuracy (%)Sensitivity (%)Specificity (%)Precision (%)F1 Score (%)
Normal4991078362530.4693.2797.7395.2294.23
Bacteria506982628830.8989.0891.7785.1887.09
Coronavirus42910121069126.1980.1891.7482.5081.32
Evaluation metric for Resnet50. For the Bacteria class (Table 6), the good value of specificity (91.77%) is obtained because the sum of true negatives was high. In addition, the value of sensitivity and precision (89.08% and 85.18%) were reasonable since the sum of false negatives and false positives were low. For accuracy, the value is equal to 30.89%. Regarding Coronavirus class (Table 6), we can see that it was distinguished relatively well because precision and sensitivity were reasonable with values: 82.50% and 80.18% respectively and also with good value of specificity (91.74%). This means that the sum of false positives and false negatives were acceptable low and also the sum of true negatives was high. The value of accuracy is equal to 26.19% (Figure 7).
Figure 7.

Confusion matrix of Resnet50.

Confusion matrix of Resnet50.

VGG16

The results given by VGG16 (Table 7) show that Normal class was detected with good sensitivity (92.71%) that is caused by the low sum of false negatives. Likewise, it distinguished with reasonable precision and specificity (77.01% and 86.58%) which is caused by the low sum of false positives and the high sum of true negatives. We can see that the value of accuracy is 30.28%.
Table 7.

Evaluation metric for VGG16.

ClassTPTNFNFPAccuracy (%)Sensitivity (%)Specificity (%)Precision (%)F1 Score (%)
Normal4969553914830.2892.7186.5877.0184.13
Bacteria44888812018227.3578.8782.9971.1174.79
Coronavirus28210212538217.2152.7192.5677.4762.73
Evaluation metric for VGG16. Bacteria class (Table 7) was identified with acceptable values of precision, sensitivity, and specificity (71.11%, 78.87%, and 82.99%). These values are produced by the low sum of false positives and sum of false negatives and by the high sum of true negatives. The value of accuracy is 27.35%. Coronavirus class (Table 7) was determined with good specificity (92.56%) given by the high sum of true negatives and with reasonable precision (77.47%) caused by the low sum of false positives. In addition, it is distinguished with wicked sensitivity (52.71%) which is a result of the poor sum of false negatives. As it is observed, the accuracy value is 17.21% (Figure 8).
Figure 8.

Confusion matrix of VGG16.

Confusion matrix of VGG16.

VGG19

Regarding VGG19 results (Table 8), we observe that Normal class was classified with good specificity (90.29%) and reasonable sensitivity and precision (85.23% and 80.99%). These values are obtained by the fact that the sum of true negatives was high (specificity) and the sum of false positives and the false negatives were low (precision and sensitivity). We can see that the value of accuracy is 27.83%.
Table 8.

Evaluation metric for VGG19.

ClassTPTNFNFPAccuracy (%)Sensitivity (%)Specificity (%)Precision (%)F1 Score (%)
Normal4569967910727.8385.2390.2980.9983.06
Bacteria4757989327228.9983.6274.5763.5872.24
Coronavirus25710322787115.6848.0393.5678.3559.55
Evaluation metric for VGG19. For Bacteria class (Table 8), it was identified relatively well since precision, sensitivity, and specificity were reasonable (63.58%, 83.62%, and 74.57%). This means that the sum of false positives was low for precision. In addition, the values of sensitivity and specificity are caused by the sum of false negatives that were low and the sum of true negatives that were high respectively. The value of the accuracy is equal to 28.99%. Analogous, Coronavirus class (Table 8) was detected with good specificity (93.56%) caused by the high sum of true negatives and with acceptable precision (78.35%) caused by the reasonable low sum of false positives. Likewise, it is identified with wicked sensitivity (48.03%) which is a result of the low sum of false negatives. For accuracy, its value is 15.68% (Figure 9).
Figure 9.

Confusion matrix of VGG19.

Confusion matrix of VGG19.

Experimental comparisons

This subsection compares the experimental results of classifying X-ray images using the different networks. The experimental results will be compared in terms of training and testing time and metrics defined in Equation (1). Table 9 is a summary of the confusion matrix performance of all used models. Thereby, as we can read, the highest performance values have been yielded an accuracy of 92.18%, sensitivity of 92.11%, specificity of 96.06%, precision of 92.38% and F1 score value of 92.07% for Inception_Resnet_V2 pre-trained model. Furthermore, good metrics values (accuracy 88.09%, sensitivity 87.99%, specificity 94.00%, precision 88.52% and F1 score 87.91%) were obtained by DensNet201.The same thing can be said for Inception_V3, the acquired performances were good (accuracy 88.03%, sensitivity 87.94%, specificity 93.97%, precision 88.30%, and F1 score 87.88%). Regarding Mobilenet_V2, the values attained can be described as follows: accuracy of 85.47%, sensitivity 85.55%, specificity 92.73%, precision 85.40%, and F1 score value of 85.45%. We can notice that good performance values were accomplished for Resnet50 (accuracy 87.54%, sensitivity 87.51%, specificity 93.75%, precision 87.63%, and F1 score 87.55%). About VGG16, the model produced low-performance values as the accuracy of 74.84%, the sensitivity of 74.76%, the specificity of 87.37%, the precision of 75.20%, and the F1 Score value of 73.88%. Moreover, the table shows us that VGG19 model achieved the lowest performance as the accuracy of 72.52%, the sensitivity of 72.29%, the specificity of 86.14%, the precision of 74.31%, and F1 score value of 71.62%. As a conclusion, the Inception_Resnet_V2 architecture furnished superiority up the other architectures both training and testing steps followed by Densnet201.
Table 9.

Evaluation metric for different models.

 Accuracy (%)Sensitivity (%)Specificity (%)Precision (%)F1 Score (%)
Inception_Resnet_V292.1892.1196.0692.3892.07
DensNet20188.0987.9994.0088.5287.91
Resnet5087.5487.5193.7587.6387.55
Mobilenet_V285.4785.5592.7385.4085.45
Inception_V388.0387.9493.9788.3087.88
VGG1674.8474.7687.3775.2073.88
VGG1972.5272.2986.1474.3171.62
Evaluation metric for different models. Furthermore, Table 10 depicts the comparative computational times in second for different models tested during this study. For inception_ResNet_V2, the elapsed time for training and testing stages was 79 184.28 s and 262 s. DensNet201 has required 68 859.73 and 225 s for training and testing steps. Likewise, for Resnet50, 58 069.93 and 194 s were required to finish training and testing steps. Similarly, for Mobilenet_V2, the elapsed time for training and testing were 58 693.21 and 196 s. For Inception_V3, it was necessary to have 58 485.06 and 193 s to achieve training and testing steps respectively. Regarding VGG16, it has required 53 621.49 and 181 s while VGG19 has required 53 493.08 and 181 s for training and testing steps respectively.
Table 10.

Comparative computational time in seconds.

ModelTraining (s)Testing (s)
Inception_Resnet_V279 184.28262
DensNet20168 859.73225
Resnet5058 069.93194
Mobilenet_V258 693.21196
Inception_V358 485.06193
VGG1653 621.49181
VGG1953 493.08181
Comparative computational time in seconds.

Discussion

In the present work, we conducted a comparative study of the most known deep learning architectures to detect and classify of coronavirus pneumonia using CT and X-ray images. From different tables above (Tables 2–8), we notice that the coronavirus class reaches low values in terms of accuracy, sensitivity, specificity, precision, and F1 score. Thereby, the main problem of the present work is the limited number of coronavirus X-ray images used for the training of different proposed deep learning models. In order to overcome this issue, we used deep transfer learning techniques. Moreover, Covid Chest X-ray Dataset (Cohen et al., 2020) contains a melange of brain and chest images that can decrease the accuracy and other metrics. The images of other classes are only images of chest. In the coming days, we are planning to improve this study with different models, if we reach more data. We compared the different pre-trained models according to accuracy, sensitivity, specificity, precision, F1 score, and training and testing times. As tabulated by Table 9, the obtained results showed that the Incpetion_Resnet_V2 gave good classification performance (92.18% of accuracy) followed by Densnet201 with 88.09% of accuracy. Contrariwise, VGG19 and VGG16 are the lowest compared with other DL architectures, since these last models help to obtain respectively 74.84% and 72.52% of accuracy. Moreover, Table 10 illustrates a comparison between the different deep learning models used in the experiments in terms of computational times. From this table, we observe that Incpetion_Resnet_V2 even it gives a good result it is not fast because it takes 79 184.28 and 262 in training and testing steps respectively followed by Densnet201. In addition, we notice that Inception_V3 is fast and provides good results (88.03% of accuracy). We can conclude that the scientist has the choice to choose between the accuracy and the computation time to finally select the DL technique to use, but since we are in the medical field, the accuracy of the DL techniques stays major selection criteria. Consequently, we recommend the Incpetion_Resnet_V2 (92.18% of accuracy, 92.11% of sensitivity, 96.06% of specificity, 92.38% of precision, and 92.07% of F1 score) model based on X-ray and CT images to be used to identify the health status of patients against the coronavirus. We hope that the results obtained during this study may serve as an initial step towards developing from X-ray and CT images a sophisticated coronavirus detection to save as many lives as possible.

Conclusion and future work

We investigated in this work automated methods used to classify the chest X-ray & CT images into bacterial pneumonia, coronavirus, and normal classes using seven deep learning architectures (VGG16, VGG19, DenseNet201, Inception_ResNet_V2, Inception_V3, Resnet50, and MobileNet_V2). The main goal is to answer the following research questions: RQ1). Is there any DL technique that distinctly outperforms other DL techniques? RQ2). Can DL use to early screen coronavirus from CT and X-ray images? RQ3). What is the diagnostic accuracy that DL can be attained based on CT and X-ray images? RQ4). Can DL assist in the efforts to accurately detect and track the progression or resolution of the coronavirus? Toward this end, the experiments were conducted using chest X-ray & CT dataset. Moreover, the performances of these experiments were evaluated using various performance metrics. Furthermore, the obtained results show that Inception_Resnet_V2 provides better results compared to other architectures cited in this work (accuracy is higher than 92%). Due to the high performance achieved by this model, we believe that these results help doctors to make decisions in clinical practice. Ongoing work intends to develop a full system for coronavirus using deep learning detection, segmentation, and classification. In addition, the performance may be improved using more datasets, more sophisticated feature extraction techniques based on deep learning such as You-Only-Look-Once (YOLO) (Al-Masni et al., 2018), and U-Net (Ronneberger et al., 2015) that was developed for biomedical image segmentation.
TP(Bacteria) :Pbb TN(Bacteria) :Pcc+Pnc+Pcn+Pnn
TP(Coronavirus) :PccTN(Coronavirus) :Pbb+Pnb+Pbn+Pnn
TP(Normal) :PnnTN(Normal) :Pbb+Pcb+Pbc+Pcc
TN(Bacteria) :Pbc+Pbn TN(Bacteria) :Pcb+Pnb
TN(Coronavirus) :Pcb+Pcn TN(Coronavirus) :Pbc+Pnc
TN(Normal) :Pnb+Pnc TN(Normal) :Pbn+Pcn
  49 in total

1.  Detection and Prevention of Virus Infection.

Authors:  Ying Wang; Bairong Shen
Journal:  Adv Exp Med Biol       Date:  2022       Impact factor: 2.622

2.  A Review on Deep Learning Techniques for the Diagnosis of Novel Coronavirus (COVID-19).

Authors:  Md Milon Islam; Fakhri Karray; Reda Alhajj; Jia Zeng
Journal:  IEEE Access       Date:  2021-02-10       Impact factor: 3.367

3.  A systematic review on AI/ML approaches against COVID-19 outbreak.

Authors:  Onur Dogan; Sanju Tiwari; M A Jabbar; Shankru Guggari
Journal:  Complex Intell Systems       Date:  2021-07-05

4.  A Systematic Review on the Use of AI and ML for Fighting the COVID-19 Pandemic.

Authors:  Muhammad Nazrul Islam; Toki Tahmid Inan; Suzzana Rafi; Syeda Sabrina Akter; Iqbal H Sarker; A K M Najmul Islam
Journal:  IEEE Trans Artif Intell       Date:  2021-03-01

5.  A Novel Multi-Stage Residual Feature Fusion Network for Detection of COVID-19 in Chest X-Ray Images.

Authors:  Zhenyu Fang; Jinchang Ren; Calum MacLellan; Huihui Li; Huimin Zhao; Amir Hussain; Giancarlo Fortino
Journal:  IEEE Trans Mol Biol Multiscale Commun       Date:  2021-07-26

6.  Feature-level ensemble approach for COVID-19 detection using chest X-ray images.

Authors:  Thi Kieu Khanh Ho; Jeonghwan Gwak
Journal:  PLoS One       Date:  2022-07-14       Impact factor: 3.752

7.  Covid-19 Imaging Tools: How Big Data is Big?

Authors:  K C Santosh; Sourodip Ghosh
Journal:  J Med Syst       Date:  2021-06-03       Impact factor: 4.460

8.  Multilevel Deep-Aggregated Boosted Network to Recognize COVID-19 Infection from Large-Scale Heterogeneous Radiographic Data.

Authors:  Muhammad Owais; Young Won Lee; Tahir Mahmood; Adnan Haider; Haseeb Sultan; Kang Ryoung Park
Journal:  IEEE J Biomed Health Inform       Date:  2021-06-03       Impact factor: 7.021

9.  Deep Learning in Classification of Covid-19 Coronavirus, Pneumonia and Healthy Lungs on CXR and CT Images.

Authors:  Mihaela-Ruxandra Lascu
Journal:  J Med Biol Eng       Date:  2021-06-10       Impact factor: 2.213

10.  A pilot study: Quantify lung volume and emphysema extent directly from two-dimensional scout images.

Authors:  Jiantao Pu; Jacob Sechrist; Xin Meng; Joseph K Leader; Frank C Sciurba
Journal:  Med Phys       Date:  2021-07-06       Impact factor: 4.506

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.