Literature DB >> 33392388

FCOD: Fast COVID-19 Detector based on deep learning techniques.

Amir Hossein Panahi1, Alireza Rafiei1, Alireza Rezaee1.   

Abstract

The sudden COVID-19 pandemic has caused a serious global concern due to infections and mortality rates. It is a hazardous disease that has recently become the biggest crisis in the modern era. Due to the limitation of test kits and the need for screening and rapid diagnosis of patients, it is essential to perform a self-operating detection model as a fast recognition system to detect COVID-19 infection and prevent the spread among the people. In this paper, we propose a novel technique called Fast COVID-19 Detector (FCOD) to have a fast detection of COVID-19 using X-ray images. The FCOD is a deep learning model based on the Inception architecture that uses 17 depthwise separable convolution layers to detect COVID-19. Depthwise separable convolution layers decrease the computation costs, time, and they can have a reducing role in the number of parameters compared to the standard convolution layers. To evaluate FCOD, we used covid-chestxray-dataset, which contains 940 publicly available typical chest X-ray images. Our results show that FCOD can provide accuracy, F1-score, and AUC of 96%, 96%, and 0.95%, respectively in classifying COVID-19 during 0.014 s for each case. The proposed model can be employed as a supportive decision-making system to assist radiologists in clinics and hospitals to screen patients immediately.
© 2020 The Authors.

Entities:  

Keywords:  COVID-19 detection; Chest X-ray images; Deep learning; Image processing; Medical applications; Radiology images

Year:  2020        PMID: 33392388      PMCID: PMC7759122          DOI: 10.1016/j.imu.2020.100506

Source DB:  PubMed          Journal:  Inform Med Unlocked        ISSN: 2352-9148


Introduction

The coronavirus illness (COVID-19) is a universal pandemic that was found in December 2019 by a Chinese doctor in Wuhan, China [1]. Below the electron microscopes, the shape of these viruses is similar to the solar corona; accordingly, the researchists called them CoV [2]. The coronavirus family can cause serious infections and diseases like Middle East Respiratory Syndrome (MERS-CoV) and Severe Acute Respiratory Syndrome (SARS-CoV). On Feb 2020, the World Health Organization (WHO) called the pandemic COVID-19 because of the type of virus [3]. It has led to more than forty million infections and over one million deaths globally (up to Oct 20, 2020). The sputum's assay via polymerase chain reaction (RT-PCR) is one of Coronaviruses recognition's principal standards, but RT-PCR is a time-consuming process for the patients to identify COVID-19 [4]. Accordingly, medical imaging modalities, like Computed Tomography (CT) and Chest X-ray (CXR), can be a significant task in checking positive COVID-19 infections, specifically about pregnant women on the way and children [5,6]. Volumetric CT thorax images have been studied in some previous researches for recognizing COVID-19 [5,7]. The main limitation of using the CT method is high cost and time-consuming [8]. In contrast, X-ray systems are accessible in most clinics and hospitals, which are used to generate 2-dimensional visualization images of the patient's rib cage. Usually, the CXR manner is a premier option, which helps the radiologists to identify the chest pathology. According to the COVID-19 condition, the CXR manner is being utilized to detect COVID-19 [5,9]. Hence, the focus of this study is entirely on the use of X-ray imaging to potentially detect COVID-19 patients. Fig. 1 shows a COVID-19 patient case with pneumonia chest X-ray images for a week is given. The patient, when using CT, may get more amount of radiation compared to the X-ray. Hence, in some cases, it is suggested to utilize X-ray imagery [8,10]. Computer-Aided Diagnosis (CAD) systems have been extended to overcome this limitation and help physicians to identify suspected diseases of vital organs in X-ray images automatically [11,12]. These systems, which are chiefly supported by the fast and advanced of computer technology (like CPU, GPU, and, TPU) are used to operate the medical vision computing algorithms, containing image enhancement, classification, segmentation, and tumor detection tasks [[13], [14], [15]]. In several medical fields, Artificial intelligence (AI) methods like a deep neural network and machine learning turn to the core of the superior CAD applications. In recent years, deep learning methods have encouraged the results to perform radiological tasks by automated examining multi-modal medical images [[16], [17], [18]]. Deep learning techniques are one of the robust neural network structures. In an intuitional way been used in several practical applications in several practical usages, such as image segmentation and classification [19]. The use of machine learning techniques in the medical area for automated analysis has gained a new reputation by developing an assistive tool for clinicians [[20], [21], [22], [23], [24]]. Deep learning technique, which is one of the modern research fields of AI, provides the development of end-to-end methods to attain promised outcomes applying input information, in the absence of handy feature exploitation [25,26]. However, deep learning methods have been strongly used in various challenges like skin disease classification [27,28], arrhythmia identification [[29], [30], [31]], brain damages classification [32], breast cancer recognition [33,34], fundus image segmentation [35], pneumonia detection from chest X-ray images [36], and segmentation of the lung [37,38].
Fig. 1

Chest X-ray images of a 50-year-old COVID-19 patient case over a week.

Chest X-ray images of a 50-year-old COVID-19 patient case over a week. The rapid development of COVID-19 outbreaks necessitates the development of automated detection algorithms based on AI methods. Due to the limited number of radiologists in different areas, it is challenging to provide specialist clinicians for each hospital. Also, accurate, fast, and simple AI algorithms can address this issue and present practical assistance for the patients. Radiologist's wide experience in this area plays a key role; in radiology, the AI equipment may help to gain a precise diagnosis [39]. Besides, AI programs may be valuable in reducing the disadvantages such as an inadequate quantity of possible RT-PCR test kits and test costs. In this paper, we proposed a fast and accurate deep learning model for the detection of COVID-19 from X-ray images. Notwithstanding overlapping features, the proposed intelligent network can classify COVID-19 and normal cases into different classes. Early detection in the case of COVID-19 plays an essential role in the handling of the disease. Currently, the detection is made for patients by a set of laboratory experiments. These experiments are a time-consuming process and need commercial kits. Also, researchers affirm that merging laboratory outcomes with clinical image characteristics can be used in the early detection of COVID-19 [[40], [41], [42]]. Radiologic data, such as images collected from COVID-19 cases, include valuable data for diagnostic and treatment. Some analyses have found differences in chest X-ray and CT images since the onset of COVID-19 symptoms [43]. Hence, alternative methods of detection are essential in the case of COVID-19. The Inception network is known as one of the Deep Convolutional Neural Network (DCNN) models with high capacity and high performance in the classification task. The primary Inception model was one of the most popular learning architectures for image processing tasks [44]. The proposed model in this research is based on the Inception structure. However, our innovation is that the middle layers change fundamentally, which is a new robust classification model. 1 Accordingly, this study will present a new robust framework of deep learning classifiers as a high-level assistive implement to help the radiologists. It can automatically detect COVID-19 in X-ray images in clinics and hospitals. The overall contributions of this paper are reviewed in the following: Designing a new framework based on deep learning assists of COVID-19 detection. Analyzing the suggested classifier architecture to classify COVID-19 virus via using chest X-ray images to achieve a lower cost process and prediction time. Reporting the results of the proposed DCNN classifier model. The proposed framework supports researchers in developing high-level artificial intelligence methods for CAD devices to prevent the spread of the COVID-19 virus. The rest of this paper is organized as follows: Section 2 gives a review of related works. System architecture, which contains the proposed model, is presented in section 3. We analyze the results of the suggested algorithm in section 4. The conclusion and discussion are given in the last section.

Related works

Many types of researches focus on COVID-19 diagnosis. In most studies, deep neural network methods are applied to chest X-ray models to recognize contagious and accurate results. A DCNN model based on pre-trained transfer architectures (ResNet50 [45], InceptionV3 [44], and Inception-ResNetV2 [46]) that can predict the coronavirus infection from chest X-ray images, is proposed in Ref. [47]. The presented CNN based model showed excellent prediction performance and accuracy from a small sample of X-ray images classified into two classes, COVID-19 and normal cases. Moreover, to dominate the inadequate information and training time, a transfer learning manner is used by the ImageNet dataset [48]. The outcomes confirmed the superiority accuracy in pairs of the training and testing grade of the ResNet50 model. Abbas et al. proposed a new CNN model based on class decomposition and transfer learning to get better performance of the X-ray image classification on pre-trained models [49]. The presented structure is called DeTraC and includes three stages. In the first stage, local feature extraction is performed by using ImageNet pre-trained CNN. In the second stage, a stochastic gradient descent (SGD) optimization approach is utilized for training. Finally, the class combination layer is adjusted for the image's ultimate classification by employing error-correction criteria related to a softmax layer. The ResNet18 [45] pre-trained ImageNet architecture is applied, and the results show an accuracy of 95.12% on chest X-ray images. Wand et al. introduced a DCNN called COVID-Net, which can detect COVID-19 patients from CXR images [50]. This architecture design includes two parts: a human-machine collaborative design approach and a machine-driven design exploration part. A lightweight residual projection-expansion-projection-extension (PEPX) design pattern is used in this architecture. Also, an explain ability-driven examination is executed for decision validation. The model results presented high sensitivity and precision of 87.1% and 96.4%, respectively, for the COVID-19 cases. Hemdan et al. proposed a structure to classifying COVID-19 disease from chest X-ray images containing seven image classifiers named COVIDX-Net with high performance and accuracy of 90% for the DenseNet201 and VGG19 classifiers [51]. Hassanien et al. presented a classification method to detect COVID-19 in lung X-ray images that utilize multi-level thresholding and a SVM [52]. Their algorithm was tested on 40 contrast-enhanced lung X-ray images. Their classification method gained a sensitivity of 95.76%, a specificity of 99.7%, and an accuracy of 97.48%. Zhang et al. proposed a novel deep anomaly detection method formed on CXR images for the screening of COVID-19 [53]. This proposed network includes some parts. The high-level features of images were extracted and then used as the input. The classification part is utilized for the image classification and is composed of a hidden convolutional layer of 100-neurons. The scalar anomaly scores are generated in the anomaly detection part (COVID-19 cases). The suggested architecture attained the false positive rate. The results illustrated a specificity and sensitivity of 70.65% and 96.00%, respectively.

Materials and methods

Data set

In this study, we have used a publicly available dataset that exists in the GitHub repository,2 which has been collected by Ref. [54]. This dataset includes X-ray images of patients infected with COVID-19, SARS, Pneumocystis, and other types of pneumonia. We just analyzed the 940 COVID-19 and Non-COVID-19 X-ray cases and divided them into two parts, 80% for the training and 20% for testing the model. In Fig. 2 some examples of the dataset are shown. The statistical characteristics (mean, minimum, and maximum) of the dataset's images are tabulated in Table 1 .
Fig. 2

Some examples from dataset. The first row are normal cases, and the second are COVID-19 cases.

Table 1

The statistical mean, minimum, and maximum of width and height of the dataset's classes.

TypeImages countMin WidthMax widthMin heightMax height
Covid43513743001564300
Non-Covid505224224224224
Mean180.522621902262
Some examples from dataset. The first row are normal cases, and the second are COVID-19 cases. The statistical mean, minimum, and maximum of width and height of the dataset's classes.

Data augmentation

Data augmentation is a popular operation applied in deep learning, which enlargement the amount of available training/testing data [55]. In this work, due to the lack of the required number of available samples, data augmentation techniques were accomplished by ImageDataGenerator in the training task. The transformations function applied random rotation in the range of 20°, width and height shift operations in the range of 0.2 pixels, zooming operation in the range of 0.8 and 1.2, and horizontal flips. ImageDataGenerator has finally produced 3760 images for the training phase of the proposed model. Data augmentation improves and enhances the network's ability to learn. Data augmentation is an effective method to prohibit network overfitting by increasing the number of training samples [56].

Depthwise separable convolutions layers

CNN's can automatically extract features, and cause the network to accomplish processing with high performance. Due to the high number of parameters, CNN has some restrictions on the computation cost, and the network overfitting. These limitations of CNN can be improved by using depthwise separable convolution that has been used in a neural network as early as 2014 [57]. Depthwise separable convolutions decrease computation cost, time, and the number of parameters. They use convolutional neural processes while improving the representational performance. Depthwise separable convolution layers are decreasing the number of parameters needed to operate at a given step, and they are more successful in image classification algorithms in achieving superior models. Standard convolution executes the channel-wise and spatial-wise computation in one step. However, our proposed network is based on depthwise separable convolution, a standard convolution layer divided into a depthwise convolution and a pointwise convolution. Depthwise separable convolution layer divided the computation into two steps: depthwise convolution and pointwise convolution. Depthwise convolution uses a convolutional filter per input channel, and pointwise convolution is utilized to mix the resulting output channels. Fig. 3 shows the comparison of standard convolution and depthwise separable convolution.
Fig. 3

(a) Standard CNN. (b) Depthwise Separable CNN. In depthwise separable convolution, standard convolutional layers divided into two different levels. Depthwise convolution executes convolution in a single depth slice, while pointwise Convolution merges the information over the entire depth.

(a) Standard CNN. (b) Depthwise Separable CNN. In depthwise separable convolution, standard convolutional layers divided into two different levels. Depthwise convolution executes convolution in a single depth slice, while pointwise Convolution merges the information over the entire depth. Suppose that K is a standard convolutional filter which size is , and F is an input feature map which size is . By applying a standard convolution on an input feature map produces an output feature map O which size is , In depthwise separable convolution, this computation is divided into two levels. The first level uses a depthwise convolution to input channel, The second level exerts pointwise convolution , in order to merge the total output, Pointwise convolution and depthwise convolution have distinct tasks in producing new features: the first one is used for obtaining channel-wise correlations, while the second one is used for obtaining spatial correlations.

Activation function

Rectified Linear Unit (ReLU) is a non-linear activation function that is utilized in neural networks models, and is defined as:where x is an input parameter. As stated in Eq (4), the maximum value between zero and function input is the ReLU output. When the input value is negative the output is equivalent to zero, and when the input is positive, the outcome is x. Accordingly, the Eq (4) readdress as follows: In our proposed model, ReLU is used after each depthwise separable convolution layer. Relu is not a vanishing gradient, more computationally efficient, and learn several times faster [26].

Proposed model

In this section, we present the Fast COVID-19 Detector (FCOD) model as a deep learning model for automatic classification of COVID-19 in 2D X-ray images. FCOD has used depthwise separable convolution layers, which is shown in Fig. 4 .
Fig. 4

Scheme of the proposed classifier architecture.

Scheme of the proposed classifier architecture. In the proposed model, we have been using fewer layers and filters in contrast to the basic architectures. Also, we have applied depthwise separable convolution layers instead of convolutional layers. The model consists of three parts. In the first part, the chest X-ray images were used for the input layer of the model. These images pass over four depthwise separable convolution layers, and then the max-pooling layer follows each of the depthwise separable convolution layers. The middle part merged with 12 depthwise separable convolution layers. Both of the first and middle parts are followed by batch normalization. Also, batch normalization enables significant gradients, and the results appear in faster convergence [58]. The dropout layer is utilized after 16 separable convolution layers with a dropout ratio of 0.2 [59]. This network ends with Global Average Pooling, a layer of depthwise separable convolution, two fully-connected (FC) layers, and a softmax activation function.

Model training

By updating the weights of the network, the loss function, which is a categorical cross-entropy can optimize the algorithm. Categorical cross-entropy is utilized for multi-class classification task to predict the probabilities over the N number of classes. The RMSprop optimization algorithm is used for training the model [60]. We have set 100 epochs, 16 batch sizes, and a learning rate of 0.001 for training our network.

Classification performance

For analyzing the performance of each model, different metrics have been used. As shown in Fig. 5 , we have used a confusion matrix for analyzing the performances of the proposed network.
Fig. 5

Confusion matrix.

Confusion matrix. This matrix has four expected parameters as follows: TP, TN, FP, FN which refer to the true positive, true negative, false positive, and false negative samples of any class. Accuracy is generally utilized as a classification metric and represents how well a classification model can distinguish the classes in the test set. The accuracy is defined by: As shown in Eq (6), the accuracy can be explained as the ratio of the predicted correct labels of the structure over the total number of labels. Precision (Eq (7)) is defined as the ratio of predicted correct labels of the architecture over the whole number of actual labels. Recall (Eq (8)) is defined as the proportion of predicted correct labels of the model over the total number of predicted labels. Measuring the proportion of negatives which are recognized without error, is done by specificity (Eq (9)). Also, F1 - score (Eq (10)) is referred to the harmonic average of the Precision and Recall.

Experimental results

In this section, we show the performance of the proposed algorithm using various evaluation metrics. The model is implemented in Python using Keras framework 2.2.4 with Tensorflow 1.14.0 as the backend, and the hardware configuration is based on free Google service.3 The Receiver Operating Characteristics (ROC) curve and the confusion matrix of the proposed model are shown in Fig. 6 (a) and 6(b), respectively.
Fig. 6

ROC curves and confusion matrix of the proposed deep learning model.

ROC curves and confusion matrix of the proposed deep learning model. The training and testing accuracy and loss curves of the proposed classification model are demonstrated in Fig. 7 . As shown in this Figure, the loss value of the training and testing phases has decreased. On the other hand, the accuracy of the training and testing sets have increased. Accordingly, these issues indicate that the proposed architecture has appropriate efficiency.
Fig. 7

Accuracy and loss curves of the training and testing phases of the proposed deep learning model.

Accuracy and loss curves of the training and testing phases of the proposed deep learning model. The performance metrics of some popular deep learning classifier models are tabulated in Table 2 .
Table 2

Comparative classification performance of deep learning models.

ClassifierSensitivitySpecificityPrecisionAccuracyF1-ScoreTraining time(s)Testing time(s)
COVID-Net [61]0.900.800.850.22
COVID-CAPS [62]0.900.950.95
Shashank [63]0980.910.960.960.97
VGG19 [51]0.831.001.000.900.8126414.0
CovidGAN [64]0.950.940.900.940.92
ResNet50 [47]1001000.980.98
DenseNet201 [51]0.831.001.000.900.8121226.00
ResNetV2 [51]1.000.620.400.700.5710862.00
InceptionV3 [51]0.500.000.5011212.00
Inception
-ResNetV2 [51]1.000.710.600.800.7519886.00
Xception [51]1.000.710.600.800.7520353.00
MobileNetV2 [51]1.000.550.200.600.333891.00
Proposed Model0.930.970.970.960.9618000.014
Comparative classification performance of deep learning models. According to Table 2, the tested DCNN image classifier model's accuracy and computational times are compared with the other models. By using the GPU's robust abilities with a small X-ray image dataset, the proposed deep learning model's running time is low. Also, the overall accuracy is 96%, and the prediction time is 0.014 s, which is one of the best prediction times among the classifier models in this task. Eventually, this model can be used as an online/real-time framework with high accuracy and low prediction time. The performance of FCOD for some randomly selected patients is shown in Fig. 8 . FCOD can detect COVID-19 cases quickly with high accuracy.
Fig. 8

Some examples from final output of the proposed model.

Some examples from final output of the proposed model.

Discussion and conclusion

Since the excessive infection rates and mortality of COVID-19 cases have been increasing, there is a need to use early, fast, and inexpensive diagnostic methods. Unfortunately, there is still no specific drug or approved vaccine to treat COVID-19 infected patients, while fast diagnosis and screening of the disease are two basic and essential issues. Nowadays, most laboratories are equipped to detect the virus gaining the otorhinolaryngological methods. The availability of testing devices and kits is the limitation of the COVID-19 tests. X-ray machines are more simply and quickly available. Hence, the use of chest X-ray images for analysis of the Coronavirus will be assistive. To address this issue, we proposed a novel architecture called FCOD, which has trained, tested, and validated on chest X-ray images for fast detection of COVID-19. This network can be efficiently utilized to help doctors in the analysis of the X-ray images. The proposed algorithm's innovation has a lower prediction time than the other existing models with high performance. The network automatically recognizes the complicated patterns from X-ray images, which is comparable with skilled radiologists. Furthermore, we can improve our network's performance by adding more COVID cases of chest X-ray images in our training data. The proposed classification method was tested and evaluated on covid-chestxray-dataset, which is a publicly available dataset. For comparison, we report the results with the other popular existing methods. The proposed network has illustrated essential improvements over the other methods, especially in prediction time. The proposed algorithm can perform detection of the COVID-19 in 0.014 s. The obtained accuracy and AUC for the proposed model are 96%, and 95%, respectively. From what has been discussed above, it can be concluded that the proposed classification algorithm can be used as an accurate and fast method to detect COVID-19 using chest X-ray images. This model can be used as an online/real-time medical assistive implement in radiologic clinics and hospitals.

Declaration of competing interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.
  10 in total

1.  Fast and Efficient Method for Optical Coherence Tomography Images Classification Using Deep Learning Approach.

Authors:  Rouhollah Kian Ara; Andrzej Matiolański; Andrzej Dziech; Remigiusz Baran; Paweł Domin; Adam Wieczorkiewicz
Journal:  Sensors (Basel)       Date:  2022-06-21       Impact factor: 3.847

2.  X-Ray Equipped with Artificial Intelligence: Changing the COVID-19 Diagnostic Paradigm during the Pandemic.

Authors:  Mustafa Ghaderzadeh; Mehrad Aria; Farkhondeh Asadi
Journal:  Biomed Res Int       Date:  2021-08-22       Impact factor: 3.411

3.  A Robust Framework for Epidemic Analysis, Prediction and Detection of COVID-19.

Authors:  Farman Hassan; Saleh Albahli; Ali Javed; Aun Irtaza
Journal:  Front Public Health       Date:  2022-05-06

Review 4.  Review of Current COVID-19 Diagnostics and Opportunities for Further Development.

Authors:  Yan Mardian; Herman Kosasih; Muhammad Karyana; Aaron Neal; Chuen-Yen Lau
Journal:  Front Med (Lausanne)       Date:  2021-05-07

Review 5.  A review of deep learning-based detection methods for COVID-19.

Authors:  Nandhini Subramanian; Omar Elharrouss; Somaya Al-Maadeed; Muhammed Chowdhury
Journal:  Comput Biol Med       Date:  2022-01-29       Impact factor: 4.589

Review 6.  Human Body Performance with COVID-19 Affectation According to Virus Specification Based on Biosensor Techniques.

Authors:  Mohammed Jawad Ahmed Alathari; Yousif Al Mashhadany; Mohd Hadri Hafiz Mokhtar; Norhafizah Burham; Mohd Saiful Dzulkefly Bin Zan; Ahmad Ashrif A Bakar; Norhana Arsad
Journal:  Sensors (Basel)       Date:  2021-12-15       Impact factor: 3.576

Review 7.  Role of Artificial Intelligence in COVID-19 Detection.

Authors:  Anjan Gudigar; U Raghavendra; Sneha Nayak; Chui Ping Ooi; Wai Yee Chan; Mokshagna Rohit Gangavarapu; Chinmay Dharmik; Jyothi Samanth; Nahrizul Adib Kadri; Khairunnisa Hasikin; Prabal Datta Barua; Subrata Chakraborty; Edward J Ciaccio; U Rajendra Acharya
Journal:  Sensors (Basel)       Date:  2021-12-01       Impact factor: 3.576

8.  A Hybrid Convolutional Neural Network Model for Diagnosis of COVID-19 Using Chest X-ray Images.

Authors:  Prabhjot Kaur; Shilpi Harnal; Rajeev Tiwari; Fahd S Alharithi; Ahmed H Almulihi; Irene Delgado Noya; Nitin Goyal
Journal:  Int J Environ Res Public Health       Date:  2021-11-20       Impact factor: 3.390

9.  Detection of COVID-19 using deep learning techniques and classification methods.

Authors:  Çinare Oğuz; Mete Yağanoğlu
Journal:  Inf Process Manag       Date:  2022-07-08       Impact factor: 7.466

Review 10.  Application of Deep Learning Techniques in Diagnosis of Covid-19 (Coronavirus): A Systematic Review.

Authors:  Yogesh H Bhosale; K Sridhar Patnaik
Journal:  Neural Process Lett       Date:  2022-09-16       Impact factor: 2.565

  10 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.