Literature DB >> 34345248

A novel deep learning based method for COVID-19 detection from CT image.

SeyyedMohammad JavadiMoghaddam1, Hossain Gholamalinejad2.   

Abstract

The novel Coronavirus named COVID-19 that World Health Organization (WHO) announced as a pandemic rapidly spread worldwide. Fast diagnosis of the virus infection is critical to prevent further spread of the virus, help identify the infected population, and cure the patients. Due to the increasing rate of infection and the limitations of the diagnosis kit, auxiliary detection tools are needed. Recent studies show that a deep learning model that comes up with the salient information of CT images can aid in the COVID-19 diagnosis. This study proposes a novel deep learning structure that the pooling layer of this model is a combination of pooling and the Squeeze Excitation Block (SE-block) layer. The proposed model uses Batch Normalization and Mish Function to optimize convergence time and performance of COVID-19 diagnosis. A dataset of two public hospitals was used to evaluate the proposed model. Moreover, it was compared to some different popular deep neural networks (DNN). The results expressed an accuracy of 99.03 with a recognition time of test mode of 0.069 ms in graphics processing unit (GPU). Furthermore, the best network results in classification metrics parameters and real-time applications belong to the proposed model.
© 2021 Elsevier Ltd. All rights reserved.

Entities:  

Keywords:  Batch normalization; COVID-19 detection method; Deep learning model; Disease diagnosis; Mish function

Year:  2021        PMID: 34345248      PMCID: PMC8318781          DOI: 10.1016/j.bspc.2021.102987

Source DB:  PubMed          Journal:  Biomed Signal Process Control        ISSN: 1746-8094            Impact factor:   3.880


Introduction

In the year 2020, a new coronavirus, called COVID-19, was marked as the cause of a pandemic disease by the WHO that first spread in China [1]. COVID-19 is a highly contagious Severe Acute Respiratory Syndrome (SARS). The number of confirmed coronavirus cases by the 27th of October 2020 was 43,341,451, of which 29,300,000 were recovered, and 1,157,509 of the infected died [1]. The standard method to detect COVID-19 cases is the Reverse Transcription-Polymerase Chain Reaction (RT-PCR). However, this detection method has some limitations such as shortage of kits, relentlessness, manual methodology, and is time-consuming. Moreover, the positive rate of the RT-PCR test is only 63% [2]. Recently, many researchers indicate that the abnormalities in the chest radiography images of patients and the computerized tomography (CT) scan are different diagnostic approaches for COVID-19 detection [3], [4], [5], [6]. These studies show that visual symptoms in the lugs are different in patients. CT images have some advantages, such as rapid triage of suspected COVID-19 patients, lower risk of transmission, and high availability. Some researchers try to develop a method for COVID-19 detection using CT images [5], [6]. Deep learning-based methods can extract the features of images that are not obvious in the original image [7]. Many studies have applied a Convolution Neural Network (CNN) to detect COVID-19 [8], [9], [10]. Ohata [11] et al. combined the CNNs with some techniques based on machine learning, such as Bayes. They showed the proposed architecture using a support vector machine (SVM) and a linear kernel with an F1-score of 98.5%. Researchers [12] suggested a method using the CNN model and class decomposition. Abraham [13] proposed a model using a multi-CNN and correlation feature selection method. They achieved an accuracy of 97.44%. Several studies have been suggested inserting a preprocessing phase using the Visual Geometry Group (VGG) model to modality the images [14], [15], [16]. The precision of these methods is up to 86%. Azemin et al. [17] presented an approach based on ResNet-101 CNN architecture. The accuracy of the method is 77.3%. The model of Mangal [18] includes a pre-train layer with a 121-layer dense convolutional network and a full-connected layer with a 90.5% accuracy. Many approaches use transfer learning with CNN architecture [19], [20], [21]. Loey introduced a method based on Generative Adversarial Network (GAN) with deep transfer learning. He used three deep transfer techniques and achieved 85.2% in testing accuracy [22]. Elaziz et al. [23] introduced a parallel framework for automatic COVID-19 diagnosis. This model uses a fraction multi-channel for feature selection. Some researchers proposed COIVD-19 detection methods for multi-class classification [24]. Khan [25] suggested a model using the deep CNN structure to COVID-19 infection diagnosis. This method uses an architecture based on Xception. The classification accuracy of the approach is 95% for the 3-class model. Karim [26] proposed to use human explanations and a deep learning model for prediction. Yoo et al. [27] have combined three binary decision trees with a CNN architecture based on the PyTorch frame. The maximum accuracy was 98% for the first decision tree. In common deep learning classification structures, the feature extraction process is performed only in the convolution layer. The polishing layer only plays the role of minimizing the dimensions of the feature maps and the network. This paper proposed a new pooling layer that performs reducing the network dimensions and feature extraction simultaneously. The feature extraction is performed using Haar wavelet [28]. Moreover, the model introduces a network structure based on Batch Normalization (BN) and Mish Function[29] to reduce the convergence time and achieve better performance.

Material and methods

CT scan image dataset

This paper has used a CT image dataset [30] to evaluate the proposed model. This dataset was gathered from two Union Hospital (HUST-UH) and Liyuan Hospital (HUST-LH) [31]. The individual CT images have been classified into three categories: The first category includes 5705 non-informative CT (NiCT) images without lung parenchyma, The second category includes 4001 positive CT (pCT) images with features related to COVID-19 pneumonia, and The third category includes 9979 negative CT (nCT) images with irrelevant features to COVID-19 pneumonia. Fig. 1 is a sample image of the dataset.
Fig. 1

A Sample image from covid-19 CT dataset.

A Sample image from covid-19 CT dataset.

The proposed model

Fig. 2 shows the proposed model in which the feature extraction process is performed in three layers:
Fig. 2

Proposed model architecture.

In the convolution layer, in the proposed polishing layer, and in the SE block layer, how to extract a feature in convolution layers similar to other structures. The proposed pooling layer is prepared using HAAR wavelet filters. In this layer, minimizing the dimensions and feature extraction are performed using wavelet HAAR simultaneously. The proposed method extracts better features than other classifying CT scan image data methods, as the test results show. Proposed model architecture. The proposed model includes a novel network structure in which batch normalization (BN) [32] is used to shorten the convergence time and achieve better performance. The activation function is Mish Function [33] to improve the classification capacity in nonlinear cases. The dropout layer with is used. Finally, an SE block is added after each dropout layer. The pooling layer is the Haar wavelet transform layer [28] to produce better features. Table 1 shows the network’s configuration in detail. In the first stages, there are some convolutional layers so that the size of their kernel is according to table 2 . Then, the SE block is embedded after any activation function. Finally, the features are fed into the Fully-Connected and SoftMax layers. Fig. 3 depicts the block diagram of the proposed network. Because the proposed model contains Wavelet and four convolution layers, it is named Wavelet CNN-4 (WCNN4).
Table 1

Configuration of the proposed network.

layerOutput sizeKernelStride
Image input256×256×3
Conv256×256×325×5×3×321
Mish + Pool + BN + Drop out (0.5) + SEBlock128×128×322
Conv128×128×325×5×3×321
Mish + Pool + BN + Drop out (0.5) + SEBlock64×64×322
Conv64×64×645×5×3×641
Mish + Pool + BN + Drop out (0.5) + SEBlock32×32×642
Conv32×32×1285×5×3×1281
Mish + Pool + BN + Drop out (0.5) + SEBlock16×16×1282
Conv16×16×2565×5×3×2561
Fully connected65,536×256
Fully connected256× (number of classes)
SoftMax(number of classes) ×1
Table 2

Training parameters.

ParameterValue
Batch size64
Epochs50
Momentum rate0.9
Learning rate0.01
Weight decay1e-3
Epsilon1e-10
SamplerWeighted random sampler
Fig. 3

Block diagram of proposed network.

Configuration of the proposed network. Training parameters. Block diagram of proposed network.

Experimental results

The experiments have been run on a PC with GeForce Turbo RTX-2080 GPU and Corei3-9100f CPU running at 3600 MHz. The implementation of the model was in Python 3.7 using the Pytorch library. All experiments have been done in GPU. Because of the difference in class numbers in any dataset, the weighted random sampler [34] was used as a Sampler in the training step. Table 2 shows all training parameters. The proposed Deep Neural network was trained using Covid-19 CT Scan images. The training phase used eight different optimizers. Table3 shows the metric results of the proposed approach. This table expresses the best results in metrics achieved using the RAdam optimizer.
Table 3

Metric results of the proposed network.

Test Cohen kappa scoreTest mean precisionTest mean recalllossTest accuracyoptimizer
95.3398.0695.480.147197.15SGDM [35]
96.6898.4296.860.086997.97GC-SGDM [36]
95.1797.9895.590.095897.05Adam [37]
95.1698.1095.730.071597.06GC-Adam [36]
93.7496.9994.040.099396.19NAdam [38]
72.2284.0380.990.378082.26GC-NAdam [36]
98.4398.7198.910.033899.03RAdam [39]
96.2798.1897.020.079097.71GC-RAdam [36]
Metric results of the proposed network.

Comparison with popular DNNs

Fig. 4 depicts a comparison between the proposed structure and the standard structure using the most common pooling called max-pooling [40]. The result shows there is not much difference in the extracted features in the network's initial layers. However, in the final layer, the difference between the extracted features is clear.
Fig. 4

Comparison of max pooling and the proposed pooling layers.

Comparison of max pooling and the proposed pooling layers. To further evaluate, some popular DNNs, including VGG [41], ResNet [42], and Inception [43], have been trained on the test dataset. VGG architecture is one of the first popular structures that produced outstanding results in the ImageNet large-scale visual recognition challenge (ILSVRC-2014) competition [44]. Its original structure cannot be trained from scratch due to the vanishing gradients problem [45]. However, today, with batch normalization, its modified version can be trained from scratch. For comparison, VGG11 with batch normalization has been trained on the test dataset. The second structure is ResNet. It has a novel structure, and having the new residual connections, can be trained on every dataset. It supports very deep layers and exists in different versions from ResNet18 to ResNet152 and even more. To compare the structures, ResNet18 and ResNet50 have been trained on the test dataset. Inception is another network that concatenates the sparse layers to make dense layers [46]. This structure reduces dimension to achieve more efficient computation and deeper networks as well as overfitting. Inception architecture takes multiple kernel filter sizes in a convolutional neural network. Table 4 presents the classification metric parameters for popular DNNs and the proposed network. Moreover, Fig. 5, Fig. 6 depict the confusion matrix for the proposed network and inception-V3. The accuracy and loss diagrams in the training phase of the tests are presented in Fig. 7, Fig. 8 . The final relationship of these networks is similar to each other. In the next step, the accuracy and loss for test data were measured for each epoch. The results depict that the proposed network has the best performance (see Fig. 9, Fig. 10 ).
Table 4

Comparison of the proposed method with some popular DNNs. Prediction times do not include image loading.

ModelNumber of ParametersPredictiontime GPUAccuracyLossCohen kappa scoreMean precisionmean recall
Popular DNNsVGG11 + BN128,792,3255.186 ms92.730.940287.8795.3589.42
ResNet1811,179,0774.262 ms93.390.959488.9895.7989.49
ResNet5023,518,2775.554 ms95.980.569793.4097.0193.98
Inception-v324,351,7196.844 ms98.110.160496.9498.4997.61
Proposed networkWCNN44,610,5310.069 ms99.030.033898.4398.7198.91
Fig. 5

Confusion matrix of Inception network.

Fig. 6

Confusion matrix of proposed network.

Fig. 7

Accuracy in the train phase for different DNNs.

Fig. 8

Loss in train phase for different DNNs.

Fig. 9

Accuracy in the train phase with test data for different DNNs.

Fig. 10

Loss in train phase with test data for different DNNs.

Comparison of the proposed method with some popular DNNs. Prediction times do not include image loading. Confusion matrix of Inception network. Confusion matrix of proposed network. Accuracy in the train phase for different DNNs. Loss in train phase for different DNNs. Accuracy in the train phase with test data for different DNNs. Loss in train phase with test data for different DNNs.

Discussion

This work proposes a deep model based on batch normalization and Mish function to detect COVID-19 cases from CT images. Two real datasets were used to evaluate the proposed model. The metric results of the proposed network showed an accuracy of 99.03. This study tries to introduce a network with kernel-based machine learning applications [47]. Therefore, some networks have been selected for the test phase suitable for real-time and online applications. According to Table 4, the worst result in popular DNNs belongs to VGG11 with a 92.73% accuracy. This network is too deep and has 128 million parameters. The best performing DNN was Inception-v3 with a 98.11% accuracy. As shown in Table 4, the best recognition time in popular DNNs is achieved using Resnet18 and is 4.262 ms. This network has more than 11 million trainable parameters and has about 93.39% accuracy on the test dataset. According to the Cohen Kappa score rate for Resnet18, it is about 88.98%; the classification is not good. Although all of the popular DNNs in our work are powerful networks in classification problems for RGB images [48], they do not have good CT image behavior. Comparing to these networks, the proposed model has four novel innovations; using wavelet [49] for pooling the feature maps, Mish [33] activation Function, Mini-Batch Normalization [32], and SE-blocks [50]. Using wavelet transform into network structure is a reason for extracting the correct and appropriate information for recognition tasks. Mish Function is a novel activation function used in powerful detection and classification networks, such as yolov4 [51], which has an essential role in the appropriate information extraction. One of the essential parts of the proposed model is using mini-batch normalization in any layer. Squeeze-and-Excitation blocks (SE block) is an architectural unit that can be plugged into a CNN structure to improve performance with only a slight increase in the total number of parameters. Squeeze-and-Excitation blocks explicitly model channel relationships and channel interdependencies and include a form of self-attention on channels. Consequently, it results in the best performance in classification metrics parameters and recognition time. Furthermore, the number of trainable parameters is more suitable for CT images than more deep structures that suffer from overfitting. The proposed network is the best in classification metric parameters comparing to the popular DNNs. Moreover, WCNN4 has the least prediction time. Therefore, it is a helpful network for real-time application for the recognition of covid-19 using CT images. Fig. 5, Fig. 6, Fig. 7, Fig. 8 show the networks are similar in the final Epochs in the training phase. However, there are differences between them in the network test phase. This difference is generally due to the complexity of the network structure. Bigger structures need more data in the training phase. As shown in Fig. 7 and Fig. 8, vgg11 has the most considerable accuracy entropy in the training phase, and the proposed network has a smooth curve. Fig. 9 and Fig. 10 are the accuracy and loss for test data after each epoch. As shown, the suggested model has decreasing trend entropy in loss and accuracy. Finally, an important achievement of the proposed network's evaluation is that the test mode's recognition time was 0.069 ms in GPU.

Conclusion

Rapid and accurate diagnosis is necessary as the number of patients with COVID-19 infection is increasing. This work proposes a deep learning-based model to detect COVID-19 disease from CT images. The model is an automated method without any feature extraction phase. The proposed model has four novel innovations, including wavelet for pooling the feature maps, Mish activation function, mini-batch normalization, and SE-blocks. For evaluation, the model was trained and tested on the datasets of two public hospitals. The results show the suggested model achieved excellent metric parameters in classifying COVID-19 cases, such as accuracy of 99.03. Moreover, it decreases trend entropy in loss and accuracy. Finally, the recognition time in test mode was 0.069 ms in GPU. The experimental results show the proposed model is a helpful network for real-time application for recognition of covid-19 using CT images.

Declaration of Competing Interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.
  15 in total

1.  Detection of SARS-CoV-2 in Different Types of Clinical Specimens.

Authors:  Wenling Wang; Yanli Xu; Ruqin Gao; Roujian Lu; Kai Han; Guizhen Wu; Wenjie Tan
Journal:  JAMA       Date:  2020-05-12       Impact factor: 56.272

2.  Chest CT Findings in Coronavirus Disease-19 (COVID-19): Relationship to Duration of Infection.

Authors:  Adam Bernheim; Xueyan Mei; Mingqian Huang; Yang Yang; Zahi A Fayad; Ning Zhang; Kaiyue Diao; Bin Lin; Xiqi Zhu; Kunwei Li; Shaolin Li; Hong Shan; Adam Jacobi; Michael Chung
Journal:  Radiology       Date:  2020-02-20       Impact factor: 11.105

3.  Automated detection of COVID-19 cases using deep neural networks with X-ray images.

Authors:  Tulin Ozturk; Muhammed Talo; Eylul Azra Yildirim; Ulas Baran Baloglu; Ozal Yildirim; U Rajendra Acharya
Journal:  Comput Biol Med       Date:  2020-04-28       Impact factor: 4.589

4.  Detection of COVID-19 from Chest X-Ray Images Using Convolutional Neural Networks.

Authors:  Boran Sekeroglu; Ilker Ozsahin
Journal:  SLAS Technol       Date:  2020-09-18       Impact factor: 3.047

5.  Transfer Learning to Detect COVID-19 Automatically from X-Ray Images Using Convolutional Neural Networks.

Authors:  Mundher Mohammed Taresh; Ningbo Zhu; Talal Ahmed Ali Ali; Asaad Shakir Hameed; Modhi Lafta Mutar
Journal:  Int J Biomed Imaging       Date:  2021-05-15

6.  Using X-ray images and deep learning for automated detection of coronavirus disease.

Authors:  Khalid El Asnaoui; Youness Chawki
Journal:  J Biomol Struct Dyn       Date:  2020-05-22

7.  CoroNet: A deep neural network for detection and diagnosis of COVID-19 from chest x-ray images.

Authors:  Asif Iqbal Khan; Junaid Latief Shah; Mohammad Mudasir Bhat
Journal:  Comput Methods Programs Biomed       Date:  2020-06-05       Impact factor: 5.428

8.  Covid-19: automatic detection from X-ray images utilizing transfer learning with convolutional neural networks.

Authors:  Ioannis D Apostolopoulos; Tzani A Mpesiana
Journal:  Phys Eng Sci Med       Date:  2020-04-03
View more
  6 in total

1.  Multi-task semantic segmentation of CT images for COVID-19 infections using DeepLabV3+ based on dilated residual network.

Authors:  Hasan Polat
Journal:  Phys Eng Sci Med       Date:  2022-03-14

2.  A lightweight CNN-based network on COVID-19 detection using X-ray and CT images.

Authors:  Mei-Ling Huang; Yu-Chieh Liao
Journal:  Comput Biol Med       Date:  2022-05-11       Impact factor: 6.698

3.  A Deep Learning and Handcrafted Based Computationally Intelligent Technique for Effective COVID-19 Detection from X-ray/CT-scan Imaging.

Authors:  Mohammed Habib; Muhammad Ramzan; Sajid Ali Khan
Journal:  J Grid Comput       Date:  2022-07-18       Impact factor: 4.674

4.  A Computational Modeling and Simulation Workflow to Investigate the Impact of Patient-Specific and Device Factors on Hemodynamic Measurements from Non-Invasive Photoplethysmography.

Authors:  Jesse Fine; Michael J McShane; Gerard L Coté; Christopher G Scully
Journal:  Biosensors (Basel)       Date:  2022-08-04

Review 5.  Recent Advances in Non-Invasive Blood Pressure Monitoring and Prediction Using a Machine Learning Approach.

Authors:  Siti Nor Ashikin Ismail; Nazrul Anuar Nayan; Rosmina Jaafar; Zazilah May
Journal:  Sensors (Basel)       Date:  2022-08-18       Impact factor: 3.847

6.  COVID-19 ground-glass opacity segmentation based on fuzzy c-means clustering and improved random walk algorithm.

Authors:  Guowei Wang; Shuli Guo; Lina Han; Zhilei Zhao; Xiaowei Song
Journal:  Biomed Signal Process Control       Date:  2022-09-12       Impact factor: 5.076

  6 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.