Literature DB >> 35607444

Detecting Covid19 and pneumonia from chest X-ray images using deep convolutional neural networks.

Nallamothu Sri Kavya1, Thotapalli Shilpa1, N Veeranjaneyulu1, D Divya Priya2.   

Abstract

With the current COVID19 pandemic, we have to weigh human life, prosperity, and value, while implicitly acknowledging that controlling case spread and mortality is a challenge. Identifying COVID19-infected patients and disconnecting them to avoid COVID transmission is one of the most difficult tasks for clinicians. As a result, figuring out who infected with covid19 is crucial. COVID19 is identified using a 4-6-hour reverse transcription-polymerase chain reaction (RT-PCR). Another way to detect Coronavirus early in the disease process is by using chest X-rays (CXR).We extracted characteristics from chest X-ray images using VGG16 and ResNet50 deep learning algorithms, then classified them into three groups: viral pneumonia, normal, and COVID19. We ran 15,153 images through the models to see how accurate they were in real-world situations. For detecting COVID19 cases, the VGG16 model has an average accuracy of 89.34 %, whereas ResNet50 has an accuracy of 91.39 %. When utilizing deep learning to identify COVID19, however, a larger dataset is necessary. It has the desired effect of detecting situations accurately.
Copyright © 2022 Elsevier Ltd. All rights reserved. Selection and peer-review under responsibility of the scientific committee of the International Conference on Advanced Materials for Innovation and Sustainability.

Entities:  

Keywords:  COVID19; Chest X-rays; Deep Learning; Pneumonia; ResNet50; VGG16

Year:  2022        PMID: 35607444      PMCID: PMC9117408          DOI: 10.1016/j.matpr.2022.05.199

Source DB:  PubMed          Journal:  Mater Today Proc        ISSN: 2214-7853


Introduction

Severe Acute Respiratory Syndrome Coronavirus 2 (SARS CoV-2) is responsible for Covid19 disease [1]. Zoonotic microorganisms are covid19 diseases that make severe pollutions to the organs of breath and started spoiling animals then conveyed from animals to people [2]. The first case of COVID is detected in the Wuhan city of China back in 2019[3].Because of this deadly virus the global pandemic was declared by the World Health Organization (WHO) [3]. Fever, dry cough, and exhaustion are the most common symptoms of COVID19. Now the virus has been spread all over the world. As of 28 February 2022, 446,511,318 people is effected by the COVID19 with deaths 6,004,421. To identify the disease in humans COVID19 reverse transcription-polymerase chain reaction(RT-PCR) is used which is time consuming and has a high positive false rate[1], [2], [4]. As of now the significant learning techniques on clinical imaging are the incredible strategy for dealing with the medical diagnosis systems. Computer-Aided Diagnosis methods such as chest X-ray and Computed Tomography techniques can be used as a complement for RT-PCR [5]. For the diagnosis of lung infections, CT and chest X-ray images are commonly used. The cost and radiation exposure are important issues for the COVID19 diagnosis, though CT images are commonly used. It is preferable to use CXR images over CT images since they are less radiation-exposed and more widely accessible [5]. Therefore, we used the chest x-ray images in the current study to identify COVID19 infected patients using deep learning techniques. Our dataset consists of total of 15,153 images in which 3616 COVID19 images, 1345 are viral Pneumonia images and 10,192 are Normal images.(See Table 1. ).
Table 1

The used model Architecture of VGG1.

Layer(type)Output ShapeParameters
Conv2d(Convo2D)(None,256,256,32)320
Max-pooling2d(None,128,128,32)0
Cov2d_1(Convo2D)(None,128,128,32)9248
Max-pooling2d_1(None,64,64,32)0
Cov2d_2(Convo2D)(None,64,64,64)18,496
Max-pooling2d_2(None,32,32,64)0
Cov2d_3(Convo2D)(None,32,32,64)36,928
Max-pooling2d_3(None,16,16,64)0
Cov2d_4(Convo2D)(None,16,16,128)73,856
Max-pooling2d_4(None,8,8,128)0
Cov2d_5(Convo2D)(None,8,8,128)147,584
Max-pooling2d_5(None,4,4,128)0
Cov2d_6(Convo2D)(None,4,4,128)147,584
Max-pooling2d_6(None,2,2,128)0
Cov2d_7(Convo2D)(None,2,2,256)295,168
Max-pooling2d_7(None,1,1,256)0
Flatten(Flatten)(None,256)0
dense(Dense)(None,128)32,896
dense_1(Dense)(None,64)8256
dense_2(Dense)(None,3)195
The used model Architecture of VGG1. Lungs are the most effected organs in the human body due to corona virus [6]. This not only effects the respiratory system, it also effects the other organs like kidneys and liver. CXR images of COVID19 infected patients have hazy lungs compared to normal healthy people lungs. These features may aid in the detection of COVID19. In recent years deep learning is using to detect many different diseases. Some examples are detection of tumor types in head, brain, lungs, etc. [7]. Many researchers have used deep learning methods and obtained good results. With the development of the pandemic machine learning and deep learning have been used in the detection of COVID19 patients. Coronaviruses are structurally analyzed in order to detect the disease. The use of convolutional neural networks has been widely adopted as a method of detecting and classifying COVID19 [8]. Generally, CNN has a high ability to categorize patients at risk for developing diseases. CNN systems are used across a wide spectrum of classification tasks, from binary classification to multiclass classification. In high-dimensional datasets with multilayer functionalities, CNN has already demonstrated impressive results for discovering convoluted structures. For 2D image processing, CNN uses 2D convolutional layers. A CNN consists of an input layer, an output layer, and hidden layers. The hidden layers include convolution layers, pooling layers, fully connected layers, and regression layer. In this paper, we look at deep learning models VGG16 and ResNet50 that are suitable for classification of images. There are several layers defined in the defined model, and each layer receives the information it needs from all previous layers. On the COVID19 Radiography Dataset, our model is evaluated. Our considered model groups the dataset into three types and produces accurate results.

Related works

In this paper, various deep learning models have been applied to diagnose COVID19 and pneumonia in chest X-ray images, some of which are described below. Li et al. [9] used transfer learning architectures such as CheXNet, DenseNet, VGG19, MobileNet, InceptionV3, ResNet18, ResNet101, and squeezeNet and divides the data into three classes. The accuracy of a model created utilising chest X-rays from 423 images of COVID-19, 1,485 images of viral pneumonia, and 1,575 normal images was 97.94%. Li et al. [10] presented CovXNet as a network architecture for diagnosing bacterial pneumonia, viral pneumonia, and COVID19. There are 1,583 images of normal people in the dataset, 1,493 images of viral pneumonia X-rays, 2980 images of pneumonia X-rays, and 305 images of COVID19 X-rays from various patients. Their model was 89.1% accurate. COVID-Net was developed by Gunraj [11] et al. to help doctors differentiate between COVID19 related and non COVID19 pneumonia. Researchers showed a new dataset of 13,975 chest X-ray scans from 13,870 patients and discovered that their algorithm was 93.3 percent accurate. The 3-class categorization, on the other hand, was simply judged on accuracy in the article. AD3D-MIL, a deep 3D multiple instance learning technique based on attention, was used by Han et al.[12] to distinguish COVID19 pneumonia from other types of viral pneumonia. An analysis of 230 CT scans was performed using data from 79 patients, 100 pneumonia patients, and 130 healthy individuals. Their system, according to their assessment, was 97.9% accurate overall. Rajaraman et al. [13] discovered COVID19 in chest Xray images by using an iteratively pruned deep learning ensemble. In their investigation, they used two models. By applying the transfer learning approach to the first model, it was trained to classify normal and abnormal chest X-rays, while the second model was taught to classify COVID19 and pneumonia cases using the first model's learning weights. They employed an ensemble strategy to improve their model's overall prediction performance and achieved a 99.01 percent accuracy rate. Hammoudi et al. [14] developed personalised models for locating COVID19 aspiratory side effects in the early stages of the drug's development. The dataset that contains two types of Pneumonia i.e, bacterial,viral and standard chest X-ray images was used to train their models. A 2D profound learning system known as the primary track COVID19 characterisation organization (FCONet) was proposed by Ko et al. [15] to analyze COVID19 pneumonia in a chest computed tomography (CT) examination image. For the preparation of the FCONet model, they used an exchange learning technique with best-in-class profound learning models as the spine. ResNet50 FCONet had the best exhibition outcomes for true positive rate, true negative rate, and accuracy on the validation dataset of 99.8 percent, 100 percent, and 99.87 percent, respectively, in all the trained FCONet models. [16] described an algorithm for detecting COVID-19 pneumonia based on CXR pictures.[17] demonstrated a model for estimating COVID-19 disorders using deep learning and laboratory data.A total of 600 patients gave laboratory results for the model to be tested. An optimization approach was utilised to diagnose COVID19 with a hybrid CNN given in  [18].To detect COVID-19 infections, researchers used the Xception architecture to construct CorNet, an image-based deep convolutional neural network[19], [29]. Deep learning was used to diagnose COVID19[20] presented this method.There were several CNN models utilised. Deep learning was used to detect COVID19 on CXR [21].There were three stages. Initially, pneumonia was discovered, then COVID19 and pneumonia were discovered, and finally, the condition was diagnosed.They employed 6523 CXR images and achieved a 97 percent accuracy rate.The authors [22] reported using a patch-based CNN to diagnose COVID19 in CXR pictures with state-of-the-art accuracy. In [23], the authors described a method for identifying diseases from COVID19 images that used COVID19 CXR image descriptors, feed-forward neural networks, and CNNs.CXR and CT images were employed in this study to detect illnesses linked to COVID19 [24], [25], [26], [27], [28].

Proposed method

The strategy for the Identification framework was completed in this work, and it will accurately analyse radiographic images accurately for COVID19 and viral pneumonia.Using chest x-rays to construct a robust system for COVID19 characterisation. Using proposed CNN pre-trained models, classify chest x-ray pictures. Examine and contrast the models' displays with performance indicators. In kaggle, a set of chest X-Ray images was gathered from the COVID19 Radiography database. A total of 15,153 images are included in the dataset, with 3616 COVID19 images, 1345 viral Pneumonia images, and 10,192 Normal images. As a result, data imbalance can be found in the collected data, potentially leading to inaccurate categorization findings. Finally, 1345 photos were chosen for our studies from each category. The key contributions of the study are as follows: To analyse and distinguish covid19 and pneumonia patients, the structure of VGG16 and ResNet50 has been evaluated. To show the training of CNN models with unbalanced data, random sampling with image augmentation is performed. A convolutional neural network model was used to identify COVID19. Performance metrics were used to address the imbalance issue. Fig. 1 depicts the COVID19 and Pneumonia screening structures.(See Fig. 2. )
Fig. 1

Screening Structure of COVID19 and Pneumonia.

Fig. 2

Overall Architecture of the Proposed System.

Screening Structure of COVID19 and Pneumonia. Overall Architecture of the Proposed System. VGG16 and ResNet50 are the models under consideration. CNN was used in the background (Convolutional Neural Networks). Multiple procedures, including as convolutional, max-pooling, dense, and softmax, are performed on CNN in multiple layers.The model's accuracy is entirely dependent on the dataset and its size. Large datasets with more epochs can sometimes result in improved training and validation accuracy. Large datasets with fewer epochs might sometimes produce higher training results but less accuracy in validation.

System architecture

To design the recommended system, the following steps must be taken: An X-ray dataset with pneumonia, COVID19, and normal X-ray pictures has been created for research purposes. Training and validation parts of the dataset were created. Preprocessing is required before resampling, scaling, and enhancing images. We are employing a developed CNN structure to calculate the output of our considered model using chest X-Ray images of pneumonia, covid-19, and normal patients. Output classification. The loss function is calculated by comparing the current output to the desired output. Using the loss function and the training process, adjust the CNN parameters. Steps 3–6 should be repeated for all datasets and epochs. Models Used: Using numerous deep learning networks, COVID-19 was successfully detected. The CNN technique is primarily used for COVID-19 categorization, segmentation, and prediction. We provide a deep learning-based COVID-19 screening in this study, in which the programme employs deep learning algorithms to predict if the imaging of the patient's suspected lungs is normal, if bacterial pneumonia has occurred, or if COVID-19 has occurred. We employed VGG16 and ResNet50 models to conduct multi-class classification on X-Ray pictures, and we applied deep learning approaches to train them. VGG16: In 2014, VGG16 was created. It is one of CNN's finest image categorization models. VGG16 has a total of 16 layers, 13 of which are convolutional and 3 of which are fully connected. The VGG16′s overall architecture is depicted in Fig. 3 . Between the convolutional layers are 2 × 2 max-pooling layers and 13 convolutional layers with 3 × 3 filters. The ReLu activation function is applied between these layers. Finally, the probabilities of each classification are calculated using a softmax function.
Fig. 3

Architecture of the VGG16.

Architecture of the VGG16. ResNet50: An input layer, four subsequent layers, and an output layer make up a ResNet Architecture. Each stage symbolizes a step in the process that we are going through in order. A CNN step is run depending on the inputs from previous stages, and a result is returned. ResNet is broken into five stages, with Stage 0 serving as an input pre-processing stage and Stages 2–4 serving as bottleneck stages. With 64 output channels and a stride of 2, input stems that execute 7–7 convolutions have 64 output channels. We have three to three max pooling layers with a stride of two in the following. This layer essentially reduces the width and height by four times, while increasing the channel width and height by 64. We have a down sampling block and residual blocks in stage 2 and all following stages. In general, residual blocks work similarly to down-sampling blocks, with the distinction that the stride of the convolutions is 1.We get different models when we modify the count of residual blocks, so we'll only say how many convolutional layers we have in ResNet50 and ResNet152. The ResNet50 architecture is depicted in Fig. 4 .(See Fig. 5. ).
Fig. 4

Architecture of the ResNet50.

Fig. 5

An example of chest X-ray images.

Architecture of the ResNet50. An example of chest X-ray images.

Dataset

COVID-19 Radiography Database was used as the dataset. COVID-19, pneumonia, and normal patients chest X-ray images are included in the collection. Related to clinical specialists, a group of scientists from Qatar University, the University of Dhaka in Bangladesh, and teammates from Pakistan and Malaysia created this dataset. A total of 15,153 images are included in the dataset, with 3616 COVID19 images, 1345 viral Pneumonia images, and 10,192 Normal images.

Performance evaluation

Many assessment criteria are used to evaluate the performance of various models in relation to the problem at hand. Some assessment metrics are more appropriate for assessing regression model performance, whereas others are more appropriate for assessing classification model performance. As previously stated, a variety of evaluation measures are available, however the accuracy, recall, precision, and F1 score were used to evaluate the models' performance in this study. Three typical CNN results are displayed for each model: Model accuracy curve Model Loss curve Confusion Matrix

Confusion Matrix

Confusion matrices are tables that describe how well a model performs on test data and allow for the calculation of actual values. Instead of rates, here are some definitions of the most essential terms: True positives (TP): Cases where we predicted yes (the patient has the disease) and, in fact, they have the disease. True negatives (TN): We expected no, and they aren't infected. False positives (FP): Yes, as expected, but they don't actually have the virus. False Negatives (FN): We had expected no, yet they do have the illness. Accuracy: Accuracy is a statistic that is used to assess the performance of classification and regression algorithms. Recall: The finding of positives that are accurately identified as positives is known as recall.A true positive rate is what it's known as. The formula is used to compute it: Precision: Precision is the number of positive predictions made by a model that explains the number of true positives divided by the number of positive predictions, which is an indicator of the model's success. F1 Score: The F1 score is a metric for how accurate models are on a given dataset. It's used to assess binary classification systems that divide the world into positive and negative categories. Here we have divided the dataset into two parts one is 80% and the other is 20%0.80% is used for training and 20% is for validation or testing in both the models. In VGG16 we got an accuracy of 89.34%, Recall is of 89%, Precision is of 89% and F1-Score is of 89%.Table2 represents the confusion matrix for VGG16 The VGG16 model got an accuracy of 89.34% and loss of 24.42%.Fig. 6 represents the VGG16 model training and validation accuracy and loss curves.
Table 2

VGG16 X-Ray Dataset Confusion Matrix.

Fig. 6

Accuracy and Loss curves for VGG16.

VGG16 X-Ray Dataset Confusion Matrix. Accuracy and Loss curves for VGG16. In ResNet50 we got an accuracy of 91.39%, Recall is of 90%, Precision is of 91.3% and F1-Score is of 91%.In ResNet50 got loss of 33.46%.Fig. 7 represents the ResNet50 model preparation and testing accuracy and loss curves.
Fig. 7

Accuracy and Loss curves for ResNet50.

Accuracy and Loss curves for ResNet50.

Conclusion

The main goal of this study is to use a variety of deep learning methods to diagnose COVID-19. For multi-class classification, we used the X-Ray dataset. VGG16 and ReNet50 models for COVID-19 classification have been validated. After the model was deployed, the accuracy, recall precision, and F1-score for COVID19 and Pneumonia diagnosis were 89.34 %, 89 %, 89 %, and 89 % for VGG16 and 91.39 %, 90 %, 91.3 %, 91% for ResNet50. ResNet50 has been shown to be effective in diagnosing COVID19 and Pneumonia patients in the two models proposed.

Declaration of Competing Interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.
  21 in total

1.  Attention-based VGG-16 model for COVID-19 chest X-ray image classification.

Authors:  Chiranjibi Sitaula; Mohammad Belayet Hossain
Journal:  Appl Intell (Dordr)       Date:  2020-11-17       Impact factor: 5.086

2.  COVID-19 detection in X-ray images using convolutional neural networks.

Authors:  Daniel Arias-Garzón; Jesús Alejandro Alzate-Grisales; Simon Orozco-Arias; Harold Brayan Arteaga-Arteaga; Mario Alejandro Bravo-Ortiz; Alejandro Mora-Rubio; Jose Manuel Saborit-Torres; Joaquim Ángel Montell Serrano; Maria de la Iglesia Vayá; Oscar Cardona-Morales; Reinel Tabares-Soto
Journal:  Mach Learn Appl       Date:  2021-08-20

3.  Iteratively Pruned Deep Learning Ensembles for COVID-19 Detection in Chest X-rays.

Authors:  Sivaramakrishnan Rajaraman; Jen Siegelman; Philip O Alderson; Lucas S Folio; Les R Folio; Sameer K Antani
Journal:  IEEE Access       Date:  2020-06-19       Impact factor: 3.367

4.  Application of deep learning techniques for detection of COVID-19 cases using chest X-ray images: A comprehensive study.

Authors:  Soumya Ranjan Nayak; Deepak Ranjan Nayak; Utkarsh Sinha; Vaibhav Arora; Ram Bilas Pachori
Journal:  Biomed Signal Process Control       Date:  2020-11-19       Impact factor: 3.880

5.  Transfer Learning to Detect COVID-19 Automatically from X-Ray Images Using Convolutional Neural Networks.

Authors:  Mundher Mohammed Taresh; Ningbo Zhu; Talal Ahmed Ali Ali; Asaad Shakir Hameed; Modhi Lafta Mutar
Journal:  Int J Biomed Imaging       Date:  2021-05-15

6.  Using Artificial Intelligence to Detect COVID-19 and Community-acquired Pneumonia Based on Pulmonary CT: Evaluation of the Diagnostic Accuracy.

Authors:  Lin Li; Lixin Qin; Zeguo Xu; Youbing Yin; Xin Wang; Bin Kong; Junjie Bai; Yi Lu; Zhenghan Fang; Qi Song; Kunlin Cao; Daliang Liu; Guisheng Wang; Qizhong Xu; Xisheng Fang; Shiqin Zhang; Juan Xia; Jun Xia
Journal:  Radiology       Date:  2020-03-19       Impact factor: 11.105

7.  An optimized deep learning architecture for the diagnosis of COVID-19 disease based on gravitational search optimization.

Authors:  Dalia Ezzat; Aboul Ella Hassanien; Hassan Aboul Ella
Journal:  Appl Soft Comput       Date:  2020-09-22       Impact factor: 6.725

View more
  1 in total

1.  Screening Lung Diseases Using Cascaded Feature Generation and Selection Strategies.

Authors:  Jawad Rasheed; Raed M Shubair
Journal:  Healthcare (Basel)       Date:  2022-07-14
  1 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.