Literature DB >> 35576824

A lightweight CNN-based network on COVID-19 detection using X-ray and CT images.

Mei-Ling Huang1, Yu-Chieh Liao2.   

Abstract

BACKGROUND AND OBJECTIVES: The traditional method of detecting COVID-19 disease mainly rely on the interpretation of computer tomography (CT) or X-ray images (X-ray) by doctors or professional researchers to identify whether it is COVID-19 disease, which is easy to cause identification mistakes. In this study, the technology of convolutional neural network is expected to be able to efficiently and accurately identify the COVID-19 disease.
METHODS: This study uses and fine-tunes seven convolutional neural networks including InceptionV3, ResNet50V2, Xception, DenseNet121, MobileNetV2, EfficientNet-B0, and EfficientNetV2 on COVID-19 detection. In addition, we proposes a lightweight convolutional neural network, LightEfficientNetV2, on small number of chest X-ray and CT images. Five-fold cross-validation was used to evaluate the performance of each model. To confirm the performance of the proposed model, LightEfficientNetV2 was carried out on three different datasets (NIH Chest X-rays, SARS-CoV-2 and COVID-CT).
RESULTS: On chest X-ray image dataset, the highest accuracy 96.50% was from InceptionV3 before fine-tuning; and the highest accuracy 97.73% was from EfficientNetV2 after fine-tuning. The accuracy of the LightEfficientNetV2 model proposed in this study is 98.33% on chest X-ray image. On CT images, the best transfer learning model before fine-tuning is MobileNetV2, with an accuracy of 94.46%; the best transfer learning model after fine-tuning is Xception, with an accuracy of 96.78%. The accuracy of the LightEfficientNetV2 model proposed in this study is 97.48% on CT image.
CONCLUSIONS: Compared with the SOTA, LightEfficientNetV2 proposed in this study demonstrates promising performance on chest X-ray images, CT images and three different datasets.
Copyright © 2022. Published by Elsevier Ltd.

Entities:  

Keywords:  COVID-19; Computer tomography; Transfer learning; X-ray

Mesh:

Year:  2022        PMID: 35576824      PMCID: PMC9090861          DOI: 10.1016/j.compbiomed.2022.105604

Source DB:  PubMed          Journal:  Comput Biol Med        ISSN: 0010-4825            Impact factor:   6.698


Introduction

COVID-19 has spread around the world. Since the virus spreads through human-to-human interaction, the speed of spread is getting faster and faster. As a result, more than 200 countries and regions around the world have been affected by the virus [1]. In the process of epidemic prevention, one of the early diagnosis of COVID-19 is through the use of reverse transcription polymerase chain reaction (RT-PCR) by doctors or medical researchers to detect positive cases of COVID-19. The method is insufficient, expensive and time-consuming [2]. Computer tomography (CT) and chest X-ray (CXR) images are used as other important detection indicators [3]. Practically, chest X-ray images are often interpreted by a radiologist, a process that is time-consuming and prone to errors in subjective assessments. The operation of neural network requires large and high-speed memory access. With the rapid improvement of modern algorithms, big data, and hardware computing capability, the use of artificial intelligence for medical image recognition can be efficient and effective. Among them, Convolutional Neural Networks (CNN) has been proven to be a powerful medical image recognition technology [4]. Some scholars use image normalization, data augmentation and image resizing for image preprocessing [5,6]. For chest X-ray image classification, related researches use methods such as deep learning, transfer learning, and ensemble learning. Science et al. [7] proposed a CNN with chaotic salt swarm algorithm, and the classification accuracy on COVID-19, Viral Pneumonia and Normal groups was 99.69%. Rashid et al. [8] proposed AutoCovNet and achieved classification accuracy of 96.45% on three groups of COVID-19, Pneumonia and Normal; classification accuracy of 99.39% on two groups of COVID-19 and Pneumonia. Based on DenseNet121, Hertel and Benlamri [9] proposed COV-SNET and achieved classification accuracy of 95.00%. Aslan et al. [10] used eight convolutional neural networks, including AlexNet, ResNet18, ResNet50, InceptionV3, DenseNet201, InceptionResNetV2, MobileNetV2, and GoogleNet, to extract features from chest X-ray images, and combined with machine learning classifiers using 3,781 chest X-ray images after data augmentation. The results show that combination of DenseNet and SVM reaches an accuracy of 96.29% on three group classifications of COVID-19, viral pneumonia and normal. Nigam et al. [11] applied VGG16, DenseNet121, Xception, NASNet and EfficientNet. The best classification accuracy of 93.48% on three groups of COVID-19, Pneumonia and Normal was from EfficientNet-B7. The best classification accuracy of 93.48% on two groups of COVID-19 and Pneumonia was from VGG19 [12]. EfficientNet-B4 model results in classification accuracy of 99.62% on two groups of COVID-19 and Normal, and classification accuracy of 96.70% on three groups of COVID-19, Pneumonia and Normal [13]. Gupta et al. [14] proposed InstaCovNet-19 using ensemble methods including ResNet101, Xception, InceptionV3, MobileNet and NASNet. Results showed that the accuracy on three groups of COVID-19, Pneumonia and Normal was 99.08%. Rahimzadeh and Attar [15] combined Xception and ReNet50V2 and obtained classification accuracy of 91.40% on three groups of COVID-19, Pneumonia and Normal. An ensemble learning network DNE-RL optimizing hyper parameters was proposed on two group classification of COVID-19 and Normal, the accuracy was as high as 99.14% [16]. A computer-aided detection model first combines several pre-trained networks for feature extraction, and then uses a sparse auto-encoder and a feed forward neural network (FFNN) to improve the classification performance. The results show that the binary classification accuracy of InceptionResNetV2+Xception in COVID-19 and Non-COVID is 95.78% [17]. On the classification of chest CT images, Wang et al. [18] first used the two-dimensional fractional Fourier entropy to extract features, and then proposed a deep stacked sparse auto-encoder (DSSAE). Data augmentation was used with a total of 1,164 CT images. The results show that classification accuracy, sensitivity and F1-Score for COVID-19, community-acquired pneumonia (CAP), secondary pulmonary tuberculosis (SPT) and normal (Healthy control, HC) are 92.46%, 92.46% and 92.32%, respectively. Rohila et al. [19] proposed ReCOV-101 convolutional neural network on two group classification of COVID-19 and non-COVID-19, the accuracy was as high as 94.90%. Ten convolutional neural networks including AlexNet, VGG16, VGG19, SqueezeNet, GoogleNet, MobileNetV2, ResNet18, ResNet50, ResNet101 and Xception were applied on two group classifications between COVID-19 and non-COVID-19. Results showed that the classification accuracy from ResNet101 was 99.51% [20]. Fu et al. [21] used UNet on chest segmentation, and proposed DenseANet model on classification of three groups of COVID-19, Pneumonia and Normal to obtain the accuracy of 96.06%. Transfer learning was executed on DenseNet121, DenseNet201, VGG16, VGG19, InceptionResNetV2 and Xception models. Further, the classification accuracy achieved 98.80% through combing DenseNet201 and GradCam algorithm on groups of COVID-19 and Pneumonia [22]. Arora et al. [23] applied MobileNet on groups of COVID-19 and Pneumonia to report the classification accuracy of 96.11%. Aswathy et al. [24] proposed a two-step classification method between COVID-19 and non-COVID-19, and the accuracy was 98.50%. Fan et al. [25] proposed a Trans-CNN Net in 2022. This method combines the existing convolutional neural network and the branch network model of the transformer structure to fuse extracted features, using a total of 194,922 CT images. The results show that Trans-CNN Net has an accuracy of 96.73% in three classifications of COVID-19, pneumonia and normal. Zhang et al. [26] proposed PZM-DSSAE in 2021, which first uses the Pseudo-Zernike moment (PZM) to extract the features of the image, including the image plane over unit circle and image plane inside the unit circle; and then use deep stacked sparse autoencoder (DSSAE) as the classifier with nine data augmentation methods. A total of 2960 CT images was used in this study. The results showed that PZM-DSSAE had a binary classification accuracy of 92.31% in COVID-19 and normal, a sensitivity of 92.06%, a specificity of 92.56%, and an F1-Score of 92.29%. Cruz [27] combined ensemble learning and transfer learning on classification between COVID-19 and non-COVID-19, but the accuracy was not high (86.70%). Based on AlexNet, GoogleNet and ResNet, Zhou et al. [28] proposed EDL-COVID ensemble convolutional neural network to extract image features on classification of three groups of COVID-19, Pneumonia and Normal. The final accuracy was as high as 99.05%. Transfer learning was applied on VGG19, ResNet101, DenseNet169 and WideResNet50-2, followed by ensemble method to obtain accuracy of 93.50% on classification between COVID-19 and non-COVID-19 [29]. Abraham and Nair [30] use ensemble learning model combining five convolutional neural networks such as MobileNetV2, ShuffleNet, Xception, DarkNet53 and EfficientNet-B0 for feature extraction, and use the kernel support vector machine (Kernel support vector machine) as the classifier, using a total of 746 CT images. The results show that ensemble learning model has a binary classification accuracy of 91.60% in COVID-19 and Non-COVID. In addition to using chest X-ray or CT images individually, some researches considered both images together in their studies. Vinod et al. [31] proposed Deep Covix-Net network on chest X-ray images of three groups of COVID-19, Pneumonia and Normal, and the classification accuracy was 96.80%; the classification accuracy was 97.00% between COVID-19 and Normal by using CT images. Li et al. [32] proposed CMT-CNN network. The classification accuracies on chest X-ray images between COVID-19 and Normal, and on three groups of COVID-19, Pneumonia and Normal were 97.23% and 93.49%, respectively; the classification accuracies on chest CT images between COVID-19 and Normal, and on three groups of COVID-19, Pneumonia and Normal were 93.46% and 91.45%, respectively. Thakur and Kumar [33] combined 6077 X-ray images and 3877 CT images to obtain high classification of 99.64% and 98.28% between COVID-19 and Normal, and on three groups of COVID-19, Pneumonia and Normal, respectively. Munusamy et al. [34] proposed FractalCovNet network by combining Fractal module and UNet to target the CT image and perform classification task. Banerjee et al. [35] combined InceptionV3, InceptionResNetV2 and DenseNet201 to form COFE-Net on two group classifications of X-ray images between COVID-19 and non-COVID-19 and three group classification of COVID-19, Pneumonia and Normal. The accuracies were 99.49% and 96.39%, respectively; the accuracies on CT images were 99.68% and 98.93%, respectively. Zhang et al. [36] used convolutional block attention module to propose an end-to-end multiple-input deep convolutional attention network (MIDCAN) in 2021, where one input receives 3D CT images and the other receives 2D chest X-ray images. Besides, in order to overcome the problem of overfitting, the authors use a variety of data augmentation methods. The results showed that MIDCAN correctly detected COVID-19 with accuracy, sensitivity and specificity of 98.02%, 98.10%, and 97.95%, respectively. Related researches have demonstrated the alternatives of using lung ultrasound for COVID-19 diagnosis, and mentioned the drawbacks of diagnosis using chest X-ray or CT images. Salvia et al. [37] adopted ResNet50, transfer learning, and data augmentation techniques on COVID-19 detection using ultrasound clips originating from linear and convex probes. A total of 5,400 ultrasound images were used. The results show that the accuracy of ResNet50 in four and seven classes are 98.32% and 98.72%, respectively. Ebadi et al. [38] proposed a Kinetics-I3D network, which can interpret lung ultrasound images quickly and reliably, and can classify the entire lung ultrasound images without using preprocessing or a frame -by- frame analysis. Kinetics-I3D network achieves an accuracy of 90% and an average precision score of 95%. Huang et al. [39] propose a Non-local Channel Attention ResNet to analyze the lung ultrasound images and automatically score the degree of pulmonary edema of patients with COVID-19 pneumonia. 2062 effective images were used. Results show that the binary classification achieves 92.34%. Compared with traditional machine learning methods, CNN have demonstrated great ability on COVID-19 detection by using chest X-ray or CT images jointly or separately. Most of the aforementioned development of the related studies proposed new networks aiming to obtain higher classification accuracy may not consider the number of parameters used in the network architecture or the number of images used in the training phase [[40], [41], [42], [43], [44], [45], [46], [47], [48], [49]]. The objective of this study is to efficiently speed up the training phase with fewer parameters when using a small number of images collected for each group while maintaining favorable classification accuracy for COVID-19 detection. The classification accuracy, Precision, Recall, and F1-Score, and the numbers of parameter used for each model were compared in this study. In addition to the pre-trained transfer learning models before and after fine-tuning, based on the architectures of AlexNet and EfficientNetV2 networks, this study proposes a lightweight convolutional neural network LightEfficientNetV2 to faster the training time for COVID-19 detection. The rest of this paper is organized as follows. Section 2 introduces the collection of datasets on chest X-ray and CT images; image preprocessing method, transfer learning models and the proposed model. Section 3 presents experiment results. Section 4 compares the classification results with the state-of-the-art COVID-19 classification models. The proposed model is further evaluated using three more open public datasets. Finally, Section 5 summarizes the proposed model and suggests the future research.

Materials and methods

Data collection

Chest X-ray images and CT images of COVID-19, Pneumonia and Normal are collected from three Kaggle datasets and one Mendeley dataset in this study. Fig. 1 shows the structure of the study.
Fig. 1

Structure of this research.

Image collection: Establish Dataset 1 and Dataset 2. Chest X-ray images (Dataset 1) are collected from Kaggle Chest X-Ray Images (Pneumonia) [50] and Kaggle COVID-19 Radiography Dataset [51]. CT images (Dataset 2) are collected from Kaggle Large COVID-19 CT scan slice dataset [52] and Mendeley Data Chest CT of multiple pneumonia [53]. Each dataset contains three groups of COVID-19, Pneumonia and Normal images. Image preprocessing: Normalize pixel values and unify image format for Dataset 1 and Dataset 2. Transfer learning models: InceptionV3, ResNet50V2, Xception, DenseNet121, MobileNetV2 and EfficientNet-B0. Fine-tuning models: the above 6 models are fine tuned. Propose a model: We propose a light-weight LightEfficientNetV2 model. Grild Search: Grid SearchCV is used to search for parameter combination. Perfrormance evaluation: Four performance indices are computed for the above model. Comparison with SOTA: The results of the above models are compared with SOTA. Structure of this research. Table 1, Table 2 present the number of images for each group, and 600 images are randomly selected from the total number of images for each group to be used for both datasets in this study. Among the 600 images, the number of images for training, validation and test subsets are 400, 100, and 100, respectively.
Table 1

Dataset 1: Chest X-ray image dataset.

CategoryTotal number of imagesNumber of images selected in this studytrainvalidationtest
COVID-193,616600400100100
Pneumonia5,618600400100100
Normal11,775600400100100
Total21,00918001,200300300
Table 2

Dataset 2: Chest CT image dataset.

CategoryTotal number of imagesNumber of images selected in this studytrainvalidationtest
COVID-199,003600400100100
Pneumonia1,120600400100100
Normal7,063600400100100
Total17,18618001,200300300
Dataset 1: Chest X-ray image dataset. Dataset 2: Chest CT image dataset.

Image preprocessing

To reduce the training time and avoid memory overload, the image size was reduced to 224 × 224 and 299 × 299, and images were converted into the portable network graphics (PNG) file format in this study. Fig. 2, Fig. 3 present the resized chest X-ray images and resized chest CT images, respectively.
Fig. 2

Resized Chest X-ray image.

Fig. 3

Resized Chest CT image.

Resized Chest X-ray image. Resized Chest CT image.

Grid SearchCV

Grid SearchCV is a useful tool on fine-tuning the parameters for models. It implements grid search and cross validation through predefined hyper-parameters to fit the estimators on the training set. The parameters selected include batch size, epochs, optimizer, learning rate and dropout in this study.

Transfer learning models

Transfer learning is mainly to transfer the trained model to another new model and use it in other different but related fields, so that we do not need to start from scratch. It can not only reduce the dependence of a large amount of data, but also improve the classification performance in the target field [[54], [55], [56]]. Considering the number of parameter in the network architecture, InceptionV3, ResNet50V2, Xception, DenseNet121, MobileNetV2, EfficientNet-B0 models were selected in the study.

InceptionV3

Google proposed GoogLeNets in 2014, called it InceptionV1 [57], and then released InceptionV2 and InceptionV3 [58] the following year. InceptionV3 uses convolutional layers with stride = 2 in parallel with pooling layers to shrink feature maps. The difference from InceptionV2 is that the first Inception module of InceptionV3 is to change the 7 × 7 convolutional layer to three 3 × 3 convolutional layers. Through the above improvements, InceptionV3 increases the width and depth of the network to improve performance. In addition, the image was resized from 224 × 224 of the original InceptionV1 to 299 × 299. Therefore, this study selects InceptionV3, which has 311 layers and 21,808,931 parameters in its network architecture.

ResNet50V2

He et al. [59] proposed the ResNet model in 2015 and won the first place in the ILSVRC-2015 classification task. ResNet adds residual learning to solve the problem of gradient and precision drop in training sets, enabling the network to become deeper and deeper, while ensuring accuracy and controlling speed. The ResNet series of models also include ResNet50, ResNet101, and ResNet152. ResNet50V2 is a modified version of ResNet50 and is pre-trained on the ImageNet dataset, which outperforms ResNet50 and ResNet101. Therefore, this study chooses ResNet50V2, which has 190 layers and 25,613,800 parameters in its network architecture, which modifies the propagation formula of connections between convolution modules.

Xception

Google proposed Xception [60] in 2016, and its model was modified from InceptionV3, using depthwise separable convolution (DSC) to replace the original Inception module, and decomposed ordinary convolution into spatial convolution and point-by-point convolution. Spatial convolution is performed on each input channel, while pointwise convolution uses a 1 × 1 kernel to convolve point by point, which reduces not only the number of parameters but also the number of computations. The Xception network architecture in this study has 132 layers and 20,867,627 parameters, and is composed of 14 modules, including 36 convolutional layers. The input image size for this model is 299 × 299.

DenseNet121

Huang et al. [61] proposed the DenseNet model, whose core is mainly composed of DenseBlock (DB), transition layer and growth rate. DenseNet also includes a series of models such as DenseNet121, DenseNet169, DenseNet201 and DenseNet264. The advantage of DenseNet121 is that the model requires less parameters, so that a deeper model can be trained during calculation, and the connection of its fully connected layer also has a regularization effect, which can reduce the problem of overfitting on smaller datasets. DenseNet121 is selected in this study, which has 427 layers and 7,040,579 parameters in its network architecture.

MobileNetV2

Google proposed the MobileNet model in 2017 [62] with a streamlined architecture. DSC is applied to MobileNetV1, and two global hyper-parameters are proposed to control the capacity of the network, which can effectively maintain a considerable high precision between latency and accuracy. MobileNetV2 was released in early 2018, and its architecture adds two new modules, reverse residual and linear bottleneck, which can achieve faster training speed and higher accuracy. MobileNetV2 is specially designed for images and can be used for classification and feature generation. With a total of 154 layers and 2,261,827 parameters, which is much fewer than other commonly used CNN models, MobileNetV2 is selected in this study.

EfficientNet-B0

In 2019, Google proposed the EfficientNet model [63], which is a new network scaling method. It uses compound coefficients to uniformly scale the depth, width and image resolution of the network. In addition, Google also developed a new baseline network using neural architecture search (NAS), and then extended it to a series of models: EfficientNet-B0 ∼ EfficientNet-B7, which is somewhat similar to MnasNet Both use the same search space, but EfficientNet has a larger floating-point operations per second (FLOPS) because the main module used is MBConv. This study chooses EfficientNet-B0, which has 237 layers and 4,053,414 parameters in its network architecture.

EfficientNetV2

EfficientNetV2, proposed by Tan & Le [64] in 2021, possesses faster training speed and better parameter efficiency. Compared to original EfficientNet, EfficientNetV2 uses both MBConv and the newly added fused-MBConv in the early layers. The expansion ratio for MBConv in EfficientNetV2 is 4, which is smaller than the expansion ratio used in EfficientNet. The advantage of smaller expansion ratio is to reduce the memory access overhead. The number of parameter of EfficientNetV2 used in this study is 20,181,331.

The proposed COVID-19 detection model

Fine-tuning transfer learning models

The most common ways of fine-tuning transfer learning are (1) using the pre-trained model as feature extraction, that is, shaving the output layer, and then using the remaining network as a feature extractor and applying it to new dataset. (2) The architecture of the pre-trained model is adopted, but all weights are randomized before training for its own dataset. (3) Keep the weights of the initial layers of the model unchanged and retrain the latter layers to obtain new weights. During this process, many attempts can be made to find the best match between frozen and retrained layers. The dataset used in this study is small and the image similarity is high, so we freeze the weights of the first a few layers in the pre-trained model, and then retrain the latter layers to reduce the number of parameters of each model to be less than one million. Table 3 compares the amount of parameters used before and after fine-tuning. In addition, we use a global average pooling layer to replace the original fully connected layer after the pre-training model, and finally add dropout before the classification layer. Fig. 4 shows the fine-tuned transfer learning model.
Table 3

Comparison of model parameters before and after fine-tuning.

Before fine-tuningParametersAfter fine-tuningParameters
InceptionV321,808,931FT-InceptionV3913,299
ResNet50V223,570,947FT-ResNet50V2957,827
Xception20,867,627FT-Xception925,083
DenseNet1217,040,579FT-DenseNet121948,611
MobileNetV22,261,827FT-MobileNetV2890,851
EfficientNet-B04,053,414FT-EfficientNet-B0912,123
EfficientNetV2-S20,181,331FT-EfficientNetV2-S863,895
Fig. 4

Architecture of fine-tuning transfer learning models.

Comparison of model parameters before and after fine-tuning. Architecture of fine-tuning transfer learning models.

Proposed model - LightEfficientNetV2

The newly introduced convolutional neural network EfficientNetV2 [64] uses training-aware neural architecture search and scaling to jointly optimize training speed and parameter efficiency. Compared with EfficientNetV2, each layer of EfficientNet adopts a compound scaling method. For example, when the depth is 2, the number of layers of each layer is doubled, but the training speed and parameter amount of each layer are not consistent. At the same time, EfficientNet continues to increase the image size, which also leads to large memory consumption, which in turn affects the training speed. EfficientNetV2 uses a non-uniform scaling strategy, which can increase the number of layers in more subsequent layers. Although many existing deep learning models have been proven to be powerful classification networks, there are still problems with a large number of parameters, which in turn leads to excessive training time. This study proposes a lightweight convolutional neural network LightEfficientNetV2 with reference to the AlexNet and EfficientNetV2 network architectures. The reasons are as follows: AlexNet achieves favorable classification results when the number of network layers is small, mainly because the size of the convolution kernel in the first two convolutional layers is relatively large, which makes it possible to extract features more accurately. In addition, the activation function ReLU is added to each convolution layer as a regularization method, which helps to improve the generalization ability of the model. Compared with previous models, EfficientNetV2 not only has faster training speed but also higher parameter efficiency. The related research points out that replacing the MBConv in the first to third layers of EfficientNet-B4 with Fused-MBConv will significantly improve the training speed. But if it is replaced in all layers, the amount of parameters and FLOPs will be greatly increased at this time. This shows that a proper combination of MBConv and Fused-MBConv can achieve beteer results. Based on the above two points, we proposed a lightweight new model LightEfficientNetV2 in this study. First use the first two convolution layers of AlexNet, and then select convolution modules of MBConv and Fused-MBConv from the original EfficientNetV2, and finally add two convolution modules we constructed to form LightEfficientNetV2 in this study. With a total parameter amount of 798,539, which is significantly reduced from the original EfficientNetV2 parameter amount of 24,000,000. This not only reduces the computational cost, but also reduces the computational complexity, enabling accurate feature extraction with a small number of parameters. Fig. 5 and Table 4 present the layer-by-layer architecture of LightEfficientNetV2 used in this study, and the model is detailed as follows:
Fig. 5

Architecture of LightEfficientNetV2.

Table 4

Architecture of LightEfficientNetV2.

Layer (type)FilterKernelSizeStridePaddingOutput shape
Image Input224 × 224 × 1
Conv 1a, k 11 × 119611 × 114156 × 56 × 96
Batch normalization 1b56 × 56 × 96
Conv 2a, k 5 × 51285 × 51056 × 56 × 128
Batch normalization 2b56 × 56 × 128
Fused-MBConv3Conv 3a243 × 31056 × 56 × 24
Batch normalization 3b56 × 56 × 24
SE 3c2456 × 56 × 24
856 × 56 × 8
Conv 3d241 × 1228 × 28 × 24
Batch normalization 3e28 × 28 × 24
MBConv4Conv 4a2561 × 11028 × 28 × 256
Batch normalization 4b28 × 28 × 256
SE 4c25628 × 28 × 256
828 × 28 × 8
DepthwiseConv 4d1283 × 3214 × 14 × 128
Batch normalization 4e14 × 14 × 128
Conv 4f5121 × 1114 × 14 × 512
Batch normalization 4g7 × 7 × 512
Conv 5a, k 3 × 3643 × 3107 × 7 × 64
Batch normalization 5b7 × 7 × 64
Conv 6a, k 1 × 1641 × 1107 × 7 × 64
Batch normalization 6b7 × 7 × 64
Global Average Pooling1 × 1 × 64
Dropout
FC1 × 1 × 3
Softmax1 × 1 × 3
Classification Output
Padding was added to the first convolutional layer in MBConv and Fused-MBConv in this study. In order to pursue higher accuracy, MBConv is the convolution module used in the Efficientnet network. Fused-MBConv performs faster. Batch normalization replaces the max pooling layer in the AlexNet convolutional module after each convolutional layer. ReLU can retain important information in the image and is added to the last two convolution modules added by this study. Adding batch normalization to the convolutional layer can make training more stable [65]. This study adds a dropout before the classification layer. Architecture of LightEfficientNetV2. Architecture of LightEfficientNetV2.

Experimental setup and implementation

In the experiment, we extract 100 images for each category as the test set, and then divide the remaining images into 80% training and 20% validation. Five-fold cross validation is used to evaluate the performance of each model. Furthermore, we use grid search to obtain the best parameters, and the values selected for each parameter are shown in Table 5 . The final result is batch size of 16; epochs count of 50; optimizer of Adam, learning rate of 1e-4 and dropout of 0.2. The equipment used in the experiment is an Intel(R) Core(TM) i9-10900F 2.81 GHz CPU, NVIDIA GeForce RTX 3070 8G GPU. The whole experiment process is performed using Python 3.8 [Python Software Foundation, Fredericksburg, Virginia, USA], which contains keras 2.6 and tensorflow 2.6.
Table 5

Parameter setting.

ParameterValue
Batch size8, 16, 32
Epochs20, 50, 100, 150, 200, 250, 300
OptimizerSGD, RMSprop, Adagrad, Adadelta, Adam, Adamax, Nadam
Learning rate0.00001, 0.0001, 0.001, 0.01, 0.1
Dropout0.0, 0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, 0.8, 0.9
Parameter setting.

Evaluation of model performance

This study uses a confusion matrix to analyze the performance of the models used, as shown in Table 6 . The four evaluation metrics used for model evaluation are accuracy, precision, recall, and F1-score, and the mathematical formulas of these indicators are shown in Eqs. (1), (2), (3), (4), respectively, where TP (true positive) represents the number of positive categories that are correctly classified as positive, FP (false positive) represents the number of negative categories that are incorrectly classified as positive, TN (true negative) refers to the number of negative categories that are correctly classified as negative, and FN (false negative) refers to the number of positive categories that are incorrectly classified as negative.
Table 6

Confusion matrix.

Confusion matrixActual situation
COVID-19PneumoniaNormal
Model classificationCOVID-19COVID-19 category that are correctly classified as COVID-19.Pneumonia category that are wrongly classified as COVID-19.Normal category that are wrongly classified as COVID-19.
PneumoniaCOVID-19 category that are wrongly classified as Pneumonia.Pneumonia category that are correctly classified as Pneumonia.Normal category that are wrongly classified as Pneumonia.
NormalCOVID-19 category that are wrongly classified as Normal.Pneumonia category that are wrongly classified as Normal.Normal category that are correctly classified as Normal.
Confusion matrix.

Results

Performance results on dataset 1

Table 7 reports the average and standard deviation of accuracies in the training, validation and test set for InceptionV3, ResNet50V2, Xception, DenseNet121, MobileNetV2, EfficientNet-B0 and EfficientNetV2 models before and after fine-tuning, and the proposed LightEfficientNetV2. The accuracy of the model after fine-tuning are higher than the accuracy of the model before fine-tuning for InceptionV3, ResNet50V2, Xception, DenseNet121, MobileNetV2, EfficientNet-B0 and EfficientNetV2 models. The training time of the model after fine-tuning is less than the training time of the model before fine-tuning for all models. LightEfficientNetV2 achieved the highest accuracy 98.33% with the least training time in the test set.
Table 7

Model accuracy on Dataset 1.

TrainValidTestTrain timeTest time
InceptionV399.71 ± 0.11%96.55 ± 0.54%96.50 ± 0.82%00:14:2700:00:03
ResNet50V299.65 ± 0.25%95.51 ± 0.99%95.10 ± 1.93%00:10:4000:00:02
Xception99.77 ± 0.14%96.32 ± 0.50%95.89 ± 0.35%00:33:0300:00:02
DenseNet12198.45 ± 0.45%95.14 ± 0.70%94.67 ± 1.28%00:14:3000:00:03
MobileNetV299.56 ± 0.41%94.87 ± 1.01%96.36 ± 1.02%00:08:4700:00:01
EfficientNetB095.09 ± 1.11%94.40 ± 0.60%94.80 ± 1.68%00:16:1400:00:02
EfficientNetV2
99.75 ± 0.14%
97.13 ± 0.87%
94.81 ± 0.58%
00:21:32
00:00:03
FT-InceptionV399.75 ± 0.11%98.10 ± 0.66%97.72 ± 0.38%00:12:2300:00:03
FT-ResNet50V299.88 ± 0.07%97.62 ± 0.64%97.13 ± 0.86%00:09:2000:00:02
FT-Xception99.86 ± 0.06%97.94 ± 0.47%97.31 ± 1.56%00:25:2500:00:02
FT-DenseNet12199.90 ± 0.07%98.05 ± 0.60%95.41 ± 1.25%00:12:3500:00:03
FT-MobileNetV299.68 ± 0.32%97.16 ± 0.60%97.71 ± 1.31%00:05:2100:00:01
FT-EfficientNet-B099.13 ± 0.43%97.32 ± 0.59%96.77 ± 1.18%00:12:2200:00:02
FT-EfficientNetV299.78 ± 0.13%96.65 ± 1.14%96.73 ± 1.19%00:09:2900:00:02
LightEfficientNetV299.80 ± 0.16%95.17 ± 0.74%98.33 ± 0.99%00:04:3000:00:01
Model accuracy on Dataset 1. Table 8 reports the average and standard deviation of precision, recall, and F1-score in the test set for models of InceptionV3, ResNet50V2, Xception, DenseNet121, MobileNetV2, EfficientNet-B0 and EfficientNetV2 models before and after fine-tuning, and the proposed LightEfficientNetV2. The precision, recall, and F1-score of the model after fine-tuning are higher than the precision, recall, and F1-score of the model before fine-tuning for all models. The highest precision, recall and F1-score from the proposed LightEfficientNetV2 model are 98.22%, 98.30% and 98.22%, respectively.
Table 8

Model performance on Dataset 1.

PrecisionRecallF1-score
InceptionV396.73 ± 0.83%96.77 ± 0.90%96.66 ± 0.83%
ResNet50V295.19 ± 1.87%95.10 ± 1.95%95.13 ± 1.82%
Xception96.54 ± 0.73%96.04 ± 0.45%96.50 ± 0.76%
DenseNet12194.74 ± 1.45%94.63 ± 1.33%95.01 ± 1.29%
MobileNetV296.30 ± 0.99%96.33 ± 1.01%96.04 ± 0.80%
EfficientNetB094.54 ± 1.43%94.49 ± 1.51%94.46 ± 1.49%
EfficientNetV2
94.77 ± 0.57%
94.83 ± 0.59%
94.63 ± 0.39%
FT-InceptionV397.66 ± 0.46%97.69 ± 0.40%97.78 ± 0.58%
FT-ResNet50V296.77 ± 0.86%96.89 ± 1.06%96.87 ± 0.81%
FT-Xception97.46 ± 1.62%97.35 ± 1.59%97.24 ± 2.18%
FT-DenseNet12195.14 ± 1.11%95.26 ± 1.31%95.19 ± 0.77%
FT-MobileNetV297.56 ± 1.46%97.65 ± 1.47%97.44 ± 1.48%
FT-EfficientNet-B096.71 ± 1.00%96.60 ± 1.10%96.63 ± 1.13%
FT-EfficientNetV296.81 ± 1.12%96.84 ± 1.12%96.70 ± 1.18%
LightEfficientNetV298.22 ± 0.88%98.30 ± 0.96%98.22 ± 0.90%
Model performance on Dataset 1. Since models after fine-tuning have better performance than models before fine-tuning, the four performance metrics for models of InceptionV3, ResNet50V2, Xception, DenseNet121, MobileNetV2, EfficientNet-B0 and EfficientNetV2 models after fine-tuning, and the proposed LightEfficientNetV2 were compared for dataset 1 in Table 9 . With the best performance, the accuracy, precision, recall and F1-score from the proposed LightEfficientNetV2 are significantly different from the other seven models. InceptionV3 model after fine-tuning scores the second place. The least standard deviations for four metrics are all from InceptionV3 model after fine-tuning, which has the best model stability.
Table 9

Performance comparison on dataset 1.

ModelsDataset 1
Accuracy (%)Precision (%)Recall (%)F1-score (%)
FT-InceptionV397.72(±0.38)b97.66(±0.46)d97.69(±0.40)e97.78(±0.58)g
FT-ResNet50V297.13(±0.86)96.77(±0.86)96.89(±1.06)96.87(±0.81)
FT-Xception97.31(±1.56)97.46(±1.62)97.35(±1.59)97.24(±2.18)
FT-DenseNet12195.41(±1.25)b,c95.14(±1.11)d95.26(±1.31)e,f95.19(±0.77)g
FT-MobileNetV297.71(±1.31)c97.56(±1.46)97.65(±1.47)f97.44(±1.48)
FT-EfficientNet-B096.77(±1.18)96.71(±1.00)96.60(±1.10)96.63(±1.13)
FT-EfficientNetV296.73(±1.19)96.81(±1.12)96.84(±1.12)96.70(±1.18)
LightEfficientNetV2
98.33(±0.99)a
98.22(±0.88)a
98.30(±0.96)a
98.22(±0.90)a
P-value0.00**0.00**0.00**0.00**

**p < 0.01.

a Means are statistically significant from the others.

b-g Means that share a letter are significantly different.

Performance comparison on dataset 1. **p < 0.01. a Means are statistically significant from the others. b-g Means that share a letter are significantly different.

Performance results on dataset 2

Table 10 reports the average and standard deviation of accuracies in the training, validation and testing set for InceptionV3, ResNet50V2, Xception, DenseNet121, MobileNetV2, EfficientNet-B0 and EfficientNetV2 models before and after fine-tuning, and the proposed LightEfficientNetV2. The accuracy of the model after fine-tuning is higher than the accuracy of the model before fine-tuning for InceptionV3, ResNet50V2, Xception, DenseNet121, MobileNetV2, EfficientNet-B0 and EfficientNetV2 models. The training time of the model after fine-tuning is less than the training time of the model before fine-tuning for all models. LightEfficientNetV2 achieved the highest accuracy 97.48% with the least training time in the test set. Table 11 reports the average and standard deviation of precision, recall, and F1-score in the test set for models of InceptionV3, ResNet50V2, Xception, DenseNet121, MobileNetV2, EfficientNet-B0 and EfficientNetV2 models before and after fine-tuning, and the proposed LightEfficientNetV2. The precision, recall, and F1-score of the model after fine-tuning are higher than the precision, recall, and F1-score of the model before fine-tuning for all models. The highest precision, recall and F1-score from the proposed LightEfficientNetV2 model are 97.41%, 97.40% and 97.49%, respectively.
Table 10

Model accuracy on Dataset 2.

TrainValidTestTrain timeTest time
InceptionV399.84 ± 0.10%92.53 ± 1.35%93.02 ± 1.15%00:14:3100:00:03
ResNet50V299.88 ± 0.10%93.65 ± 0.85%93.43 ± 1.91%00:10:3300:00:02
Xception98.20 ± 1.59%93.41 ± 1.25%93.46 ± 0.80%00:33:0600:00:02
DenseNet12197.61 ± 0.88%93.68 ± 1.30%93.12 ± 2.21%00:14:3300:00:03
MobileNetV299.77 ± 0.13%94.36 ± 0.59%94.46 ± 1.44%00:08:4900:00:01
EfficientNetB094.51 ± 1.59%92.50 ± 1.16%93.40 ± 2.17%00:17:3200:00:02
EfficientNetV2
99.81 ± 0.07%
93.46 ± 0.37%
93.42 ± 0.31%
00:21:37
00:00:03
FT-InceptionV399.46 ± 0.35%94.70 ± 0.45%94.51 ± 0.76%00:12:2000:00:03
FT-ResNet50V299.84 ± 0.06%95.64 ± 0.54%96.01 ± 1.83%00:09:2100:00:02
FT-Xception99.88 ± 0.08%96.01 ± 0.67%96.78 ± 1.19%00:25:4500:00:02
FT-DenseNet12199.88 ± 0.05%96.35 ± 1.03%96.58 ± 1.07%00:12:4000:00:03
FT-MobileNetV299.87 ± 0.09%95.58 ± 1.10%95.82 ± 1.42%00:05:2400:00:01
FT-EfficientNet-B099.44 ± 0.16%95.67 ± 0.39%95.73 ± 1.05%00:12:0800:00:02
FT-EfficientNetV299.82 ± 0.09%95.11 ± 0.83%96.67 ± 1.35%00:09:2700:00:02
LightEfficientNetV299.89 ± 0.08%96.65 ± 0.56%97.48 ± 0.85%00:04:3000:00:01
Table 11

Model performance on Dataset 2.

PrecisionRecallF1-score
InceptionV393.20 ± 1.02%92.86 ± 1.02%92.83 ± 0.98%
ResNet50V293.61 ± 2.07%93.56 ± 2.00%93.46 ± 2.07%
Xception94.09 ± 0.82%94.06 ± 0.90%93.89 ± 0.77%
DenseNet12193.11 ± 2.25%93.07 ± 2.23%93.06 ± 2.31%
MobileNetV294.51 ± 1.40%94.45 ± 1.40%94.41 ± 1.36%
EfficientNetB093.31 ± 2.20%93.30 ± 2.10%93.15 ± 2.20%
EfficientNetV2
93.46 ± 0.22%
93.48 ± 0.24%
93.34 ± 0.34%
FT-InceptionV394.77 ± 1.00%94.60 ± 0.85%94.57 ± 1.17%
FT-ResNet50V295.67 ± 1.90%95.67 ± 1.86%95.66 ± 2.07%
FT-Xception96.43 ± 1.15%96.52 ± 1.28%96.48 ± 1.35%
FT-DenseNet12196.59 ± 1.02%96.53 ± 1.09%96.48 ± 1.18%
FT-MobileNetV295.74 ± 1.22%95.83 ± 1.47%95.06 ± 0.67%
FT-EfficientNet-B095.48 ± 0.23%95.74 ± 1.01%95.55 ± 0.74%
FT-EfficientNetV296.11 ± 1.30%96.09 ± 1.38%95.98 ± 1.35%
LightEfficientNetV297.41 ± 1.03%97.40 ± 0.96%97.49 ± 0.92%
Model accuracy on Dataset 2. Model performance on Dataset 2. Since models after fine-tuning have better performance than models before fine-tuning, the four performance metrics for models of InceptionV3, ResNet50V2, Xception, DenseNet121, MobileNetV2, EfficientNet-B0 and EfficientNetV2 models after fine-tuning, and the proposed LightEfficientNetV2 were compared for dataset 2 in Table 12 . With the best performance, the accuracy, precision, recall and F1-score from the proposed LightEfficientNetV2 are significantly different from the other seven models. Unlike the results found in Dataset 1, InceptionV3 model after fine-tuning does not score the second place. The performance of InceptionV3 model after fine-tuning was unfavorable among eight models.
Table 12

Performance comparison on dataset 2.

ModelsDataset 2
Accuracy (%)Precision (%)Recall (%)F1-score (%)
FT-InceptionV394.51(±0.76)b,c,d94.77(±1.00)e94.60(±0.85)f,g94.57(±1.17)h,i
FT-ResNet50V296.01(±1.83)95.67(±1.90)95.67(±1.86)95.66(±2.07)
FT-Xception96.78(±1.19)b96.43(±1.15)96.52(±1.28)g96.48(±1.35)i
FT-DenseNet12196.58(±1.07)c96.59(±1.02)e96.53(±1.09)f96.48(±1.18)h
FT-MobileNetV295.82(±1.42)95.74(±1.22)95.83(±1.47)95.06(±0.67)
FT-EfficientNet-B095.73(±1.05)95.48(±0.23)95.74(±1.01)95.55(±0.74)
FT-EfficientNetV296.67(±1.35)d96.11(±1.30)96.09(±1.38)95.98(±1.35)
LightEfficientNetV2
97.48(±0.85)a
97.41(±1.03)a
97.40(±0.96)a
97.49(±0.92)a
P-value0.00**0.00**0.00**0.00**

**p < 0.01.

aMeans are statistically significant from the others.

b-g Means that share a letter are significantly different.

Performance comparison on dataset 2. **p < 0.01. aMeans are statistically significant from the others. b-g Means that share a letter are significantly different.

Discussion

We compare seven models after fine-tuning and the proposed LightEfficientNetV2 model against the related state-of-the-art models in this section. Table 13 shows quantitative comparison results including the metrics of accuracy and parameters in its network architecture. In fact, the accuracy from our proposed LightEfficientNetV2 model was slightly inferior or close to some of the studies [5,41,42,46,66]. However, in terms of the total number of parameters used in the model, the model proposed in this study is less than those in most other studies. The least amount of parameters 131,842 used is from Ref. [67], which is less than that in our proposed LightEfficientNetV2 model in this study, but the model [55] only achieved an accuracy of 88.00% in the chest X-ray image dataset. In addition, the number of images selected for each category in our study is 600, which is much fewer than most studies. With fewer images, the training time can be greatly reduced in our models.
Table 13

Comparison of the proposed model with SOTA.

Study(s)DatasetArchitectureClassAccuracyParameter
Panwar et al. (2020) [67]Total: 5863 X-ray images.nCOVnet288.00%131,842
Das et al. (2021) [40]219 COVID-19,1345 viral pneumonia,1341 normal X-ray images.TLCoV + VGG16397.67%12,410,021
Tahir et al. (2021) [41]11596 COVID-19,11263 Non-COVID-19,10701 normal X-ray images.FPN + InceptionV4399.23%4,357,000
Sheykhivand et al. (2021) [42]2842 COVID-19,2840 viral pneumonia,2778 bacterial pneumonia,2923 normal X-ray images.Proposed DNN499.50%23,070,232
Saha et al. (2021) [5]2300 COVID-19,2300 normal X-ray images.EMCNet298.91%3,955,009
Ibrahim et al. (2021) [43]Total: 33676 (X-ray + CT images)VGG19+ CNN498.05%22,337,604
Hussain et al. (2021) [44]500 COVID-19,400 viral pneumonia,400 bacterial pneumonia,800 normal X-ray images.CoroDet394.20%2,873,609
Panwar et al. (2020) [45]Total: 5856 X-ray images.Proposed CNN295.61%5,244,098
Total: 3008 CT images.
Moghaddam et al. (2021) [46]4001 COVID-19,5705 non-informative,9979 normal CT images.WCNN4399.03%4,610,531
Ahamed et al. (2021) [66]1143 covid-19,1150 bacterial pneumonia,1150 viral pneumonia,1150 normal X-ray images.Modified & Tuned ResNet50V2496.45%49,210,756
1000 covid-19,1000 normal,1000 CAP CT images.399.01%
ProposedX-ray images600 COVID-19,600 pneumonia,600 normal.FT-EfficientNet-B0396.77%912,123
FT-DenseNet12195.41%948,611
FT-ResNet50V297.13%957,827
FT-MobileNetV297.71%890,851
FT-InceptionV397.81%913,299
FT-Xception97.31%925,083
FT-EfficientNetV296.73%863,895
LightEfficientNetV298.33%798,539
CT images600 COVID-19,600 pneumonia,600 normal.FT-EfficientNet-B0395.73%912,123
FT-DenseNet12196.58%948,611
FT-ResNet50V296.01%957,827
FT-MobileNetV295.82%890,851
FT-InceptionV394.51%913,299
FT-Xception96.78%925,083
FT-EfficientNetV296.00%863,895
LightEfficientNetV297.48%798,539
Comparison of the proposed model with SOTA. To verify the performance of the proposed LightEfficientNetV2 model, three more open public datasets NIH Chest X-rays (https://www.kaggle.com/nih-chest-xrays/data/code), SARS-CoV-2 (https://www.kaggle.com/plameneduardo/sarscov2-ctscan-dataset) and COVID-CT (https://github.com/UCSD-AI4H/COVID-CT) were tested and the performance metrics were displayed in Table 14 . The first dataset NIH Chest X-rays is a larger dataset, which contains 15 categories, 112,120 chest X-ray images. The classification accuracies reported from related studies were around 73%–81%, and the highest accuracy 81.60% was from Guan and Huang [68] in 2020. The accuracy from our proposed LightEfficientNetV2 model is 83.42%, which dominates the related studies.
Table 14

Comparisons on NIH Chest X-rays, SARS-CoV-2 and COVID-CT datasets.

DatasetStudy(s)AccuracyPrecisionRecallF1-scoreParameter
NIH Chest X-raysWang et al. (2017) [72]74.50%
Yao et al. (2018) [73]76.10%
Guendel et al. (2018) [74]80.70%
Li et al. (2018) [75]75.50%
Shen and Gao (2018) [76]77.50%
Tang et al. (2018) [77]80.30%
Liu et al. (2019) [3]81.50%
Bharati et al. (2020) [78]73.00%69.00%63.00%68.00%15,488,051
Guan and Huang (2020) [68]81.60%
Proposed (LightEfficientNetV2)
83.42%
81.94%
82.76%
82.39%
798,539
SARS-CoV-2Özkaya et al. (2019) [79]96.00%
Soares et al. (2020) [80]97.30%99.10%95.50%97.30%
Panwar et al. (2020) [45]95.00%95.30%94.00%94.30%5,244,098
Jaiswal et al. (2021) [81]96.20%20,242,984
Shaik and Cherukuri (2021) [69]98.99%98.98%99.00%98.99%
Proposed (LightEfficientNetV2)
99.47%
99.19%
99.25%
99.04%
798,539
COVID-CTMishra et al. (2020) [82]88.30%86.70%
Saqib et al. (2020) [83]80.30%78.20%85.70%81.80%
He et al. (2020) [70]86.00%85.00%14,149,480
Mobiny et al. (2020) [84]84.00%
Polsinelli et al. (2020) [71]85.00%85.00%87.00%86.00%12,600,000
Yang et al. (2020) [85]89.10%25,600,000
Cruz (2021) [24]86.00%89.00%85.00%
Shaik and Cherukuri (2021) [69]93.33%93.17%93.54%93.29%
Proposed (LightEfficientNetV2)88.67%87.28%87.43%87.55%798,539
Comparisons on NIH Chest X-rays, SARS-CoV-2 and COVID-CT datasets. The second dataset SARS-CoV-2 contains 2 categories with 2,482 CT images in total. With accuracies higher than 95%, the performance of related studies are favorable. The accuracy of 98.99% was from Shaik and Cherukuri [69] using ensemble learning model. Even with fewer parameter in the network architecture, our proposed LightEfficientNetV2 model scores the highest accuracy of 99.47%. The third dataset COVID-CT contains 2 categories with 746 CT images in total. Most of the accuracies from related studies are less than 90%, except the one 93.33% from Shaik and Cherukuri [69]. Although the accuracy from our proposed LightEfficientNetV2 model is 88.67%, which is lower than that in Ref. [69], the parameter in LightEfficientNetV2 model is much less than those in Refs. [70,71].

Conclusion

Based on AlexNet and EfficientNetV2, a light-weight convolutional neural network LightEfficientNetV2 is introduced on classification of COVID-19, Pneumonia and Normal images. With less number of parameters used in the network during the training phase as compared to state-of-the-art models, the proposed LightEfficientNetV2 model achieves high accuracies of 98.33% and 97.48% on chest X-ray image and CT image, respectively. LightEfficientNetV2 saves computation and running time as compared with the original models. LightEfficientNetV2 excels favorable performance on three other public datasets NIH Chest X-rays, SARS-CoV-2 and COVID-CT. In addition to classification of three groups of COVID-19, Pneumonia and Normal images in this study, Bacterial Pneumonia, Tuberculosis, or Chest Abdomen Pelvis are encouraged to be considered in the future study. Although the model parameter of the proposed LightEfficientNetV2 is much less than those used in the related study, the accuracy achieved from LightEfficientNetV2 is not the best as compared to the related works. Using the advanced ensemble learning network which reduces the misclassification rate from single models might be a direction in the future research.

Declaration of competing interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.
  50 in total

1.  SDFN: Segmentation-based deep fusion network for thoracic disease classification in chest X-ray images.

Authors:  Han Liu; Lei Wang; Yandong Nan; Faguang Jin; Qi Wang; Jiantao Pu
Journal:  Comput Med Imaging Graph       Date:  2019-05-28       Impact factor: 4.790

2.  Classification of the COVID-19 infected patients using DenseNet201 based deep transfer learning.

Authors:  Aayush Jaiswal; Neha Gianchandani; Dilbag Singh; Vijay Kumar; Manjit Kaur
Journal:  J Biomol Struct Dyn       Date:  2020-07-03

3.  COVID-19 detection in X-ray images using convolutional neural networks.

Authors:  Daniel Arias-Garzón; Jesús Alejandro Alzate-Grisales; Simon Orozco-Arias; Harold Brayan Arteaga-Arteaga; Mario Alejandro Bravo-Ortiz; Alejandro Mora-Rubio; Jose Manuel Saborit-Torres; Joaquim Ángel Montell Serrano; Maria de la Iglesia Vayá; Oscar Cardona-Morales; Reinel Tabares-Soto
Journal:  Mach Learn Appl       Date:  2021-08-20

4.  A modified deep convolutional neural network for detecting COVID-19 and pneumonia from chest X-ray images based on the concatenation of Xception and ResNet50V2.

Authors:  Mohammad Rahimzadeh; Abolfazl Attar
Journal:  Inform Med Unlocked       Date:  2020-05-26

5.  A computer-aided diagnosis system for the classification of COVID-19 and non-COVID-19 pneumonia on chest X-ray images by integrating CNN with sparse autoencoder and feed forward neural network.

Authors:  Gayathri J L; Bejoy Abraham; Sujarani M S; Madhu S Nair
Journal:  Comput Biol Med       Date:  2021-12-14       Impact factor: 4.589

6.  COVID-19 infection localization and severity grading from chest X-ray images.

Authors:  Anas M Tahir; Muhammad E H Chowdhury; Amith Khandakar; Tawsifur Rahman; Yazan Qiblawey; Uzair Khurshid; Serkan Kiranyaz; Nabil Ibtehaz; M Sohel Rahman; Somaya Al-Maadeed; Sakib Mahmud; Maymouna Ezeddin; Khaled Hameed; Tahir Hamid
Journal:  Comput Biol Med       Date:  2021-10-30       Impact factor: 4.589

7.  COV-SNET: A deep learning model for X-ray-based COVID-19 classification.

Authors:  Robert Hertel; Rachid Benlamri
Journal:  Inform Med Unlocked       Date:  2021-05-27

8.  Application of deep learning for fast detection of COVID-19 in X-Rays using nCOVnet.

Authors:  Harsh Panwar; P K Gupta; Mohammad Khubeb Siddiqui; Ruben Morales-Menendez; Vaishnavi Singh
Journal:  Chaos Solitons Fractals       Date:  2020-05-28       Impact factor: 5.944

9.  Evaluation of Pulmonary Edema Using Ultrasound Imaging in Patients With COVID-19 Pneumonia Based on a Non-local Channel Attention ResNet.

Authors:  Qinghua Huang; Ye Lei; Wenyu Xing; Chao He; Gaofeng Wei; Zhaoji Miao; Yifan Hao; Guannan Li; Yan Wang; Qingli Li; Xuelong Li; Wenfang Li; Jiangang Chen
Journal:  Ultrasound Med Biol       Date:  2022-02-07       Impact factor: 2.998

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.