Literature DB >> 34971978

A computer-aided diagnosis system for the classification of COVID-19 and non-COVID-19 pneumonia on chest X-ray images by integrating CNN with sparse autoencoder and feed forward neural network.

Gayathri J L1, Bejoy Abraham2, Sujarani M S3, Madhu S Nair4.   

Abstract

Several infectious diseases have affected the lives of many people and have caused great dilemmas all over the world. COVID-19 was declared a pandemic caused by a newly discovered virus named Severe Acute Respiratory Syndrome Coronavirus 2 (SARS-CoV-2) by the World Health Organisation in 2019. RT-PCR is considered the golden standard for COVID-19 detection. Due to the limited RT-PCR resources, early diagnosis of the disease has become a challenge. Radiographic images such as Ultrasound, CT scans, X-rays can be used for the detection of the deathly disease. Developing deep learning models using radiographic images for detecting COVID-19 can assist in countering the outbreak of the virus. This paper presents a computer-aided detection model utilizing chest X-ray images for combating the pandemic. Several pre-trained networks and their combinations have been used for developing the model. The method uses features extracted from pre-trained networks along with Sparse autoencoder for dimensionality reduction and a Feed Forward Neural Network (FFNN) for the detection of COVID-19. Two publicly available chest X-ray image datasets, consisting of 504 COVID-19 images and 542 non-COVID-19 images, have been combined to train the model. The method was able to achieve an accuracy of 0.9578 and an AUC of 0.9821, using the combination of InceptionResnetV2 and Xception. Experiments have proved that the accuracy of the model improves with the usage of sparse autoencoder as the dimensionality reduction technique.
Copyright © 2021 Elsevier Ltd. All rights reserved.

Entities:  

Keywords:  CNN; COVID-19; Computer-aided detection; Feed forward neural network; Sparse autoencoder

Mesh:

Year:  2021        PMID: 34971978      PMCID: PMC8668604          DOI: 10.1016/j.compbiomed.2021.105134

Source DB:  PubMed          Journal:  Comput Biol Med        ISSN: 0010-4825            Impact factor:   4.589


Introduction

The novel coronavirus outbreak was reported by officials in December 2019 in Wuhan City, China. The virulence of COVID-19 has affected more than 200 million lives and killed more than four million people across the world [1]. The Reverse Transcription-Polymerase Chain Reaction (RT-PCR) test is treated as the golden standard for detection of the SARS-CoV-2 virus. The rapid increase in the number of patients, and the lack of sufficient RT-PCR test facilities in several parts of the world causes delay in testing and detection of the disease. Accessible, fast and affordable methods could play an important role in the diagnosis of the disease. Radiographic methods are easily available and affordable compared to RT-PCR test. A computer-aided diagnosis method using X-ray images could assist the medical practitioners in the detection of COVID-19 at an early stage. Modalities for COVID-19 detection include Computed tomography (CT), X-Ray and Ultrasound imaging. COVID-19 manifests common abnormal X-ray findings including ground glass and consolidative opacities in the peripheral lung regions, nodular opacities and bilateral patchy and confluent patterns. Some image findings include small amounts of pleural effusions, which are uncommon.Viral pneumonia exhibits patches in bilateral areas of consolidation, thickening of bronchial walls, bilateral consolidations, and ground glass opacities or centrilobular nodules poorly defined. Since COVID-19 and other pneumonia share some of these X-ray characteristics, it is not easy to differentiate the COVID-19 images from other pneumonia. Typical radiological COVID-19 pattern including bilateral peripheral or focal round ground glass opacities with or without consolidation differentiates COVID-19 from other pneumonia [2]. However, manual examination of the image modalities is time-consuming as the number of cases are increasing day by day. The application of machine learning in the biomedical field can assist physicians in the computer-aided diagnosis of medical images efficiently and effectively. CAD systems aid radiologists to expound the medical images. Hence computer-aided detection could assist the radiologists in distinguishing COVID-19 infected radiographic images. CAD systems using Artificial Neural Networks (ANNs) and Deep Learning (DL) have shown tremendous success in the field of medical data analysis [3]. Deep learning technologies widely used in disease diagnosis include CNNs, autoencoder, Deep Belief Networks (DBN) and Generative Adversarial Network (GAN) [3]. Several works using ANNs and Deep Learning has been published in the application of disease diagnosis, including detection of interstitial lung disease [4], depression screening [5], schizophrenia [6], ECG arrhythmias classification [7], and ischemic heart disease [8]. Various works related to computer-aided detection of COVID-19 was published. Abraham et al. [9,10] developed models comprising an ensemble of CNNs to detect COVID-19. Ardakani et al. [11] discussed the usage of ten convolutional networks on COVID-19 detection. Horry et al. [12] highlight the use of different image modalities to help faster detection of the disease. Shaban et al. [13] adopted a new methodology for feature selection by integrating filter and wrapper methods and classifying using an ensemble learning technique. Phankokkruad et al. [14] developed a transfer learning technique that involves fine-tuning of the pre-trained network. Rekha Hanumanthu et al. [15] discussed different deep learning and transfer learning methods adopted for the early diagnosis of the disease. Wang et al. [16] adopted a method where the features are extracted using a UNet, and later, the classification was performed using a progressive classifier. Hassantabar et al. [17] used a convolutional neural network, where the Softmax layer helps detect SARS-COV2 infection. Rahimzadeh and Attar et al. [18] used transfer learning methodology, which involves fine-tuning a concatenation of Xception and Resnet50V2 for diagnosing COVID-19. Phankokkruad et al. [14] implemented a model involving transfer learning experimented on three different pre-trained networks such as VGG16, Xception and InceptionResnetV2. Ucar and Korkmax et al. [19] presented an Artificial Intelligence structure based on Squeezenet pre-trained network accompanied by Bayesian optimization. Li et al. [20] explored a multi-task contrastive learning for COVID-19 diagnosis. The contrastive learning task has been implemented using supervised neural networks. The method involves aggregation through contrastive loss. Pandit et al. [21] adopted fine-tuning of VGG-16 network for the diagnosis of the COVID-19 from chest radiographs. Chandra et al. [22] utilized an ensemble learning methodology for the detection of coronavirus. The methodology involves majority voting from different weak learners. Saufi et al. [23] used stacked sparse autoencoders to extract features from X-ray and CT scans for the detection of COVID-19. Lazrag et al. [24] explored wavelet analysis for feature extraction followed by autoencoder for feature modelling to detect COVID-19. Behura et al. [25] used XGBoost and sparse autoencoder for feature selection and classification. Ismael and Sengur et al. [26] developed a model involving classification using Support Vector Machine(SVM) prospecting the features extracted from X-ray using Resnet50 for the diagnosis of COVID-19. Toraman et al. [27] accomplished a methodology for discovering COVID-19 infections using capsule networks utilizing X-ray lung imaging. Most of the existing works have used Convolutional Neural Networks for the detection of COVID-19. No methods have explored the combination of CNN with sparse autoencoder for the diagnosis of COVID-19. The proposed method chose to explore sparse autoencoder as a dimensionality reduction method. Sparse autoencoder has been found successful in the field of disease diagnosis, including Alzheimer's [28], Parkinson's disease [29], heart disease [30], identification of neonatal sleep state [31], glaucoma [32], cerebral microbleeds [33], etc to name a few. Sparse autoencoder enforces the sparsity constraint directing the single layer network for code learning resulting in error minimization while code reconstruction [34]. The sparsity penalty imposed on the hidden layers on top of the reconstruction error eliminates overfitting [35]. Sparse representation of data has benefits in denoising robustness and improved classification performance in high dimensional latent spaces [36]. The proposed work has the following contributions. The method uses an ensemble of Xception and InceptionResnetV2 for feature extraction. The features are passed to a custom made sparse autoencoder for reducing dimensionality of feature vector, followed by a Feed Forward Neural Network (FFNN) for classification. No state-of-the-art methods have employed such a pipeline for the detection of COVID-19. The proposed method explores neural network techniques for all stages of computer-aided diagnosis of the disease, namely feature extraction, dimensionality reduction and classification. The technique utilizes neural networks, namely, CNN, Sparse autoencoder, and FFNN, for feature extraction, dimensionality reduction, and classification. Most of the existing methods have either performed transfer learning using CNN or used non-neural network methods for feature selection and classification combined with features extracted using CNN. The method has chosen the ensemble of CNNs, sparse autoencoder and FFNN empirically based on experimental analysis. The experiments we performed prove the effectiveness of the novel framework composed of CNN, sparse autoencoder, and FFNN in diagnosing COVID-19. The rest of the paper is organized as follows. Section 2.1 discusses datasets used for training the model. Section 2.2 describes the architecture of the proposed model. Section 3 is a result analysis phase utilizing single-CNN and different combinations of pre-trained networks. The section also gives an overview of the comparison of the proposed model with other classifiers and dimensionality reduction techniques. Section 4 presents the conclusion of our work based on the result analysis phase.

Materials and methods

Dataset

Two publicly available datasets have been used to train the model. The first dataset is a public dataset by Cohen et al. [37], available in Github, consisting of both CT and Chest X-ray images of COVID-19 observations, other types of pneumonia and healthy patients. From this dataset X-ray images are filtered for training the model. The dataset consists of 783 X-ray images, among which 504 are COVID-19 images and 279 are non-COVID-19 images. The second dataset is a public open data set from Kaggle created by Paul Mooney [38]. The dataset consists of 390 chest X-ray images of bacterial and other viral pneumonia. It was constructed before the COVID-19 outbreak. A balanced dataset is essential for building an effective model [39]. To balance the dataset for achieving an effective model, the first 263 X-ray images of Pneumonia affected patients have been extracted from the second dataset. The combined dataset consists of 504 COVID-19 images and 542 non-COVID-19 images. The non-COVID-19 images consist of both normal and pneumonia images.

Proposed method

The proposed architecture is divided into three phases. The model consists of feature extraction, dimensionality reduction and the classification phase.

Feature extraction

Feature extraction phase reduces the dimension of the initial raw dataset into manageable groups for optimizing the processing. Many of the recent research studies have worked on models based on pre-trained networks as a feature extractor [9]. In the proposed model, pre-trained networks are used as a feature extractor. Pre-trained networks are trained on an Imagenet database [40] consisting of 1000 image classes. Even though trained on non-biomedical images, pre-trained CNNs in combination with off-the-shelf classifiers were found successful in the detection of a wide range of diseases from X-ray images, including tuberculosis [41], breast cancer [42] and pneumonia [43]. The convolutional layers built on top of each other, learn more complex features for reliable classification tasks. Automated feature extraction by CNN makes these networks highly efficient for classification tasks. In the proposed model, images are pre-processed according to the input size in the input layer of the chosen pre-trained model, and then the dataset is fed into the network. Both single-CNN and multi-CNN has been utilized for the analysis. InceptionResnetV2 [44], Xception [45], EfficientnetB0 [46], Darknet-53 [47], Resnet101 [48] are used for the experimentation. The input size of different pre-trained networks differs in size. Table 1 denotes some of the pre-trained networks, their depth and their input size used for the analysis. The images are pre-processed to the respective input sizes of the pre-trained models before the feature extraction phase. The dataset used for the proposed model consists of 1046 instances.
Table 1

Pre-trained networks.

NetworkDepthInput Size
InceptionResnetV2164299 × 299
Resnet101101224 × 224
Xception72299 × 299
EfficientnetB082224 × 224
Darknet5353256 × 256
Pre-trained networks. The method has used CNN as a feature extractor and not as a transfer learning method, where parameters of an end-to-end pre-trained CNN are fine-tuned to suit the data at hand. While using pre-trained CNN as a feature extractor, activations from any of the deep layers except Softmax layer can be used as features for classification using an off-the-shelf classifier like FFNN. The layer from which features are to be extracted is a design choice [49,50]. However, selecting features from the fully connected layer right before the Softmax classification layer is a good option [49,50]. The activations of the last fully connected layer represent global feature representation of the image [51,52]. Another function of the last fully connected layer is dimension reduction [53]. Softmax layer output the vector of probability values of an input image belonging to one among the 1000 classes and hence it cannot be used as a feature extractor. In the study by Abidin et al. [54], features extracted from the last fully-connected layer outperformed features from the other layers. Several research works have used the last fully-connected layer to extract features for classification using an off-the-shelf classifier [51,[55], [56], [57], [58], [59], [60], [61], [62], [63], [64]]. Based on the above-mentioned reasons, we have chosen activations of the last fully-connected layer as the feature vector. The output set after the feature extraction phase is a feature set with dimension 1046 × 1000. The CNN includes three basic layers: convolution layer, pooling layer and a Softmax layer. The center of the convolutional neural network is the convolution layer. Convolutional operation is performed in this layer which is the linear multiplication of the filter mask and the input array image to produce a feature map. Consider f(x, y) as the input image and h(x, y) be the filter mask. Convolution operation can be mathematically expressed as: An activation function is applied to the output of the convolution layer. The activation function used is ReLU. The next block after the convolution layer is the pooling layer. The most commonly used pooling layer is max-pooling layer. The pooling layer reduces the number of parameters used for the computation of the network. The final layer of the convolutional neural network is the classification layer, where the instances are classified according to the respective classes. For multi-CNN the features are extracted from two CNN models and then concatenated to produce a new feature set. The dimension of the feature set is n × 1000 m, where n is the number of X-ray images and m is the number of pre-trained CNN models.

Dimensionality reduction

Dimensionality reduction process reduces the dimension of feature vector by eliminating the features that will less contribute to the predictor variable. The presence of these irrelevant features may result in a decrease in the overall performance of the model. The proposed model uses sparse autoencoder for dimensionality reduction. Autoencoder is a network that imposes a bottleneck architecture, representing the input image in a compressed knowledge representation form [65]. The network follows an unsupervised learning technique for the task of representation learning [66]. The basic idea of the autoencoder is that it encodes the input sensor data using its hidden layer and outputs the best feature expression. The concept of autoencoder lies in taking in an unlabelled set and framing it as a supervised problem to get an output , which is a reconstruction of the original input x. The amount of information that traverses in the whole network is constrained in the bottleneck that drives a learned compression of the input image. Only the variations in the input data are maintained by the model for avoiding the redundancies. An autoencoder is composed of an encoder and a decoder. An encoder maps the input to a latent space using encoder activation. Later the input is reconstructed by the decoder using a decoder activation function. The activation of a basic autoencoder is represented as:where E denotes the pre-activation values, l is the squared loss function, g denotes the encoder activation function, and g indicates the decoder function. The cost function of an autoencoder is represented as:where , n, ρ and w denotes the reconstruction error loss, number of training samples (X-ray images), weight decay parameter and the weight at (i,j) location, respectively. Different kinds of activation autoencoders are available, among which our proposed model uses sparse autoencoder. Sparse autoencoder constructs a loss function, and the network is allowed to learn encoding and decoding, which relies only on a small number of neurons. Rather than regularizing the activations, sparse autoencoders regularizes the weights of the network. Activations of different nodes of the neural network are data-dependent, as different inputs will activate different neurons. Sparse autoencoders learn patterns by imposing sparsity constraints on the hidden layers [66]. The difference between the sparse autoencoder and the basic autoencoders lies in the cost function. Global regularization is used to solve the main objective function, whereas the sparsity penalty solves the trivial identity mapping and overfitting. In the model, features extracted using multiple pre-trained networks of the dimension 1046 × 1000 m, where m denotes the number of pre-trained networks is reduced to most relevant features of the dimension 1046 × 1000 using the sparse encoder for improving the performance of the model.

Classification

The model utilizes Feed Forward Neural Network for classifying the predictor variables. Feed Forward Neural Networks were found successful in a wide range of medical applications including Alzheimer's disease [67], chronic kidney disease [68] and lung cancer detection [69] to name a few. The wide usage of Feed Forward Neural Network in pattern classification is due to their prediction capability regardless of the probability distribution information of distinct labels. These networks gain efficiency from their parallel structure and their ability to improve their performance by experience. Hence, they can be used to efficiently classify the observations into different classes [67]. They can store the information in the network with less fault tolerance capacity. Advanced developments have proved these networks as the function approximators as they can approximate any arbitrary functions by fine-tuning the number of hidden layers and their parameters [70]. The network includes connections, with each of the links designated to different weights. The information flow in the network exists only in one direction. The output of the previous layer serves as the input for its successive layer. The output of a neuron is represented as the weighted sum of its inputs. Equation (4) depicts the weighted sum of inputs of a layer.where represents the weighted sum of inputs, m represents the number of nodes in (k − 1) layer, represents the output of the pth node in (k − 1) layer and w denotes the weight of the link. Equation (5) denotes the output of a layer.where h represents the output function and p ranges from 1 to m . Error loss is computed from the predicted and actual output by reforming the connection weights. Equation (6) depicts the error loss.where δ represents the error loss, r and t are the predicted and actual outputs, respectively. Accounting the error loss, the connection weights are updated. The weights are updated such that the δ values are minimized. In each epoch of training, the weight reformation process happens for minimized error loss. Weight reformation takes place from the last layers proceeding towards the lower layers. Fig. 1 shows the diagrammatic representation of the model. The set f 1, f 2 …. , f denotes the feature set extracted by the pre-trained networks, denotes the features after reducing dimensionality and h 1, h 2, h 3, h 4, …‥, h 10 denotes the 10 nodes in the hidden layer of FFNN. These features are fed as input to the Feed Forward Neural Network.
Fig. 1

Architecture of the proposed method.

Architecture of the proposed method.

Results and discussions

Experimental setup

The experimentation was performed on Intel core i5 processor with GPU support of 4 GB and 8 GB RAM. The model has been implemented using MATLAB. Neural network toolbox by Jingwei Too [71] is used for implementing FFNN. Accuracy, F1-Score, Precision, Specificity, Sensitivity, Area Under Curve(AUC) and Matthews Correlation Coefficient (MCC) have been computed to evaluate the model.

Classification results

The experiments were performed on both single pre-trained networks and concatenation of multiple pre-trained networks. For single CNN, the features were extracted and passed to Feed Forward Neural Network. Ten-fold cross-validation has been performed for analysing the model. Random partitioning of data into 10 equal folds with 9 folds of data treated as training and the remaining 1 fold serves as the testing dataset at each iteration. A single iteration takes 90% of data for training and 10% for testing. Analysing the model using a single pre-trained network as the feature extractor and a Feed Forward Neural Network as an off-the-shelf classifier exhibits good performance. The pre-trained networks used for the analysis phase are EfficientnetB0, Resnet101, Darknet-53, InceptionResnetV2 and Xception. The last fully connected layers of the pre-trained networks have been used as the feature extractor, which outputs a feature set of dimension 1046 × 1000. EfficientnetB0, Resnet101, Darknet-53, InceptionResnetV2 and Xception was able to achieve an accuracy of 0.9321, 0.9216, 0.86043, 0.9149 and 0.9081, respectively. It was noticed that EfficientnetB0 achieved the highest performance among the five networks used for training. The specificity, sensitivity, precision, F1-Score and AUC of EfficientnetB0 was 0.9452, 0.9188, 0.9305, 0.9425 and 0.9756, respectively. Out of 504 COVID radiographic images, 475 instances were correctly classified by the FFNN. Among 542 non-COVID-19 images, 500 instances were correctly classified. Fig. 2 represents the graphical analysis of the model performance using single-CNN without using the dimensionality reduction module. Table 2 indicates the performance of the model using the Feed Forward Neural Network as the off-the-shelf classifier and without performing dimensionality reduction using sparse autoencoder. Using multi-CNN, the greatest performance was achieved using the combination of Xception and EfficientnetB0 with an accuracy of 0.9301. Out of 542 instances, 461 instances were predicted correctly. Fig. 3 depicts the graphical analysis of the multi-CNN model without using the dimensionality reduction module.
Fig. 2

Graphical analysis of model performance using Single-CNN without using sparse autoencoder.

Table 2

Performance of the model using CNN and FFNN, without using sparse autoencoder.

Pre-trained modelSpecificitySensitivityF1-ScorePrecisionAccuracyAUCMCC
InceptionResnetV2+Xception0.92060.93500.92370.91270.92730.97500.8546
InceptionResnetV2+Resnet1010.92060.93500.92370.91270.92730.97500.8546
Xception + Resnet1010.92710.91000.91630.92260.91870.97330.8374
Darknet53+Resnet1010.91170.94580.92280.90080.92720.95640.8552
Xception + EfficientnetB00.92250.93890.91470.92660.93010.97870.8604
Xception0.91450.90160.90870.90510.90810.95360.8163
Darknet530.83790.88910.81150.84850.86040.94720.7222
Resnet1010.90210.94510.88890.91620.92160.97270.8441
EfficientnetB00.94520.91880.94250.93050.93210.97560.8645
InceptionResnetV20.90660.92430.89680.91040.91490.95990.8298
Fig. 3

Graphical analysis of model performance using Multi-CNN without using sparse autoencoder.

Graphical analysis of model performance using Single-CNN without using sparse autoencoder. Performance of the model using CNN and FFNN, without using sparse autoencoder. Graphical analysis of model performance using Multi-CNN without using sparse autoencoder. Dimensionality reduction using sparse autoencoder was performed in the second phase of analysis, which shows an improved result. Table 3 indicates the model's performance while using Sparse autoencoder and Feed Forward Neural Network. The best performance with an accuracy of 0.9578, F1-Score of 0.9563 and precision of 0.9563 was achieved with the combination of InceptionResnetV2 and Xception. The computation time taken by the model for testing 10% of the data is 276.87 s. The computation time taken by Darknet53+Resnet101, InceptionResnetV2+Resnet101, Xception+Resnet101, and Xception + EfficientnetB0 were 550.938 s, 397.66 s, 367.03 s and 312.18 s, respectively.
Table 3

Performance of the proposed model using Sparse Autoencoder and FFNN.

Pre-trained modelSpecificitySensitivityF1-ScorePrecisionAccuracyAUCMCC
InceptionResnetV2+Xception0.95940.95630.95630.95630.95780.98210.9158
Darknet53+Resnet1010.94230.96130.94870.93650.95110.98000.9025
InceptionResnetV2+Resnet1010.92270.96440.93780.95040.94170.98050.8842
Xception + Resnet1010.94540.95370.94710.940 50.94930.98930.8986
Xception + EfficientnetB00.94360.95360.94600.93850.94830.97920.8967
Performance of the proposed model using Sparse Autoencoder and FFNN. To ensure that the results are statistically significant, p-value based on chi-square has been computed for the best performing model of InceptionResnetV2 and Xception. A p-value less than 0.00001 is achieved that show that the results are statistically significant at p less than 0.05. Statistical model evaluation test, Matthews Correlation Coefficient (MCC) measure has also been analyzed for estimating model performance. MCC values close to one account for a strong correlation between predicted and the actual class. The highest MCC value of 0.9158 exhibited by the combination of Xception and InceptionResnetV2 denotes that the model is worth for distinguishing COVID-19 and non-COVID-19. Xception networks denote extreme inception, with the inception architectures as the backbone of these networks. The convolutions in the original inceptions modules are restored with depthwise separable convolutions in Xception networks. This correlation scanning of 2D followed by 1D mapping is easier and more effective than full 3D mapping [72]. InceptionResnetV2, on the other hand, is an updated inceptionv3 network capable of better performance achievement than other convolutional networks. A combination of the inception block followed by the residual block in the architecture and the shortcut connections adds to the performance enhancement of the model. Thus the concatenation of these two efficient networks results in improved feature set generation and improved results. The combination of InceptionResnetV2 and Resnet 101 achieved the highest sensitivity of 0.9644 and the combination of Xception and Resnet101 achieved the highest AUC of 0.9893. Even though the above-mentioned combinations achieved better sensitivity and AUC, the combination of InceptionResnetV2 and Xception outperformed the other combinations in all other performance measures. The experiments were repeated multiple times to ensure the stability of the results. As seed values were used to generate random weights and cross-validation folds, we could reproduce the same results each time. The results show that the accuracy of the model has been improved with sparse encoder dimensionality reduction technique. However, no pre-trained CNNs were able to achieve 100% accuracy. Few X-ray images were misclassified in all the methods. The major cause of false positives and false negatives are the similarities in X-ray images of COVID-19 and pneumonia images, which makes the accurate prediction difficult. Among 504 COVID-19 instances, 482 images were correctly classified and the remaining 22 images were misclassified using the best performing combination of Xception and InceptionResnetV2 with sparse autoencoder and FFNN. Out of 542 non-COVID-19 images, 520 images were correctly classified, and 22 images were misclassified. The statistical MCC test has proven that the model has improved its efficiency by incorporating sparse autoencoder technique. Fig. 4 represents the graphical analysis of the model after performing dimensionality reduction using sparse autoencoder. Fig. 5 presents the graphical analysis of the accuracy comparison, integrating dimension reduction and without dimension reduction methodology.
Fig. 4

Graphical analysis of Multi-CNN model performance after performing dimension reduction using sparse autoencoder.

Fig. 5

Comparison of accuracy before and after employing sparse autoencoder as the dimensionality reduction technique.

Graphical analysis of Multi-CNN model performance after performing dimension reduction using sparse autoencoder. Comparison of accuracy before and after employing sparse autoencoder as the dimensionality reduction technique. To compare the performance of deep pre-trained CNN with shallow network, we further performed feature extraction using a single hidden layer sparse autoencoder and classification using FFNN. Feature extraction using the shallow sparse autoencoder and classification using FFNN achieved accuracy, specificity, sensitivity, f1-score, precision and AUC of 0.8528, 0.8345, 0.8755, 0.8412, 0.8095 and 0.9217, respectively. The proposed method, which used deep pre-trained CNNs for feature extraction significantly outperformed feature extraction using shallow sparse autoencoder.

Parameter setting

For training the model using sparse autoencoder and FFNN, some of the parameters have been assigned. The parameters for the model construction has been empirically initialized using trial and error method. Table 4 denotes the parameter setting used for training the model. The hidden size parameter for sparse autoencoder specifies the number of features to be extracted. The model is trained for an epoch of 100.
Table 4

Parameter setting for Sparse autoencoder & FFNN.

Parameters(Sparse autoencoder)Value
Hidden Size1000
Random Seed2
L2WeightRegularization0.001
Sparsity Regularization4.000
Sparsity Proportion0.0500
Decoder Transfer Function
purelin
Parameters(FFNN)
Value
Hidden Layers1
Random Seed1
Maxepochs100
K-Fold10
trainFcntrainlm
Netfeed forward net
Parameter setting for Sparse autoencoder & FFNN.

Comparison of the results with other feature selection techniques

Sparse autoencoder has been empirically chosen for dimensionality reduction after performing experimental analysis with two major feature selection techniques, namely, Principal Component Analysis (PCA) and Correlation Feature Selection (CFS). Attribute selection using CFS and PCA has been performed on Weka 3.6. The selected features were then passed to feed forward neural network. Table 5 presents the results obtained while passing the feature extracted from InceptionResnetV2 and Xception to different feature selection techniques. CFS and PCA were able to achieve an accuracy of 0.8556 and 0.8899, respectively. Among PCA and CFS, the proposed Sparse Autoencoder dimensionality reduction technique has proven its effectiveness with an accuracy of 0.9578. The Sparse Autoencoder has achieved an AUC of 0.9821, while CFS and PCA have acquired an AUC of 0.9326 and 0.9469, respectively. The time consumption for the sparse autoencoder, CFS and PCA were 190.308 s, 26 s and 260 s, respectively.
Table 5

Performance achieved using various dimensionality reduction/feature selection techniques, in combination with the proposed pre-trained model and FFNN.

Pre-trained modelMethodSpecificitySensitivityF1-ScorePrecisionAccuracyAUC
InceptionResnetV2+XceptionProposed0.95940.95630.95630.95630.95780.9821
InceptionResnetV2+XceptionCFS0.90310.81460.85820.90670.85560.9326
InceptionResnetV2+XceptionPCA0.92280.85950.89000.92260.88990.9469
Performance achieved using various dimensionality reduction/feature selection techniques, in combination with the proposed pre-trained model and FFNN. The superior performance of sparse autoencoder is attributed in the following. Sparse autoencoders learn the data projections more efficiently with the dimension and the sparsity constraints rather than the other feature selection techniques. Autoencoder networks learn nonlinear transformations and are also more constructive in terms of model parameters with various layers than PCA with a single transformation [34].

Comparison of the results with other classifiers

FFNN has been empirically chosen as off-the-shelf classifier from experimental analysis with various other classifiers. Table 6 presents the performance of various classifiers with the best performing multi-CNN and sparse autoencoder. It is evident that only FFNN achieved an accuracy above 90%. Accuracy of Bayesnet, SVM, KNN, Random Forest and Adaboost falls in the range of 0.70–0.80. Among the different classifiers used for training, FFNN has achieved an accuracy of 0.9578 using InceptionResnetV2 and Xception as the backbone of the model and Sparse Autoencoder as the dimensionality reduction technique. The superior performance of FFNN owes to its ability to arrest more complex patterns. The time taken for the classification phase by the FFNN, Bayesnet, SVM, KNN, Random Forest and Adaboost were 0.128 s, 0.81 s, 2.17 s, 0.87 s, 2.1 s and 3.35 s, respectively.
Table 6

Performance achieved using various classifiers, in combination with the proposed pre-trained model and sparse autoencoder.

Pre-trained modelClassifierSpecificitySensitivityF1-ScorePrecisionAccuracyAUC
InceptionResnetV2+XceptionBayesnet0.82990.70380.77130.85320.75620.7970
InceptionResnetV2+XceptionSVM0.79490.78990.78280.77580.79250.7920
InceptionResnetV2+XceptionKNN0.78510.79300.77610.75990.78870.7850
InceptionResnetV2+XceptionRandom Forest0.86670.74500.80730.88100.79730.8790
InceptionResnetV2+XceptionAdaboost0.80170.70790.75870.81750.74950.8300
InceptionResnetV2+XceptionProposed0.95940.95630.95630.95630.95780.9821
Performance achieved using various classifiers, in combination with the proposed pre-trained model and sparse autoencoder.

Comparison with other state-of-the-art methods

Different state-of-the-art methods were analyzed for proving the effectiveness of the proposed model. Table 7 presents a consolidation of results achieved by other state-of-the-art methods and the proposed method. The comparison has considered only the methods using X-ray images for the classification of COVID-19 and non-COVID-19 images. The methods employing CT scans and ultrasound for COVID-19 detection have not been considered, as they are entirely different modalities.
Table 7

Comparison of the results achieved by the proposed method with other state-of-the-art techniques.

MethodNumber of imagesSpecificitySensitivityF1-ScorePrecisionAccuracyAUC
Proposed504 COVID-19 vs. 542 non-COVID-190.95940.95630.95630.95630.95780.9821
Abraham et al. [9]453 COVID-19 vs. 497 non-COVID-19-0.98500.91400.85300.91150.9630
Li et al. [20]231 COVID-19 vs. 1583 Normal0.91910.9297--0.97230.9213
Pandit et al. [21]224 COVID-19 vs. 1204 non-COVID-190.97270.9264--0.9600-
Hassantabar et al. [17]315 COVID-19 vs. 367 non-COVID-19-0.9610--0.9320-
Chandra et al. [22]696 COVID-19 vs. 696 non-COVID-19----0.91320.8310
Sethy et al. [73]25 COVID-19 vs. 25 non-COVID-19--0.9141-0.9538-
Waheed et al. [74]403 COVID-19 vs. 721 Normal0.97000.9000-0.95600.9500-
Panwar et al. [75]142 COVID-19 vs. 142 Normal0.97000.9000-0.95600.88100.8809
Hemdan et al. [76]25 COVID-19 vs. 25 non-COVID-19-1.00000.91000.8300--
Zhang et al. [77]100 COVID-19 vs. 1431 non-COVID-19-0.9600---0.9500
Ismael and Sengur et al. [26]180 COVID-19 vs. 200 non-COVID-19----0.9470-
Toraman et al. [27]231 COVID-19 vs. 500 non-COVID-190.80950.96000.93750.91600.9124-
Comparison of the results achieved by the proposed method with other state-of-the-art techniques. The number of instances used for training the model is different for multiple methods analyzed. The methods by Pandit et al. [21], Panwar et al. [75], Sethy et al. [73], Ismael and Sengur et al. [26] and Hemdan et al. [76] used a held-out validation set for the evaluation, whereas the other methods have used cross-validation. The results are on par with the state-of-the-art methods. Even though all the methods have achieved significant results, the proposed method achieved a better AUC, F1-score and precision than the other methods. Fair comparison between the different results analyzed are not possible since each method has utilized different number of images for the study.

Limitations and future research directions

Even though the method achieved significant results, some of the constraints are worth noting. The method is developed for binary classification of COVID-19 and non-COVID-19 images. The model has not been explored in a multi-class scenario in classifying normal, COVID-19 and pneumonia images. Also, a wide range of sparse autoencoder parameter values were not experimented in developing the model, which can be customized using the grid search method for relatively higher model performance. The method developed is specific to the diagnosis of COVID-19 from X-ray images. However, after empirical studies, it can be extended to diagnose other diseases. The method seems to have prospects in diagnosing other lung diseases and diseases that can be detected using X-rays. The strategy can also be applied to other imaging modalities, after customization. As a future research study, we propose applying the method for diagnosing COVID-19 from other imaging modalities such as CT and Ultrasound.

Conclusion

The proposed model implements a computer-aided model for COVID-19 detection utilizing chest X-ray images using Sparse Autoencoder and Feed Forward Neural Network. The concatenation of multiple pre-trained networks for feature extraction has been implemented which outperforms single-CNN. The usage of Sparse Autoencoder has greatly contributed in improving the accuracy of the model. It is worth noting that the performance of the model has considerably increased with the usage of the dimensionality reduction phase rather than using the Feed Forward Neural Network alone. From the analysis phase it is observed that combination of Xception and InceptionResnetV2 achieved greatest accuracy in combination with the custom-made sparse autoencoder and FFNN.

Declaration of conflict of interest

The authors declare that there are no conflicts of interest in this work.
  8 in total

1.  Rethinking U-Net from an Attention Perspective with Transformers for Osteosarcoma MRI Image Segmentation.

Authors:  Tianxiang Ouyang; Shun Yang; Fangfang Gou; Zhehao Dai; Jia Wu
Journal:  Comput Intell Neurosci       Date:  2022-06-06

2.  An agent-based transmission model of COVID-19 for re-opening policy design.

Authors:  Alma Rodríguez; Erik Cuevas; Daniel Zaldivar; Bernardo Morales-Castañeda; Ram Sarkar; Essam H Houssein
Journal:  Comput Biol Med       Date:  2022-07-19       Impact factor: 6.698

3.  A lightweight CNN-based network on COVID-19 detection using X-ray and CT images.

Authors:  Mei-Ling Huang; Yu-Chieh Liao
Journal:  Comput Biol Med       Date:  2022-05-11       Impact factor: 6.698

Review 4.  Machine learning applications for COVID-19 outbreak management.

Authors:  Arash Heidari; Nima Jafari Navimipour; Mehmet Unal; Shiva Toumaj
Journal:  Neural Comput Appl       Date:  2022-06-10       Impact factor: 5.102

5.  A modified DeepLabV3+ based semantic segmentation of chest computed tomography images for COVID-19 lung infections.

Authors:  Hasan Polat
Journal:  Int J Imaging Syst Technol       Date:  2022-06-11       Impact factor: 2.177

6.  A Novel Method for COVID-19 Detection Based on DCNNs and Hierarchical Structure.

Authors:  Yuqin Li; Ke Zhang; Weili Shi; Zhengang Jiang
Journal:  Comput Math Methods Med       Date:  2022-08-31       Impact factor: 2.809

7.  AI-CenterNet CXR: An artificial intelligence (AI) enabled system for localization and classification of chest X-ray disease.

Authors:  Saleh Albahli; Tahira Nazir
Journal:  Front Med (Lausanne)       Date:  2022-08-30

8.  Multi-texture features and optimized DeepNet for COVID-19 detection using chest x-ray images.

Authors:  Anandbabu Gopatoti; Vijayalakshmi P
Journal:  Concurr Comput       Date:  2022-08-01       Impact factor: 1.831

  8 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.