Literature DB >> 35875199

Deep learning-based important weights-only transfer learning approach for COVID-19 CT-scan classification.

Tejalal Choudhary1, Shubham Gujar2, Anurag Goswami1, Vipul Mishra1, Tapas Badal1.   

Abstract

COVID-19 has become a pandemic for the entire world, and it has significantly affected the world economy. The importance of early detection and treatment of the infection cannot be overstated. The traditional diagnosis techniques take more time in detecting the infection. Although, numerous deep learning-based automated solutions have recently been developed in this regard, nevertheless, the limitation of computational and battery power in resource-constrained devices makes it difficult to deploy trained models for real-time inference. In this paper, to detect the presence of COVID-19 in CT-scan images, an important weights-only transfer learning method has been proposed for devices with limited runt-time resources. In the proposed method, the pre-trained models are made point-of-care devices friendly by pruning less important weight parameters of the model. The experiments were performed on two popular VGG16 and ResNet34 models and the empirical results showed that pruned ResNet34 model achieved 95.47% accuracy, 0.9216 sensitivity, 0.9567 F-score, and 0.9942 specificity with 41.96% fewer FLOPs and 20.64% fewer weight parameters on the SARS-CoV-2 CT-scan dataset. The results of our experiments showed that the proposed method significantly reduces the run-time resource requirements of the computationally intensive models and makes them ready to be utilized on the point-of-care devices.
© The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature 2022.

Entities:  

Keywords:  Automated diagnosis; COVID-19; Convolutional neural network; Deep learning; Pruning

Year:  2022        PMID: 35875199      PMCID: PMC9289654          DOI: 10.1007/s10489-022-03893-7

Source DB:  PubMed          Journal:  Appl Intell (Dordr)        ISSN: 0924-669X            Impact factor:   5.019


Introduction

In recent times, millions of people have been infected worldwide with the COVID-19 pandemic. According to the WHO, more than 231M people have been infected with the COVID-19 to date leading to 4.74M deaths worldwide [1]. In the USA alone, 42.5M people were infected, and 681K people lost their lives. Moreover, the world’s second-largest populated country, India, lost 447K lives, and 33.6M people get infected [1]. COVID-19 infection causes mild to severe respiratory disease. It is more dangerous for older persons and those with other disease such as cardiovascular, diabetes, chronic respiratory illness, and cancer. The most prevalent COVID-19 symptoms are dry cough, fever, and fatigue. However, the serious symptoms include loss of smell and taste, trouble in breathing, chest discomfort, aches, and pain [2]. A wide variety of tests are available to determine the presence of COVID-19. Some of the popular test includes reverse transcription-polymerase chain reaction (RT-PCR) [3] also known as molecular test, antigen test [4] for rapid testing, antibody test [5] also known as serology test, and using radiological images; CT-scan and chest X-ray [6, 7]. The antigen test is faster and inexpensive when compared to the RT-PCR. It has high false-positive rates and is thus less reliable. The rapid antibody test looks for antibodies in the patient’s blood sample. The RT-PCR test is the most prevalent test that is utilized for the detection of COVID-19. If the antibody test along with RT-PCR is positive, it confirms that the person is COVID-19 positive. However, one of the main disadvantages of RT-PCR is that it is time-consuming. The COVID-19 virus spreads quickly; by the time it gets detected, it can spread to others. The COVID-19 virus can also be detected using chest computed tomography (CT) scan images. A bilateral change can be observed in these CT-scan images [7, 8]. Moreover, the examination of the CT-scan images is challenging, time-consuming, and requires skilled radiologists. The radiologist’s opinion also suffers from inter-observer variability [9]. In recent years, computerized medical diagnosis have gained a lot of attention. In the medical field, artificial intelligence (AI) methods such as deep learning have been applied in a variety of diagnosis [10, 11]. In addition, a convolutional neural network (CNN) is a deep learning variant capable of learning discriminating properties from images. CNN has already proven its superior performance on different classification [12], [13, 14], segmentation [15, 16] and detection [17] applications in the past compared to the traditional methods. Further, an affordable and acute medical facility in remote areas is limited due to the lack of specialized laboratories. Point-of-care devices have the ability to play an important role in generating quick diagnosis results in this regard. For practical applications, deploying the trained deep learning model onto the point-of-care devices is must for a widespread use. However, the deep learning models suffer from over-parameterization; they contain millions of learnable weight parameters and during inference, perform a lot of floating-point operations (FLOPs) [18]. When these models are utilized for transfer learning, their generalization performance on medical datasets becomes poor. Hence, this limits their deployability on the point-of-care devices as the deep learning models need an adequate amount of resources to execute the trained model, i.e. computational, battery power, and memory [19, 20]. Another issue with deep learning models is that they were trained on the massive ImageNet dataset containing 1.2 million images from 1000 classes. However, the datasets in the medical field contain less number of classes. Therefore, a lot of weight parameters become redundant and less important in the network and it also results in low generalization. As a result, it is important to reduce run-time resource requirements before deploying trained models onto point-of-care devices. In this regard, filter pruning [21-23] has become a popular technique for reducing the amount of computational and battery power required during inference and improving the overall inference performance. Owing to the ability of the deep learning models to perform superior in various automated medical diagnosis, the advantages of bringing intelligence to point-of-care devices and their limited resource ability motivated us to develop an inference efficient method for COVID-19 classification. Particularly, in this paper, we proposed an important weights-only transfer learning approach to automatically classify the CT-scan images as COVID-19 infected or not in the context of point-of-care devices. We propose to prune less essential filters from the convolutional layers to lower the insignificant weight parameters of the pre-trained models. Thus, while training the models on the CT-scan dataset only important weight parameters are used and CT-scan images are then classified using the pruned models. The proposed method is significantly different from [18, 21, 22], and [23]. The work presented in [18] is similar to ours, the authors worked on pruning the smaller weight filters from the convolutional layers. However, in contrast to [18] wherein the authors worked with invasive ductal carcinoma (IDC) dataset for classifying breast cancer images. In our research, we evaluate the performance of the proposed approach on the SARS-Cov-2 CT-scan dataset. Further, in [18] the authors worked with 50 × 50 × 3 images, whereas the SARS-Cov-2 CT-scan dataset has relatively higher dimension images. In [21], the authors utilized batch normalization parameters to find the unimportant filters. However, according to [21], to identify the weak filters, the model needs to be trained from the scratch and thereby limits its applicability for the standard pre-trained networks. In another words, the pruning of pre-trained networks require retraining on the ImageNet dataset. Unlike [21], there is no such restriction with the proposed method. Moreover, the proposed approach differs from [22], where the authors represent pruning as an optimization problem and prune the filters depending on the next layer’s statistics instead of the current layer. We believe that pruning the filters based on the current layers’ filter statistics is more significant. The authors of [23] proposed asymptotic soft filter pruning in which the filters are set to zero before first training epoch and during the retraining, the authors devise an strategy to update the previously pruned filters. The same procedure is repeated for the rest of the training epochs. This also requires model to be trained from scratch to identify the pruning candidates. Moreover, setting filter weights to zero does not completely remove the filters and makes the network sparse. In contrast, our method is aimed at finding the pruning candidate filters and completely removing the filters and their respective feature maps. Following are the main contributions of the paper: A novel important weights-only transfer learning approach for the classification of COVID-19 CT-scan images. We proposed transferring only the significant weights to decrease the models’ run-time resource requirements by pruning the least important weights. To identify the less important filters of the model, we evaluate the importance of each filter based on their absolute sum. Performed multiple experiments with both the models; unpruned pre-trained models and the pruned. The experiments were performed on the SARS-CoV-2 CT-scan dataset [24], and the experimental results indicate the superiority of the work presented here. This paper has been organized as follows. Section 2 presents the related work wherein the various deep learning methods proposed for the classification of COVID-19 are summarized. Section 3 provides the detail of the proposed important weights only transfer learning approach. The experiments performed to validate the method’s effectiveness, dataset & models used, their results, and analysis is elucidated in Section 4. Finally, the discussion on the proposed methods and conclusion is included in Sections 5 and 6, respectively.

Related work

This section contains the details of the existing research methods wherein the authors show the need to ship the deep learning models on resource-limited devices. Further, the section also provides an overview of the important transfer learning and filter pruning methods as the proposed work is based on these, followed by a review of the existing research methods proposed for COVID-19 CT-scan image classification. The popularity of the deep learning methods has opened tremendous opportunities for researchers to extend their uses for on-device practical applications on resource-limited devices. In recent years, there is a growing demand to move the inference step from high computing machines to resource-limited devices [25]. In this context, deploying trained deep learning models on resource-limited devices is becoming more prevalent for a variety of applications [26, 27]. Multiple studies have shown the need and importance of deploying deep learning models on point-of-care devices as well [28, 29]. However, due to a large number of multiply and accumulate (MAC) operations and memory access operations, a typical deep learning application can quickly exhaust resource-limited device [30]. In [31], the authors show that deep learning architectures contain enormous parameters, requiring large storage space and computational resources. Transfer learning plays an important role to improve the performance of the deep learning models wherein the knowledge learned by the model trained on the source domain is applied to a target domain. The research has shown that transfer learning significantly enhances model performance. In [13] and [18] it was found that the model trained using transfer learning performed superior compared to the model trained from scratch. Apart from transfer learning, in recent years, to improve the inference performance of the deep learning models, filter pruning has emerged as an important technique [32]. The researchers have shown that pruning not only speeds up the inference performance but also reduces overfitting problems and helps the model in learning clear and expressive features. In a research [33], the author proposed a flexible-rate pruning method to compress and accelerate the trained models in which to identify the filters to be pruned, the authors employ a greedy-based strategy and execute an iterative loss-aware pruning procedure. In another research [34], the authors argue that the filters are affected by other filters, therefore, only the magnitude is not enough to decide the importance of the filters. The authors proposed a meta-attribute-based pruning method in which the geometric distance of filters is considered as a pruning criterion. The method of [35], pruned the filters by calculating the learned representation median in the frequency domain in contrast to the existing method that prunes less important filters in the spatial domain. In short, there are various filter pruning methods to accelerate and compress the deep learning models, however, there are limited studies evaluating and analyzing pruning methods in the context of point-of-care devices. Moreover, the deep learning model has shown outstanding results in classifying CT-scan images over traditional machine learning methods. In a research [36], the authors proposed a multi-modality medical image diagnosis approach using deep neural networks. The non-subsampled contourlet transform (NSCT) domain method was used to construct multi-modality images. In another research [37], the authors proposed automatic segmentation and classification method for COVID-19 CT-scan images. The authors used a total of 1069 images for training and 150 images were used for validation and testing. The U-Net model was compared for segmentation and the AlexNet, DenseNet, ResNet50, InceptionV3, VGG16, VGG19, and EfficientNet were compared with the proposed architecture for classification. The segmentation architecture performed better than U-Net and obtained 88% accuracy, while the classification architecture achieved 94.67% accuracy. In a research [38], the authors proposed a transfer learning-based classification approach for the detection of COVID-19 from CT-scan images. The experiments were performed on a total of 2492 images. A pre-trained DenseNet201 architecture was used as a base model. The authors also compared the results with other CNN architectures; VGG16, InceptionResNet, and ResNet152V2. The DenseNet model outperformed other CNN models and obtained 96.25% accuracy on the test dataset. In one of the research [39], a multi-objective differential evolution-based approach was proposed to classify COVID-19 CT-scans. Compared to the other methods, the proposed approach achieved 1.97% better accuracy. The authors of the research [40] proposed an automated classification approach using an ensemble of densely connected convolutional networks. An ensemble model was trained using ResNet152V2, DenseNet201, and VGG16 models. The authors found that compared to the other methods, the proposed method obtained 1.27% improved accuracy. In another study [41], machine learning, deep learning, and texture analysis based approaches were compared for early diagnosis. Along with texture analysis methods, support vector machines (SVM) and K-nearest neighbors (KNN) were implemented. Moreover, a 23-layer custom CNN was also trained other than pre-trained AlexNet and MobileNetV2. In a study [42], the authors proposed a stacked ensemble of pre-trained CNN architectures. The authors implemented four different architectures and performed extensive experimentation with the stacked ensemble. The proposed model achieved 90.75% accuracy. Similarly, in another study [43], the authors proposed a stacked ensemble-based method and claimed higher recall. The proposed method achieved 94% accuracy and 0.98 recall. In one of the research [44], a combination of transfer learning with the U-Net segmentation architecture was proposed to classify the SARS-Cov-2 CT-scan dataset. The pre-trained DenseNet-169 on U-Net segmented data achieved 89.92% classification accuracy. In a research [45], the authors proposed a novel redesigned framework from existing COVID-Net along with contrastive learning for enhanced training. The proposed method achieved 90.83% accuracy. In another research [46], a transfer learning-based DenseNet-121 architecture was proposed and achieved 92% accuracy. The authors of [47] also proposed a transfer learning-based method to classify the COVID-19 CT-scan images. The authors also implemented grad-CAM based color visualization to interpret the predictions. The method achieved 95.61% accuracy. From the study of the existing methods, it was found that none of the existing methods were designed to address the constraints of the resource-limited devices. On the other hand, our proposed approach aims at reducing the inference-time requirements of the models to enable faster and acute diagnosis in resource-limited devices.

Methodology

This section elucidates the details of the proposed important weights-only transfer learning approach. The overall method can be divided into different parts; pruning, fine-tuning, and training. In a nutshell, first, we propose to prune the filters that are least important from the convolutional layers. The pruned models were then retrained using the ImageNet dataset in the second part of the study. Finally, the resulting fine-tuned models were then used for training and testing on the SARS-CoV-2 CT-scan dataset. Figure 1 shows the overall pruning, fine-tuning, and training pipeline. Further, the main contributions of the proposed work includes: 1) A novel important weights-only transfer learning approach for the classification of COVID-19 CT-scan images. 2) To reduce the models’ run-time resource requirements, we proposed transferring only the significant weights by pruning the least important weights. 3) In order to identify the less less important filters of the model, we evaluate the importance of each filter based on their absolute sum. 4) The effectiveness of the proposed work is validated through multiple experiments with both the models; unpruned pre-trained models and the pruned. The experiments are performed on the SARS-CoV-2 CT-scan dataset [24]. The following subsections contain the details of the proposed methodology.
Fig. 1

Proposed important weights-only transfer learning approach

Proposed important weights-only transfer learning approach

Step I: prune the least important filters

Convolutional filters are the backbone of any CNN architecture. However, in earlier research, it is found that not all filters are essential, and the removal of a few filters can be done without a significant accuracy loss [48]. In this context, filter pruning has become popular in recent years. It does not necessitate any model architectural changes. The resulting model can be deployed without requiring any additional hardware or software for acceleration. Particularly in our research work, a layer-by-layer filter pruning was performed. One of the important tasks in filter pruning is to assess the filters’ importance. The filter’s magnitude, impact on loss/error, and batch normalization parameter can all be used to evaluate the filter’s relevance. Different approaches can find distinct pruning filters. In order to determine the optimal criterion, we conducted several experiments. To find the number of pruning candidate filters using the batch normalization parameter, the model needs to be trained from the scratch. There are already pre-trained models on the ImageNet dataset, hence, training from scratch makes it computationally expensive. In a different experiment, the filters were pruned based on their impact on the loss/error to determine whether the removal of the filters increases or decreases the loss/error. We found that pruning the filters based on their impact of the loss/error is a time-consuming process. It requires the model to be trained after the removal of each filter. The magnitude of the weight parameters impact filters activations. The filter with a smaller magnitude generates weaker activation. Hence, we considered small magnitude (absolute sum) filters less important than the filters with larger magnitude. Each convolutional layer k in the model produces the output as A(, where u, v, and c correspond to the height, width, and the channels, respectively. The generated output A(, works as an input for the k + 1 convolutional layer. Each convolutional filter generates one feature map when all the feature maps are combined they produce the feature maps of size A(. In the proposed approach, filters that fail to meet the evaluation criterion were pruned since the objective was to remove the less significant filters from the trained model. Further, we employed a binary masking approach to designate whether or not any filter in the layer would be pruned. All the filters of the layer were first sorted in decreasing order by their absolute sum. The aim of the research was to create an optimal model for point-of-care devices with the minimal number of convolutional filters while least compromising the model performance. Let N represent the original model, N represent the pruned model and N has K convolutional layers, the k layer is given by k[, and l ∈ (1,2,…,K). The number of filters for layer k[ is n and the generated activation map is given by Q. The activation Q works as input for the subsequent layer. Further, = represent the set of filter for layer k[. The original model weights of layer k[ is and pruned model weights of the k[ layer is , n≠n. For dataset and a pruning threshold p, the filter pruning problem is formulated as: In (2), is the standard cross entropy loss. If the ℓ1-norm or the absolute sum of filter g is given by γ and , the norm of filters g is ∥γ∥ and defined as. For every γ, ∥γ∥≥ 0 for all , and ∥γ∥= 0 iff γ= 0. Let represent the relative filter index for layer k[ and U ∈ (0,1). The set of filters of the layer are ranked in descending order according to their absolute sum. Then, the pruning threshold p finds the number of filters to be pruned (X) for each layer, If is than ∥X[v]∥, where v = 0, the relative filter index U is to zero otherwise to one as The filters index with value zero are less important and will be pruned, while those with value one will be preserved. In the proposed work, the filters whose corresponding index in U were zero, pruned from the model and a new architecture is created, and the weighs are copied from model N to N for the remaining filters. Further, the identified candidate pruning filters were then pruned in a single pass from the layer. For each of the convolutional layers, this process was repeated. When a filter from the layer k was pruned, the activation map associated with that filter was also pruned, resulting in reducing the #channels for the k + 1 layer.

Step II: re-train pruned models

The pruning of convolutional filters from the model leads to performance degradation. Therefore, it is required to re-train the pruned architecture on the original dataset before applying transfer learning. In this regard, before training the resulting pruned architectures on the CT-scan dataset, the resulting architectures are re-trained on the ImageNet dataset with standard data splits. The pre-trained models were used from the PyTorch deep learning framework. After pruning, during fine-tuning the pruned models, standard hyper-parameters were used for training and testing purposes.

Step III: training on the COVID dataset

After pruning and fine-tuning the pre-trained models, the SARS-CoV-2 dataset was utilized for training and testing the pruned models. The process of training pre-trained models on another dataset is referred to as transfer learning. Since we transferred only important weights of the pre-trained models, we called this an important weight-only transfer learning. The models were trained with data splits discussed in Section 4.1. On the SARS-CoV-2 dataset, we conducted various experiments with the unpruned models and the ImageNet re-trained pruned models. The base and re-trained models were trained and tested in the following ways. Training the entire model (both original and the pruned & fine-tuned) on the SARS-CoV-2 dataset. Training the last layer of the model (both pruned & fine-tuned and the original) on the SARS-CoV-2 dataset. Training the entire dense layers on the SARS-CoV-2 dataset (both VGG16 original and the pruned & fine-tuned).

Experimental setup and results

This section includes the detail of the dataset, CNN models, experimental setup, and evaluation metrics used to perform the experiments and measure the performance of the original & the pruned models.

Dataset

Deep learning algorithms require a large amount of labelled data to learn the distinguishing characteristics from images. In our study, we used a publicly available SARS-CoV-2 CT-scan dataset [24] to validate the effectiveness of the selective transfer learning method. The SARS-CoV-2 dataset comprises 2482 images. Out of the 2482 images, 1252 belong to the COVID-19 infected, and 1230 images belong to the COVID-19 non-infected class. The dataset was prepared by collecting images of actual patients in the hospital in Sao Paulo, Brazil. Figure 2 shows some random COVID-19 positive as well as COVID-19 negative images. The dimension of the images in the dataset is 224 × 224. The dataset is available for download from the link 1. As shown in Table 1, the dataset was divided into three parts as 68% training, 17% validation, and 15% testing.
Fig. 2

COVID-19 positive (bottom row) and negative (top row) images from the dataset

Table 1

Normal and infected images from the SARS-CoV-2 CT-scan dataset

ClassTrainingValidationTestingTotal Images
Normal8362081861,230
Infected8512121891,252
Total1,6874203752,482
COVID-19 positive (bottom row) and negative (top row) images from the dataset Normal and infected images from the SARS-CoV-2 CT-scan dataset

CNN models used

The experiments were performed on two CNN classification models, VGG16 and ResNet34. VGG16 [49] is a popular CNN architecture that achieved 92.7% top-5 test accuracy in ILSVRC 2014 challenge on the ImageNet dataset. VGG16 consists of convolutional, dense, and pooling layers, 3 × 3 filters for convolutional and 2 × 2 for max pooling, respectively. The model accepts 224 × 224 input images followed by two convolution layers with 64 filters and a max pooling to reduce the output height and width to 112 × 112 × 64. Further, two convolutional layers with 128 filters are used, followed by a max pooling layer that reduces the activation size to 56 × 56 × 128. Similarly, three convolutional layers with 256 filters are followed by a pooling layer that reduces the output activation to 28 × 28 × 256. Finally, there are two stacks of three convolutional layers with 512 filters, separated by pooling layers. Next, dense layers with 4096 nodes accept the output of the last pooling layer which is 7 × 7 × 512. The dense layer is followed by one more dense layer with 4096 nodes. Finally, the model has the softmax layer with 1000 nodes. ResNet34 [50] is another well-known architecture that performed better than VGG16 in ILSVRC challenge in the year 2015 and also archived first place in the competition. The ResNet34 design is made up of four residual blocks and is based on skip connections. The first block comprises six convolutional layers, each of which has 64, 3 × 3 filters. Eight convolutional layers and 128, 3 × 3 filters make up the second block. 256, 3 × 3 filters are used in the third block, consisting of 12 convolutional layers. The final block consists of six convolutional layers with 512 filters. Finally, it is followed by an average pooling and a softmax layer with 1000 nodes.

Evaluation metrics

The standard evaluation metrics were used to validate the performance of the CNN models. The evaluation metrics used in the experiments were confusion matrix, accuracy, precision, F1-score, recall/sensitivity, and specificity. The different evaluation metrics are defined as: (In the below equations, FN = false negatives, TN = true negatives, TP = true positives, FP = false positives).

Confusion matrix:

The confusion matrix is one of the important evaluation metrics to measure the performance of binary classification problems. It measures the performance by comparing the actual values and the predicted values. When the actual sample is positive/negative and the model classifieds it as a positive/negative, it is known as true positive (TP) and true negative (TN), respectively. When the actual sample is negative (a person does have COVID), and the model predicted it as positive, it is known as a false positive (FP). Finally, if the actual sample is positive (a person has COVID) and the model predicted it as negative, it is known as false negative (FN).

Accuracy:

Accuracy is defined as the total number of the correct classification made by the model to the total number of samples in the dataset. It is the most widely used evaluation metric to evaluate the classification performance and it is given by

Precision:

Precision is the ratio of the correct positives samples identified by the model out of the total number of positive predictions made by the model. The precision value ranges from zero to one; if the model does not have any false positives (FP = 0), it will always have a precision of one.

Sensitivity or recall:

Sensitivity or recall is defined as the measure of what proportion of actual positives are identified correctly. If the model has no false negatives (FN = 0), then the recall will be one, which states that all the actual positives samples were classified correctly.

Specificity:

It is the ratio of the total true negatives identified by the model to the sum of true negatives and false positives. It is given by

F1-score:

The harmonic mean of precision and recall is used to calculate the F1-score. The value of the F1-score also varies between zero and one, where the values close to one are considered best. F1-score is defined as

ROC curve:

The receiver operating characteristic (ROC) curve is another important metric to measure the performance of binary classification problems. ROC curve is plotted between the true positive rate (y-axis) and false-positive rate (x-axis) for varying thresholds between zero and one.

Training VGG16 and ResNet34 on the SARS-CoV-2 dataset

The PyTorch deep learning framework was utilized to implement the experiments. NVIDIA DGX-1 V100 supercomputer was used as computing power. The initial experiments were performed on the VGG16 pruned and original model. The following were the hyper-parameters that were utilized for training, validating, and testing. The models were trained for 200 epochs using a 0.001 learning rate. The training was carried out with a 0.9 momentum stochastic gradient descent (SGD) optimizer. All of the images were normalized before being fed into the model. The images were presented to the model in a mini-batch size of 32. In addition, the images were resized to 256 × 256 and were center cropped at 224 × 224. Further, the experiments were also performed with and without data augmentation on original and pruned models. Random horizontal flip, random vertical flip, and 20-degree random rotation were mainly applied to the training images. Two experiments were carried out on the ResNet 34 model, one to train only the last softmax layer and the other to train the entire model. The VGG16 model, on the other hand, was subjected to three sorts of experiments: training entire layers, dense layer, and the last layer. As given in Table 1, the dataset was split into training, validation, and test in the ration of 68%, 17%, and 15%, respectively. The result of training the original and pruned models on the SARS-CoV-2 CT-scan dataset are summarized in Tables 2, and 3, respectively. A detailed discussion of the result is given in the next section.
Table 2

Different performance measures for the ResNet34 and VGG16 on the test data (original pre-trained models)

Model trainedAugmentationAccuracyPrecisionRecallF1-scoreSpecificityROC-AUC
ResNet34No95.730.94710.96760.95720.94740.9931
ResNet34Yes97.870.99470.96410.97920.99440.9967
VGG16, denseNo90.130.92060.88780.90390.91620.9667
VGG16, denseYes89.870.94710.86470.90400.94050.9797
VGG16, allNo90.400.91530.89640.90580.91210.9660
VGG, allYes96.530.96300.96810.96550.96260.9957
Table 3

Different performance measures for the ResNet34 and VGG16 on the test data (pruned models)

Model trainedAugmentationAccuracyPrecisionRecallF1-scoreSpecificityROC-AUC
ResNet34No94.930.93120.96700.94880.93260.9888
ResNet34Yes95.470.99470.92160.95670.99420.9974
VGG16, denseNo89.330.87300.91160.89190.87630.9669
VGG16, denseYes89.330.88890.89840.89360.88830.9698
VG16, allNo93.070.94180.92230.93190.93960.9744
VG16, allYes92.800.96300.90100.93090.95950.9878
Different performance measures for the ResNet34 and VGG16 on the test data (original pre-trained models) Different performance measures for the ResNet34 and VGG16 on the test data (pruned models)

Results

The results of all the experiments utilizing the SARS-CoV-2 dataset are discussed in this section. The section also contains the detail of the comparative study, complexity analysis, and statistical test. In this study, first, the experiments were performed with the pertained deep learning models VGG16 and ResNet34. Table 2 details the various evaluation metrics calculated after using test data on trained models. Table 2 shows that when the complete model was trained with data augmentation, the VGG16 model achieved higher accuracy. In this case, the model achieved 0.9630 precision, 0.9681 sensitivity, 0.9655 F1-score, 96.53% accuracy, 0.9626 specificity, and 0.9957 AUC. The ResNet34 model achieved higher accuracy when the entire model was trained with data augmentation. In this case, the model achieved 0.9947 precision, 0.9641 sensitivity, 0.9792 F1-score, 97.87% accuracy, 0.9944 specificity, and 0.9967 AUC. Table 3 shows the detail of different evaluation metrics calculated after applying test data on the pruned models. The pruned version of the VGG16 achieved 93.07% accuracy, 0.9223 sensitivity, 0.9319 F1-score, and 0.9396 specificity, and 0.9744 AUC. On the pruned version of the ResNet34 model 95.47% accuracy, 0.9216 sensitivity, 0.9567 F1-score, 0.9942 specificity, and 0.9974 AUC was achieved. In addition, for the ResNet34 and VGG16 models, Figs. 3 and 4 illustrate the confusion matrix, ROC curve, and precision-recall curve, respectively. On the basis of #parameters, #FLOPs, and accuracy, in Table 4 we compared the original & pruned VGG16 and ResNet34 models. It is evident from Table 4 that the pruned VGG16 has 41.66% less weight parameters, and the #FLOPs also reduced by 77.47%. On the other hand, ResNet34 pruned model has 20.64% less weight parameters, and the #FLOPs were also reduced by 41.96%.
Fig. 3

The confusion matrix, precision-recall curve, and ROC curve for the VGG16 original (left) and pruned (right) model

Fig. 4

The confusion matrix, precision-recall curve, and ROC curve for the ResNet34 original (left) and pruned (right) model

Table 4

Pruned and original model comparison, Augmentation = Aug, Million = M, Billion = B

OriginalPruned
ModelAug.Para (M)FLOP (B)Acc.Para (M)FLOP (B)Acc.%Para %FLOP Acc(±)
ResNet34No21.283.6795.7316.892.1394.9320.6441.96-0.80
ResNet34Yes21.283.6797.8716.892.1395.4720.6441.96-2.40
VGG16, denseNo134.2615.4990.1378.333.4989.3341.6677.47-0.80
VGG16, denseYes134.2615.4989.8778.333.4989.3341.6677.47-0.53
VG16, allNo134.2615.4990.4078.333.4993.0741.6677.472.67
VG16, allYes134.2615.4996.5378.333.4992.8041.6677.47-3.73
The confusion matrix, precision-recall curve, and ROC curve for the VGG16 original (left) and pruned (right) model The confusion matrix, precision-recall curve, and ROC curve for the ResNet34 original (left) and pruned (right) model Pruned and original model comparison, Augmentation = Aug, Million = M, Billion = B

Complexity analysis

The complexity of the proposed work was analyzed based on the time taken to classify the single image and the whole test set. For this, the pruned and original models were deployed on GPU & CPU. In addition, the complexity of the VGG16 model was assessed based on the number of parameters, and the FLOPs decreased layer-by-layer. Table 5 shows the inference time of the VGG16 and ResNet34 pruned and original models on CPU & GPU. It should be noted here from Table 5 that the pruned models have significant improvement in the inference time compared to the original models. Further, the test set was applied 50 times to record the inference time and the average was taken. The CPU and GPU inference time of the models for a single image and an entire set is shown in Figs. 5 and 6, respectively. It is also evident from Figs. 5 and 6 that the CPU and GPU time is less for the pruned models. Table 6 compares the complexity of the pruned and original VGG16 models in terms of weight parameters and FLOPs. The 13 layers of the VGG 16 model are divided into five blocks, where each block further follows the pooling layer. While counting the FLOPs, addition and multiplication were considered as a single operation. The pruned model contains fewer FLOPs and weight parameters, resulting in a considerable improvement in model inference performance, as shown in Table 6.
Table 5

Inference time (seconds), #parameters, FLOPs, and filters of the models

Original modelPruned model
MetricVGG16ResNet34VGG16ResNet34
Parameters (M)134.2621.2878.3316.89
Parameter reduction (%)0041.6620.64
FLOPs (B)15.493.673.492.13
FLOPs reduction (%)0077.4741.96
Convolutional filters4224851220737362
Convolutional filter reduction (%)0050.9213.51
GPU inference-time, single image (s)0.0052190.0046370.0041540.004250
GPU inference-time, test set (s)1.9570301.7388011.5577871.593882
CPU inference-time, single image (s)0.1581530.0588110.0562850.040023
CPU inference-time, test set (s)59.30741122.05404921.10668815.008648
Fig. 5

Inference time (CPU and GPU) for single image

Fig. 6

Inference time (CPU and GPU) on the entire test set for all the models

Table 6

VGG16 model complexity analysis

Convolutional block and layersBefore pruningAfter pruningReduction
BlockFilter, LayerParametersFLOPsParametersFLOPs%Para red%FLOP red
Conv block 164, 2387201952448512837442458931278.3778.25
Conv block 2128, 222144027825602564082251400294481.5781.53
Conv block 3256, 314753284629839872348173109341344076.4076.38
Conv block 4512, 3589977646266286081306313102472014477.8677.85
Conv block 5512, 370794241388269568183756936051318874.0474.03
FC14,096 (neuron)102764544102764544580075525800755243.5543.55
FC24,096 (neuron)167813121678131216781312167813120.000.00
FC32 (neuron)81948194819481940.000.00
Total134.26M15.49B78.33M3.49B41.6677.47
Inference time (seconds), #parameters, FLOPs, and filters of the models Inference time (CPU and GPU) for single image Inference time (CPU and GPU) on the entire test set for all the models VGG16 model complexity analysis

Paired statistical test

The inference-time was used as a dependent variable in a paired statistical t-test to validate the performance of the VGG16 and ResNet34 original and pruned models. The mean inference-time difference between the original and pruned models before and after pruning was compared using a paired sample t-test. For this, the hypotheses were established (null and alternate). The null hypothesis was that the mean inferencetime of the original and pruned models was the same (H0 : μ = μ). The mean inference-time of the two models was not the same under the alternate hypothesis H1 : μ≠μ. The inference-time of the original and pruned models was determined by evaluating them on the test set with various test set splits. The test set was split into one, two, three, four, five, and ten equal parts. On each test split, the original and pruned models were evaluated, and the model inference time was recorded. α was set to 0.05 as the significance level. The VGG16 model had a p-value of less than 0.001 and a t-value of 4.504. There is sufficient evidence to reject the null hypothesis because the p-value is smaller than alpha. The ResNet34 model has a t-value of 2.735 and a p-value less than 0.001. With alpha = 0.05, the ResNet34 model has a lower p-value, indicating that there is enough evidence to reject the null hypotheses.

Comparison with other methods.

The proposed method was compared with existing state-of-the-art methods [24, 38, 42–47, 51] and the results are presented in Table 7. The authors of the research [42] proposed a stacked ensemble of heterogeneously pre-trained CNN models. The ensemble model was created by the combination of VGG19, ResNet101, Densenet169, and WideResNet50-2. No data augmentation was applied on the SARS-CoV-2 CT-scan dataset. As a result, the model failed to achieve good classification accuracy. The authors achieved 91.5% accuracy, 0.915 sensitivity, and 0.915 F-score. In addition, [43] also proposed a stacked ensemble method. The authors claimed that the stacked ensemble method achieved higher recall and accuracy. The authors achieved 94% accuracy, 0.98 sensitivity, and 0.94 F-score. The methods proposed by [42] and [43] did not make any optimization and were less suitable for the point-of-care devices. The authors of the research [44] concluded that the combination of transfer learning with segmentation methods such as U-Net improves the classification performance. Transfer learning with U-Net architecture outperformed other state-of-the-art transfer learning-based CNN methods. Without segmentation, the authors achieved 89.31% accuracy, 0.8240 sensitivity, 0.8860 F-score, and 0.9634 specificity. On the other hand, the authors achieved 89.92% accuracy, 0.8680 sensitivity, 0.8967 F-score, and 0.9309 specificity with the segmentation scheme. However, both the schemes failed to achieve competitive performance.
Table 7

Comparison of the proposed important weights only approach with other methods on the SARS-CoV-2 dataset

MethodAccuracySensitivityF1-scoreSpecificityFLOP(%) Para.(%)
[42] Varied threshold91.50.9150.915-
[44] Without segmentation89.310.82400.88600.9634
[44] With segmentation89.920.86800.89670.9309
[38] VGG1695.450.95230.95490.9567
[38] DenseNet96.250.96290.96290.9621
[24] GoogleNet91.730.93500.9182-
[24] ResNet94.960.97150.9503-
[24] AdaBoost95.160.96710.9514-
[45] Contrastive Learning90.830.85890.9087-
[46] DenseNet-121920.950.89-
[43] Stacked Ensemble940.980.94-
[47] DNN95.61---
Ours, VGG16 pruned93.070.92230.93190.939677.4741.66
Ours, ResNet34 pruned95.470.92160.95670.994241.9620.64

The bold text shows our research result

Comparison of the proposed important weights only approach with other methods on the SARS-CoV-2 dataset The bold text shows our research result Moreover, [38] also proposed a transfer learning-based approach for SARS-CoV-2 CT-scan classification. Particularly, the authors worked with the VGG16 and DenseNet201 models. The accuracy, sensitivity, F-score, and specificity for the VGG16 model were 95.45%, 0.9523, 0.9549, and 0.9567, respectively. The accuracy, sensitivity, F-score, and specificity achieved by the DenseNet201 model were 96.25%, 0.9629, 0.9629, and 0.9621, respectively. The results of the research [38] were improved compared to the other methods. However, the storage, energy, and computational requirement of the pre-trained models were high. Hasan et al [46] worked with the DenseNet121 convolutional architecture and obtained 0.95 sensitivity, 0.89 F-score, and 92% accuracy. The authors of the research [45] proposed a redesigned COVID-Net for improved performance along with the objective of contrastive learning for cross-site learning. The authors achieved 90.83% accuracy, 0.8589 sensitivity, and 0.9087 F1-score. Angelov and Almeida Soares [24] worked with the GoogleNet, ResNet, and AdaBoost methods. AdaBoost achieved higher accuracy compared to GoogleNet and ResNet. However, the GoogleNet model showed higher sensitivity. The authors of the research [47] achieved 95.61% accuracy with the VGG16 pre-trained model. The authors also implemented grad-CAM-based color visualization to interpret the predictions. It should be noted from Table 7 that none of the existing methods takes into consideration the constraints of the point-of-care devices. The last two-column of Table 7 shows that the existing methods make no reduction in FLOPs and learnable weight parameter. Further, it can be seen from Table 7 that the proposed method significantly reduces the inference-time needs of the models and also achieved competitive performance. Furthermore, our pruned ResNet-34 model achieved 95.47% classification accuracy, 0.9216 sensitivity, 0.9567 F-score, and 0.9942 specificity. On the other hand, the VGG16 pruned model achieved 93.07% accuracy, 0.9223 sensitivity, 0.9319 F1-score, and 0.9396 specificity. The pruned VGG16 and ResNet34 models reveal that the pre-trained models are over-parameterized and that removing low-importance parameters enhances the model’s performance for point-of-care devices.

Discussion

Early detection and treatment of infectious diseases play an important role in medical diagnosis. Many researchers have recently recommended radiological imaging-based approaches, given the present constraints of reverse transcription-polymerase chain reaction (RT-PCR)-based testing for diagnosing COVID19. Furthermore, with the development of AI-based technology, significant progress in automated medical diagnosis has been made. However, during our research, it was found that high-performance techniques such as deep learning methods need high computational resources. For widespread benefits, the trained deep learning model must be deployed in point-of-care devices. However, the point-of-care devices have limited resources to execute the large, trained models. Motivated by the deep learning models’ ability to generate the diagnosis results accurately, timely, and the limitations of the point-of-care devices, a selective transfer learning approach was suggested in this study to classify CT-scan images as COVID-19 positive or negative. The result of the study indicates that the selective transfer learning approach effectively makes the deep learning models inference efficient for point-of-care devices in the medical domain for early diagnosis. It will help speed up the diagnosis process and significantly reduce the dependability on the skilled technicians, laboratories, and make the automated diagnosis more affordable in underprivileged areas. The comparative analysis found that the proposed method performed superior to existing methods in classifying chest CT-scan images. Moreover, none of the existing methods minimizes the trained models’ run-time resource requirements for point-of-care devices. The VGG16 pruned model achieved 93.07% accuracy, while the Resnet-34 pruned model achieved 95.47% accuracy. Another noteworthy finding from this study is that the VGG16 model has 41.66 percent fewer parameters and 77.47 percent less floating-point operations than the standard model. Similarly, the ResNet-34 model has 20.64% fewer parameters and 41.96% fewer FLOPs than the standard model. Furthermore, this research finds that pre-trained CNN architectures are over-parameterized, and that filter pruning improves inference performance. The proposed method has advantages over other existing filter pruning methods. Unlike, other methods in which to identify the pruning candidate filters, the author remove the filters one by one and evaluate model loss after each pruned filter. Removing the filters one by one is a time consuming and computational intensive task. In contrast, in the proposed method, one shot pruning is applied to find all the candidate pruning filters. Moreover, the method is also different from those in which the convolutional filters are sparsified by setting some of the weights to zero. Such methods require specialized hardware and software to process the resulting sparse model [48]. On the contrary, the proposed method completely removes the unimportant filters and their corresponding feature maps. Unlike [21], the proposed method doesn’t required training the model from scratch to find the less important filters. In contrast, the proposed method can be applied to prune any pretrained model. Further, the current work focuses only on COVID-19 disease; however, the proposed important weights-only learning approach can be used for other applications in point-of-care devices. Some of the applications include detecting skin lesions, Pneumonia, and Tuberculosis, to name a few. In addition to various advantages of using CT-scan-based automatic image diagnosis for COVID-19 detection, such models can help radiologists effectively detect the virus. Also, these models not only show predictions or classifications over the CT-scan but can also be used to monitor the outcome of the treatment [52, 53].

Conclusion

In this paper, we proposed an important weights-only transfer learning method to classify the CT-scan images as COVID-19 infected or not. The proposed method has superior performance, and the trained models have reduced inference-time resource requirements. Another important conclusion of the study was that deploying the pruned models on point-of-care devices is advantageous. The computational requirement of the VGG16 and ResNet34 models was reduced by 77.47% and 41.66% in terms of the number of floating-point operations to be performed during inference. The pruned and original architectures were tested on GPU and CPU to see the practical speed-up and a significant speed-up is measured in the performance of the pruned architectures. Future studies can be aimed at developing inference-efficient models for other kinds of diseases for point-of-care devices. Moreover, the new structured filter pruning techniques can be developed to better find the filters that produce weaker activation. Further, the proposed pruning approach can be combined with other acceleration techniques to improve inference performance on point-of-care devices.
  34 in total

1.  A transfer learning with structured filter pruning approach for improved breast cancer classification on point-of-care devices.

Authors:  Tejalal Choudhary; Vipul Mishra; Anurag Goswami; Jagannathan Sarangapani
Journal:  Comput Biol Med       Date:  2021-04-30       Impact factor: 4.589

2.  Shallowing Deep Networks: Layer-Wise Pruning Based on Feature Representations.

Authors:  Shi Chen; Qi Zhao
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  2018-10-08       Impact factor: 6.226

3.  A novel comparative study for detection of Covid-19 on CT lung images using texture analysis, machine learning, and deep learning methods.

Authors:  Huseyin Yasar; Murat Ceylan
Journal:  Multimed Tools Appl       Date:  2020-10-06       Impact factor: 2.757

Review 4.  A Review of Coronavirus Disease-2019 (COVID-19).

Authors:  Tanu Singhal
Journal:  Indian J Pediatr       Date:  2020-03-13       Impact factor: 1.967

5.  Point-of-care cervical cancer screening using deep learning-based microholography.

Authors:  Divya Pathania; Christian Landeros; Lucas Rohrer; Victoria D'Agostino; Seonki Hong; Ismail Degani; Maria Avila-Wallace; Misha Pivovarov; Thomas Randall; Ralph Weissleder; Hakho Lee; Hyungsoon Im; Cesar M Castro
Journal:  Theranostics       Date:  2019-11-26       Impact factor: 11.556

Review 6.  Deep Learning in Medical Imaging: General Overview.

Authors:  June-Goo Lee; Sanghoon Jun; Young-Won Cho; Hyunna Lee; Guk Bae Kim; Joon Beom Seo; Namkug Kim
Journal:  Korean J Radiol       Date:  2017-05-19       Impact factor: 3.500

7.  Coronavirus disease (COVID-19) detection in Chest X-Ray images using majority voting based classifier ensemble.

Authors:  Tej Bahadur Chandra; Kesari Verma; Bikesh Kumar Singh; Deepak Jain; Satyabhuwan Singh Netam
Journal:  Expert Syst Appl       Date:  2020-08-26       Impact factor: 6.954

8.  A stacked ensemble for the detection of COVID-19 with high recall and accuracy.

Authors:  Ebenezer Jangam; Chandra Sekhara Rao Annavarapu
Journal:  Comput Biol Med       Date:  2021-06-30       Impact factor: 4.589

9.  DenseNet Convolutional Neural Networks Application for Predicting COVID-19 Using CT Image.

Authors:  Najmul Hasan; Yukun Bao; Ashadullah Shawon; Yanmei Huang
Journal:  SN Comput Sci       Date:  2021-07-23
View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.