| Literature DB >> 33948047 |
Fouzia Altaf1, Syed M S Islam1, Naeem Khalid Janjua1.
Abstract
Deep learning has provided numerous breakthroughs in natural imaging tasks. However, its successful application to medical images is severely handicapped with the limited amount of annotated training data. Transfer learning is commonly adopted for the medical imaging tasks. However, a large covariant shift between the source domain of natural images and target domain of medical images results in poor transfer learning. Moreover, scarcity of annotated data for the medical imaging tasks causes further problems for effective transfer learning. To address these problems, we develop an augmented ensemble transfer learning technique that leads to significant performance gain over the conventional transfer learning. Our technique uses an ensemble of deep learning models, where the architecture of each network is modified with extra layers to account for dimensionality change between the images of source and target data domains. Moreover, the model is hierarchically tuned to the target domain with augmented training data. Along with the network ensemble, we also utilize an ensemble of dictionaries that are based on features extracted from the augmented models. The dictionary ensemble provides an additional performance boost to our method. We first establish the effectiveness of our technique with the challenging ChestXray-14 radiography data set. Our experimental results show more than 50% reduction in the error rate with our method as compared to the baseline transfer learning technique. We then apply our technique to a recent COVID-19 data set for binary and multi-class classification tasks. Our technique achieves 99.49% accuracy for the binary classification, and 99.24% for multi-class classification.Entities:
Keywords: COVID-19; Chest radiography; Computer-aided diagnosis; Deep learning; Dictionary learning; Thoracic disease classification; Transfer learning
Year: 2021 PMID: 33948047 PMCID: PMC8083924 DOI: 10.1007/s00521-021-06044-0
Source DB: PubMed Journal: Neural Comput Appl ISSN: 0941-0643 Impact factor: 5.606
Fig. 1Schematics of the proposed technique: A set of natural (colour) image deep learning models are augmented with additional input and modified output layers. The augmented models are hierarchically fine-tuned with limited (greyscale) images of chest X-rays. Features extracted from augmented models are also used as dictionaries to compute dense and sparse representations of unseen samples. Outputs of model ensemble and dictionary codes are combined to predict output labels
Network adaption for transfer learning at input stage
| Network | Original | Input | Modified | Input | Activations |
|---|---|---|---|---|---|
| DenseNet201 | Input_1 | 224 | Input_grey | 448 | 224 |
| Conv 7, 3, [2,2] | |||||
| Batch-N, ReLU | |||||
| ResNet50 | Input_1 | 224 | Input_grey | 448 | 224 |
| Conv 7, 3, [2,2] | |||||
| Batch-N, ReLU | |||||
| Inception-V3 | Input_1 | 299 | Input_grey | 598 | 299 |
| Conv 3, 3, [2,2] | |||||
| Batch-N, ReLU | |||||
| VGG-16 | Input | 224 | Input_grey | 448 | 224 |
| Conv 3, 3, [2,2] | |||||
| Batch-N, ReLU |
‘Original’ names of the altered layers are given along the ‘Input’ dimensions expected. For the ‘Modified’ network, conv K, F, [S,S] indicates a convolutional kernel with kernel size , with F number of filters and a stride of [S, S]. The activations of convolutional layer are batch-normalized [22], indicated by ‘Batch-N’, followed by ReLU activations [37]. The output ‘Activations’ of the modified layer are given in the last column. At the output stage, the fully connected layers are modified to have N neurons instead of 1000, where N is the number of output classes considered
Fig. 2Samples of chest X-ray images from COVID-19 data set
Results summary on Chest X-ray14 data set: conventional transfer learning (TL) with DenseNet201 is the ‘Baseline’
| Models | Spec. | Sens. | F1 | Acc. | ERR | ||
|---|---|---|---|---|---|---|---|
| Baseline (TL) | 89 | 80 | – | 83.33 | – | 16.76 | – |
| DenseNet201 (Den.) | 94 | 55 | 46 | 89.65 | 37.91 | 48.27 | 188.0 |
| Den+VGG | 94 | 55 | 47 | 90.00 | 3.38 | 50.03 | 3.65 |
| Den+VGG+Res | 94 | 58 | 51 | 90.63 | 6.29 | 53.17 | 6.27 |
| Den+VGG+Res+IV3 | 95 | 60 | 53 | 91.03 | 4.27 | 55.17 | 3.76 |
| Proposed | 95 | 60 | 53 | 91.38 | 3.90 | 56.90 | 3.13 |
‘Dense’ denotes DenseNet201 augmented with our technique. Similarly, ‘VGG’, ‘Res’ and ‘IV3’ are augmented versions of VGG-16, ResNet50 and Inception-V3 using our method. ‘Full ensemble’ is the final technique. The Error Reduction Rate (ERR) is computed using Accuracy (Acc.) of two consecutive rows. The Gain is computed with two consecutive rows of Acc
Computational times for the used models: step 1 is the training of input and output layers with frozen inner layers for 5 epochs with learning rate 0.001
| Models | Training time | Testing time (milliseconds) | ||
|---|---|---|---|---|
| Step1 | Step2 | Step3 | ||
| DenseNet201 | 3 h 20 m 5 s | 6 h 15 m 20 s | 7 h 29 m 5 s | 6 |
| VGG | 0 h 55 m 15 s | 1 h 51 m 25 s | 2 h 5 m 20 s | 8 |
| ResNet | 0 h 47 m 20 s | 1 h 26 m 15 s | 1 h 35 m 10 s | 6 |
| InceptionV3 | 1 h 14 m 25 s | 2 h 23 m 15 s | 2 h 31 m 15 s | 6 |
Step 2 is the training on augmented data with the model resulting from step 1, using learning rate 0.0001 for five epochs. Step 3 is the training on augmented data with learning rate 0.00001 and fine tuning the complete model weights for 5 epochs. We also include the dictionary computation time in this stage. Test time is for a single image, including sparse coding stage
Results of individual classes on Chest X-ray14 data set: for A/B, A is the value computed for the proposed technique, B is the value of ‘Baseline’ that uses transfer learning with DenseNet201
| Class | Spec. | Sens. | F1-Score | Acc. |
|---|---|---|---|---|
| Atelectasis | 0.96/0.94 | 0.61/0.08 | 0.56/0.09 | 93.46/88.73 |
| Cardiomegaly | 0.97/0.75 | 0.86/0.22 | 0.72/0.08 | 96.43/72.27 |
| Effusion | 0.97/0.86 | 0.59/0.06 | 0.66/0.05 | 93.36/ 76.60 |
| Infiltration | 0.93/0.86 | 0.36/0.09 | 0.42/0.09 | 85.13/74.43 |
| Mass | 0.97/0.63 | 0.67/0.19 | 0.55/0.03 | 96.56/62.40 |
| Nodule | 0.97/0.96 | 0.41/0.06 | 0.33/0.04 | 96.63/94.83 |
| Pneumothorax | 0.94/0.97 | 0.7/0.02 | 0.66/0.04 | 92.10/86.90 |
| Consolidation | 0.88/0.99 | 0.73/0.00 | 0.47/– | 87.30/92.2 |
| Pleural thickening | 0.93/0.97 | 0.46/0.01 | 0.25/0.01 | 92.46/94.80 |
| No finding | 0.95/0.99 | 0.52/0.00 | 0.64/– | 80.63/65.13 |
Results summary on COVID-19 data set: dense denotes DenseNet201 augmented with our technique. Similarly, ‘VGG’, ‘Res’ and ‘IV3’ are augmented versions of VGG-16, ResNet50 and Inception-V3 using our method
| Model | Spec. | Sens. | F1 | Acc. without Dict. | Acc. with Dict. | ||
|---|---|---|---|---|---|---|---|
| Dense | 98.11 | 96.21 | 96.20 | 97.47 | 97.47 | 96.21 | 96.21 |
| VGG | 98.86 | 97.73 | 97.74 | 96.97 | 98.48 | 95.45 | 97.73 |
| Res | 97.73 | 95.45 | 95.48 | 94.44 | 96.97 | 91.67 | 95.45 |
| IV3 | 96.59 | 93.18 | 93.08 | 95.45 | 95.45 | 93.18 | 93.18 |
| Ensemble | 99.24 | 98.48 | 98.49 | 98.99 | 98.48 |
‘Ensemble’ is the ensemble of the four models. ‘Acc.’ denotes the accuracy for binary classification and ’Acc.’ is the accuracy for multiclass classification
Results of individual classes on COVID-19 data set
| Class | Spec. | Sens. | F1 | Acc. |
|---|---|---|---|---|
| Covid-19 | 100 | 100 | 100 | 100 |
| Pneumonia | 98.86 | 100 | 98.88 | 99.24 |
| Normal | 100 | 97.73 | 98.88 | 99.24 |
Comparison of the proposed technique with other deep learning techniques for COVID-19 diagnostic using chest X-ray images
| Study | No. of cases | Architecture | Data set | Accuracy |
|---|---|---|---|---|
| Ioannis et al. [ | 224 COVID-19, 700 Pneumonia, 504 Healthy | VGG-19 | [ | 98.75 |
| Wang et al. [ | 53 COVID(+), 5526 COVID(−), 8066 Healthy | COVID-Net | [ | 92.4 |
| Ozturk et al. [ | 125 COVID-19, 500 Pneumonia, 500 No finding | Dark COVID-Net | [ | 98 |
| Asif et al. [ | 1300 images of COVID-19, normal, pneumonia | CoroNet | [ | 95 |
| Tougaccar et al. [ | 295 COVID-19, 98 Pneumonia, 65 No findings | MobileNetV2 | [ | 99.27 |
| Narin et al. [ | 50 COVID-19, 50 No findings | ResNet-50 | [ | 98 |
| Hemaden et al. [ | 25 COVID-19, 25 No findings | VGG-19, DenseNet-121 | [ | 90 |
| Sethy et al. [ | 25 COVID-19, 25 No findings | ResNet-50 | [ | 95.38 |
| Toraman et al. [ | 1050 COVID, 1050 No finding | CapsNet | [ | 97.24 |
| Panwar et al. [ | 192 COVID-19, 145 No findings | nCOVnet | [ | 97.62 |
| Ucar et al. [ | 76 COVID-19, 1583 normal, 4290 pneumonia | Bayes-SqeezeNet | [ | 98.3 |
| 219 COVID-19, 219 Viral Pneumonia, 219 Normal | Augmented | [ |