| Literature DB >> 33362346 |
Deepika Selvaraj1, Arunachalam Venkatesan1, Vijayalakshmi G V Mahesh2, Alex Noel Joseph Raj3.
Abstract
The novel coronavirus disease (SARS-CoV-2 or COVID-19) is spreading across the world and is affecting public health and the world economy. Artificial Intelligence (AI) can play a key role in enhancing COVID-19 detection. However, lung infection by COVID-19 is not quantifiable due to a lack of studies and the difficulty involved in the collection of large datasets. Segmentation is a preferred technique to quantify and contour the COVID-19 region on the lungs using computed tomography (CT) scan images. To address the dataset problem, we propose a deep neural network (DNN) model trained on a limited dataset where features are selected using a region-specific approach. Specifically, we apply the Zernike moment (ZM) and gray level co-occurrence matrix (GLCM) to extract the unique shape and texture features. The feature vectors computed from these techniques enable segmentation that illustrates the severity of the COVID-19 infection. The proposed algorithm was compared with other existing state-of-the-art deep neural networks using the Radiopedia and COVID-19 CT Segmentation datasets presented specificity, sensitivity, sensitivity, mean absolute error (MAE), enhance-alignment measure (EMφ), and structure measure (S m) of 0.942, 0.701, 0.082, 0.867, and 0.783, respectively. The metrics demonstrate the performance of the model in quantifying the COVID-19 infection with limited datasets.Entities:
Keywords: Zernike moment; artificial intelligence; computed tomography image; deep neural network; feature extraction; limited training points; segmentation
Year: 2020 PMID: 33362346 PMCID: PMC7753711 DOI: 10.1002/ima.22525
Source DB: PubMed Journal: Int J Imaging Syst Technol ISSN: 0899-9457 Impact factor: 2.177
Literature summary of related techniques for COVID‐19 screening
| Reference | Technique | Dataset | Findings |
|---|---|---|---|
|
| Aggregated residual transformations to learn a robust and expressive feature representation. | 110 COVID‐19 CT image of size 512 × 512 are collected from 60 patients. |
The model gives 0.95 precision and 0.89 accuracy. 12.7% of accuracy and 14.5% of precision are improvement with U‐Net model. |
|
| Contrastive learning method. | MedSeg dataset:110 CT COVID‐19 images. COVID‐19 CT dataset: 349 CT COVID‐19 images 397 Non‐COVID images. | ResNet‐50 performance gives accuracy of 0.868, recall of 0.872, AUC of 0.931. |
|
| ResNet50, InceptionV3 and InceptionResNetV2. | Chest X‐ray 100 images: 50 non‐infected image and 50 infected images. All images are 224 × 224 pixel size. | ResNet50 model provides good classification performance with 98% accuracy than other two proposed models (97% accuracy for InceptionV3 and 87% accuracy for Inception‐ResNetV2). |
|
| An infection Size Aware Random Forest method (iSARF). | COVID‐19 images are taken from 1658 patients. | iSARF yielded sensitivity of 0.907, specificity of 0.833, and accuracy of 0.879. |
|
| A patch‐based convolutional neural network approach. | COVID‐19 dataset: 180 infected image of size 224 × 224. |
Global approach gives 70.7% of accuracy, 60.6% of precision, 60.1% of recall, 59% of F1‐score. Local approach gives 88.9% of accuracy, 83.4% of precision, 85.9% of recall, 84.4% of F1‐score. |
|
| COVID‐19 Lung Infection Segmentation Deep Network (Inf‐Net) with a parallel partial decoder method is to aggregate the high‐level features. |
Dataset consists of 100 axial CT images from different COVID‐19 patients. 45 CT images randomly selected as training, 5 CT images for validation, and the remaining 50 images for testing. | Inf‐Net achieves 0.692 of sensitivity, 0.943 of specificity, 0.725 of sensitivity, 0.960 of specificity. |
|
| Semi‐supervised shallow neural network model using Parallel self‐supervised neural network model (PQIS‐Net). |
Variable size 2482 lung CT scans: 1252 COVID‐19 infected CT image and 1230 Non‐infected COVID image. 20 labelled COVID‐19 CT image of size 512 × 512 which includes infection region masks, lung masks (left and right) and lung‐infection pair masks. | Accuracy, precision, f1 score, and AUC are 0.931, 0.890, 0.826 and 0.982, respectively. |
|
|
Truncated VGG‐19 with fine‐tuning method. Classification is done with four different classifier method such as deep convolutional neural network (CNN), Bagging Ensemble with support vector machine (SVM), Extreme Learning Machine (ELM) and Online sequential ELM. | CT Scan Dataset: 344 COVID‐19 images, 358 Non‐COVID‐19 images. | Best performing classifier Bagging Ensemble with SVM achieves 95.7% accuracy, 95.8% precision, 0.958 AUC and 95.3% F1‐score. |
|
| Transfer Learning method with argumentation: ResNet18, ResNet 50, ResNet101, SqueezeNet. | COVID‐19:349 CT scan images and Non‐COVID‐19 CT: 397 CT scan images. |
ResNet18 achieves the best performance model. Training accuracy is 99.82%, validation accuracy is 97.32% and testing accuracy is 99.4%. |
|
| Social‐group‐optimization with Kapur's entropy thresholding and then followed by k‐means clustering and morphology‐based segmentation. | COVID‐19 dataset consists of 400 numbers of greyscale lung CT image (200 Non‐infected image and 200 infected images). | Morphology‐based segmentation and KNN gives more than 91% and 87% accuracy. |
FIGURE 1Block diagram for the flow of binary classification for computed tomography image
FIGURE 2A, Original computed tomography image; B, Image after enhancement
CII values for the pre‐processed COVID‐19 CT images
| Number of images | Contrast for the original CT image | Contrast for the pre‐processed image | CII for the pre‐processed image |
|---|---|---|---|
| 1 | 0.5984 | 0.8166 | 1.3646 |
| 2 | 0.4264 | 0.7348 | 1.7234 |
| 3 | 0.1650 | 0.3279 | 1.9872 |
| 4 | 0.4808 | 0.6698 | 1.3931 |
| 5 | 0.4846 | 0.9497 | 1.9599 |
Note: Default CII value for the original image is 1.000.
FIGURE 3Zernike moment features with different combination of n and m [Color figure can be viewed at wileyonlinelibrary.com]
FIGURE 4Experimental analysis of low‐level and high‐level feature from Zernike moment (total 36 Zernike feature, red – infected point, green – background point) [Color figure can be viewed at wileyonlinelibrary.com]
FIGURE 5Plots for Zernike moment features from specific region of the training image (Red points represent background and Green points represents infected pixels) [Color figure can be viewed at wileyonlinelibrary.com]
FIGURE 6Plots for GLCM features from specific region of the training image (Red points represents background and Green points represents infected pixels) [Color figure can be viewed at wileyonlinelibrary.com]
FIGURE 7A, Specific region of background (Bf) and infection features (If) are separated out for training. B, Histogram plot for Bf (blue colour) and If (brown colour) [Color figure can be viewed at wileyonlinelibrary.com]
Evaluation of training points metrics for different classifiers
| Different classifiers | AUC | Sensitivity | Specificity | Accuracy (%) |
|---|---|---|---|---|
|
| 0.8723 | 0.5549 | 0.8988 | 88.6 |
|
| 0.9107 | 0.4025 | 0.9735 | 83.3 |
|
| 0.8187 | 0.5211 | 0.8950 | 88.2 |
|
| 0.94 | 0.756 | 0.9593 | 93 |
|
|
|
|
|
|
Note: The significance of the bold numbers are highlighted results of the proposed algorithm.
Confusion matrix of training points for DNN classifier
|
| ||||
|---|---|---|---|---|
| Infected pixels | Background pixels | Accuracy | ||
| Training output | Infected pixels | 46 469 (7.9%) | 12 684 (2.2%) | 93.8% |
| Background pixels | 23 540 (4.0%) | 506 687 (85.9%) | ||
FIGURE 8ROC for different classifiers [Color figure can be viewed at wileyonlinelibrary.com]
FIGURE 9Deep neural network (DNN) architecture
FIGURE 10Test dataset plot for morphological structuring element vs average accuracy [Color figure can be viewed at wileyonlinelibrary.com]
FIGURE 11Segmentation results for the COVID‐19 dataset. A, Column shows computed tomography axial view input image; B, Column shows ground truth image; C, Column shows the output from deep neural network model and D, Column shows overlay the GT contour on output image for visual understanding [Color figure can be viewed at wileyonlinelibrary.com]
Performance evaluation for the test dataset (36 ZM features)
| Number of test images | Accuracy (A) = | IOU score | Mean IOU = | Sensitivity or Recall (R) = | Specificity = | F1 score = | Precision (P) = | |
|---|---|---|---|---|---|---|---|---|
| Class 0 ( | Class 1 ( | |||||||
| 1 | 83.5 | 0.393 | 0.815 | 0.604 | 0.414 | 0.98 | 0.564 | 0.884 |
| 2 | 85.6 | 0.337 | 0.843 | 0.59 | 0.381 | 0.969 | 0.505 | 0.747 |
| 3 | 88.8 | 0.459 | 0.936 | 0.697 | 0.490 | 0.992 | 0.629 | 0.879 |
| 4 | 91.5 | 0.166 | 0.928 | 0.547 | 0.177 | 0.994 | 0.285 | 0.736 |
| 5 | 89.3 | 0.333 | 0.925 | 0.629 | 0.371 | 0.987 | 0.5 | 0.765 |
| 6 | 82.2 | 0.382 | 0.881 | 0.631 | 0.453 | 0.967 | 0.552 | 0.708 |
| 7 | 97.7 | 0.058 | 0.895 | 0.476 | 0.060 | 0.99 | 0.110 | 0.583 |
| 8 | 86.3 | 0.086 | 0.872 | 0.479 | 0.101 | 0.977 | 0.16 | 0.375 |
| 9 | 83 | 0.66 | 0.922 | 0.791 | 0.711 | 0.984 | 0.600 | 0.915 |
| 10 | 83.3 | 0.238 | 0.824 | 0.531 | 0.285 | 0.955 | 0.385 | 0.590 |
Confusion matrix for test image 1
| GT | ||||
|---|---|---|---|---|
| Infected pixels | Background pixels | Accuracy | ||
| Testing output | Infected pixels | 14 025 (21.4%) | 2228 (3.4%) | 92.30% |
| Background pixels | 2818 (4.3%) | 46 465 (70.9%) | ||
Performance evaluation for the random 10 test image from COVID‐19 dataset (36 ZM and 2 GLCM features)
| Number of test images | Accuracy (A) = | IOU score | Mean IOU = | Sensitivity or Recall (R) = | Specificity = | F1 score = | Precision (P) = | Dice index | MAE | EMφ |
| |
|---|---|---|---|---|---|---|---|---|---|---|---|---|
| Class 0 ( | Class 1 ( | |||||||||||
| 1 | 0.9230 | 0.90 | 0.73 | 0.815 | 0.832 | 0.95 | 0.847 | 0.86 | 0.8028 | 0.1003 | 0.9033 | 0.8363 |
| 2 | 0.9237 | 0.907 | 0.68 | 0.793 | 0.876 | 0.93 | 0.815 | 0.76 | 0.7369 | 0.0945 | 0.9277 | 0.8431 |
| 3 | 0.9610 | 0.95 | 0.66 | 0.805 | 0.723 | 0.98 | 0.795 | 0.88 | 0.8121 | 0.1025 | 0.8512 | 0.8017 |
| 4 | 0.9447 | 0.94 | 0.54 | 0.74 | 0.835 | 0.95 | 0.705 | 0.61 | 0.7923 | 0.0678 | 0.7778 | 0.8411 |
| 5 | 0.9553 | 0.95 | 0.56 | 0.75 | 0.600 | 0.99 | 0.720 | 0.906 | 0.6479 | 0.0863 | 0.8122 | 0.7501 |
| 6 | 0.9579 | 0.95 | 0.74 | 0.845 | 0.882 | 0.96 | 0.851 | 0.82 | 0.7488 | 0.0703 | 0.9366 | 0.8416 |
| 7 | 0.9788 | 0.97 | 0.3 | 0.62 | 0.330 | 0.98 | 0.432 | 0.66 | 0.673 | 0.0394 | 0.9646 | 0.6475 |
| 8 | 0.9394 | 0.938 | 0.31 | 0.624 | 0.346 | 0.98 | 0.473 | 0.81 | 0.6366 | 0.0808 | 0.7627 | 0.6103 |
| 9 | 0.8771 | 0.848 | 0.61 | 0.729 | 0.828 | 0.89 | 0.759 | 0.70 | 0.6665 | 0.1624 | 0.8422 | 0.7391 |
| 10 | 0.9248 | 0.910 | 0.67 | 0.79 | 0.852 | 0.94 | 0.704 | 0.76 | 0.6945 | 0.1128 | 0.8962 | 0.8212 |
Comparative performance evaluation of other existing methods
| Existing reference | Methods | Sensitivity | Specificity | Dice | MAE | EMφ |
|
|---|---|---|---|---|---|---|---|
| Ronneberger et al | U‐Net | 0.534 | 0.858 | 0.439 | 0.186 | 0.625 | 0.622 |
| Oktay et al | Attention‐UNet | 0.637 | 0.744 | 0.583 | 0.112 | 0.739 | 0.744 |
| Schlemper et al | Gated‐UNet | 0.658 | 0.725 | 0.623 | 0.102 | 0.814 | 0.725 |
| Li et al | Dense‐UNet | 0.594 | 0.655 | 0.515 | 0.184 | 0.662 | 0.655 |
| Zhou et al | U‐Net++ | 0.672 | 0.722 | 0.581 | 0.120 | 0.720 | 0.722 |
| Deng‐Ping Fan et al | Inf‐Net | 0.692 | 0.781 | 0.682 | 0.082 | 0.838 | 0.781 |
| Deng‐Ping Fan et al | Semi‐Inf‐Net | 0.725 | 0.800 | 0.739 | 0.064 | 0.894 | 0.800 |
| Our method | DNN (specific background and infected region) | 0.701 | 0.942 | 0.757 | 0.082 | 0.867 | 0.783 |
Models are analysed in Reference 19 using 45 training images and 50 test images.
589 824 training point from 30 images and 20 test images, 30 test images.
Percentage improvement compared between our model with existing work
| Models | Sensitivity | Specificity | Dice | MAE | EMφ |
|
|---|---|---|---|---|---|---|
| U‐Net | 31.2% | 9.79% | 72.43% | 44% | 38.7% | 25.8% |
| Attention‐UNet | 10.04% | 26.61% | 29% | 73% | 17.3% | 5.2% |
| Gated‐UNet | 6.53% | 29.93% | 21% | 80% | 6.5% | 8% |
| Dense‐UNet | 18.01% | 43.8% | 46% | 44.4% | 30.9% | 19% |
| U‐Net++ | 4.31% | 30.47% | 30% | 68% | 20.4% | 6% |
| Inf‐Net | 1.30% | 20.6% | 10.9% | — | 3.4% | 0.25% |
| Semi‐Inf‐Net | — | 17.75% | 2.4% | — | — | — |
| Avg. improvement with our method | 10.19% | 25.5% | 30.2% | 44.2% | 16.7% | 9.17% |
Note: —, Unavailability of data.