Literature DB >> 33362346

An integrated feature frame work for automated segmentation of COVID-19 infection from lung CT images.

Deepika Selvaraj1, Arunachalam Venkatesan1, Vijayalakshmi G V Mahesh2, Alex Noel Joseph Raj3.   

Abstract

The novel coronavirus disease (SARS-CoV-2 or COVID-19) is spreading across the world and is affecting public health and the world economy. Artificial Intelligence (AI) can play a key role in enhancing COVID-19 detection. However, lung infection by COVID-19 is not quantifiable due to a lack of studies and the difficulty involved in the collection of large datasets. Segmentation is a preferred technique to quantify and contour the COVID-19 region on the lungs using computed tomography (CT) scan images. To address the dataset problem, we propose a deep neural network (DNN) model trained on a limited dataset where features are selected using a region-specific approach. Specifically, we apply the Zernike moment (ZM) and gray level co-occurrence matrix (GLCM) to extract the unique shape and texture features. The feature vectors computed from these techniques enable segmentation that illustrates the severity of the COVID-19 infection. The proposed algorithm was compared with other existing state-of-the-art deep neural networks using the Radiopedia and COVID-19 CT Segmentation datasets presented specificity, sensitivity, sensitivity, mean absolute error (MAE), enhance-alignment measure (EMφ), and structure measure (S m) of 0.942, 0.701, 0.082, 0.867, and 0.783, respectively. The metrics demonstrate the performance of the model in quantifying the COVID-19 infection with limited datasets.
© 2020 Wiley Periodicals LLC.

Entities:  

Keywords:  Zernike moment; artificial intelligence; computed tomography image; deep neural network; feature extraction; limited training points; segmentation

Year:  2020        PMID: 33362346      PMCID: PMC7753711          DOI: 10.1002/ima.22525

Source DB:  PubMed          Journal:  Int J Imaging Syst Technol        ISSN: 0899-9457            Impact factor:   2.177


INTRODUCTION

The novel coronavirus disease‐2019 (COVID‐19) is a lung infection caused by Severe Acute Respiratory Syndrome (SARS) and is transmitted from person to person easily. The first human COVID‐19 case has been reported in Wuhan city, China in December 2019. This disease was traced during the sewage water study from Milan and Turin in Italy before China identified the first case and the related information was shared in a recent article. The World Health Organization (WHO) has reported 10.59 million COVID cases across 213 countries and territories, as of June 2020. A large percentage of the population at present is affected by this disease. The recovery and mortality rates are 54.22% and 4.91% approximately. All over the world, researchers from biomedical departments are involved in an effort to find the effective vaccine for the COVID‐19. However, early detection is very important for choosing the right type of treatment to control the disease and avoid the spread. AI has been crucial in helping and serving the automated diagnostics and public health surveillance to analyse the severity of the COVID‐19 as well as in protecting the people from the virus spread. In the human body, the lungs are the major organs of the respiratory system. The disease in the lung will affect the entire air circulation system and may even result in death. The other major diseases that affect the lungs are lung cancer, tuberculosis, pneumonia. , Also, some microorganisms such as fungus, bacteria, and virus may cause illness which requires proper medical treatment to cure the infection/disease. Likewise, the infection caused due to the COVID‐19 virus requires the same. The most commonly used COVID‐ 19 tests are RT‐PCR and antigen testing. Reverse transcription polymerase chain reaction (RT‐PCR) is the standard COVID‐19 test used to detect the viral nucleic acid that provides the genetic information of the virus. The test requires nasopharyngeal and throat swabs collected from the infected person. The duration of the test is too long and the result of RT‐PCR is affected by sampling errors and low viral load. The antigen test is designed to detect the spike protein that is responsible for facilitating the entry of virus into the human cell. Though the test is fast, its sensitivity is poor. An alternate detection method for COVID‐19 is based on medical imaging techniques that identify the infected location and infection rate. Medical imaging records the information in the lung region using chest radiography (X‐rays) and Computed‐Tomography (CT) scan. , , , Compared to the X‐rays, a CT scan is mostly preferred for detecting the radiographic features of COVID‐19 from the three‐dimensional view of the lung's region. Further, the CT scans can be used to give two‐dimension views: axial view, coronal view, and sagittal view for the proper COVID infection detection. CT scan imaging of lungs gives good sensitivity to examine the COVID‐19 infected region even before the medical symptoms occur. The spread of the infection due to COVID‐19 in the lungs are identified through Ground glass opacification (GGO) and pulmonary consolidation phases. GGO is identified in the early stage of the infection whereas pulmonary consolidation indicates the final stage of the disease. These phases are observed from the CT scan images of the COVID cases. , The qualitative rate of the disease indicated in CT scans provides important information to guide against COVID‐19. Rajinikanth et al. suggested that infection visibility analysis using CT scans gives better capability and reliability in the detection of the disease using the two‐dimensional (coronal) view. The dataset used is obtained from the Radiopedia database. The experimental analysis and the simulation of the model were performed in a MATLAB environment. Accurate segmentation of radiographic features is important to quantify the disease in CT scan images. Segmentation of medical images needs marking manually by experts. As the number of COVID cases is increasing rapidly, it is difficult to label the infected region manually. Thus, automatic segmentation is needed from the CT scans to detect the infected region. From the above consideration, our contributions are as following: We propose a Deep Neural Network (DNN) model to detect and segment the COVID‐19 from the axial view CT scan image from Radiopedia and COVID‐19 CT Segmentation database. Our model provides better performance with limited number of training points. Using the region‐specific approach, instead of taking all the training points or images, limited training points are chosen from the infected region as well as the background region of the CT scan image. Before selecting the training points, for each image, the shape and texture descriptors are extracted using Zernike moments and GLCM, respectively. The performance of the proposed method is evaluated using the standard metrics for the different test images and evaluated metrics are compared with the existing models. , , , , , The methodology needs enhancement techniques for CT scans of the COVID‐19 dataset, feature extraction and appropriate classifier model to detect and segment the infected region. The rest of the paper is arranged as follows. Section 2 discusses the related works on the COVID‐19 using AI techniques. Section 3 gives a detailed explanation and workflow of methodology and implementation. Sections 4 and 5 outline the experimental design, results and discussion. Section 6 provides the conclusion.

RELATED WORKS ON THE COVID‐19 USING AI

A comprehensive list of work for COVID‐19 image‐based AI techniques are found in Maga et al. and Fu et al. Developing a deep network gives more benefits for automatic and fast segmentation of the medical images. Camouflaged object detection (COD) is introduced to identify the embedded object with their surroundings. COD is beneficial in medical image applications such as lung infection segmentation. Ali et al. proposed the integration approach consisting of unsupervised machine learning (self‐organizing map), dimensionality reduction (principal component analysis) and computational classification (Adam Deep Learning) to present a better classification performance. Particularly, the U‐Net model works well in many of the segmentation tasks for medical images. There are a few applications that have adopted U‐Net for liver, heart, and multi‐organ segmentation. , , Chen et al. has proposed a deep learning model for multiple regions auto segmentation for COVID‐19 and used the aggregated residual transformations to extract the features from the CT image. Contrastive learning is proposed to train the encoder module, which provides the expressive feature information from the publicly available CT image COVID‐19 dataset. The Deep Learning model requires a lesser number of image samples for training the dataset to provide an automatic classification of COVID‐19 infected images. Also, the authors concluded that the contrast learning method achieves better performance than ResNet‐50. Narin et al. have analysed the ROC result of three Convolutional Neural Network (CNN) models such as ResNet‐50, InceptionV3, and Inception‐ResNetV2 using a chest X‐ray image. Shi et al. have discussed pre‐processing the COVID CT images using location‐specific feature extraction and infection Size Aware Random Forest method (iSARF) to distinguish the infection level and then classification using random forest. The iSARF performed well by providing Sensitivity, Specificity and Accuracy of 0.907, 0.833 and 0.879, respectively. Yujin et al. proposed, the patch‐based CNN using a novel probabilistic Grad‐CAM saliency map with a limited number of training points from the chest X‐rays. Wang et al. proposed an open‐source CNN model called COVIDNet, which is trained and tested by the COVIDx dataset. It is useful to identify the COVID‐19 infection from chest X‐rays and has achieved a sensitivity of 80%. Lung Infection Segmentation Deep Network (Inf‐Net) used to detect the COVID‐19 infection in the lungs, from the CT image slices where a parallel partial decoder is adapted to extract the high‐level features. Later implicit reverse attention and explicit edge attention are used to enhance the features. Charmaine et al. have presented the deep learning model using the location‐attention oriented approach to calculate the probability of COVID‐19 infection region. Chest CT dataset is used for training the deep learning system. A novel semi‐supervised shallow learning model including Parallel Quantum‐Inspired Self‐supervised Network (PQIS‐Net) with Fully‐Connected (FC) layers for automatic segmentation of COVID‐19 CT image is proposed in Konar et al. The patch‐based classification was applied to the segmented images for the diagnosis of COVID‐19 using the two publicly available datasets. The efficiency (F1‐score and AUC) of the PQIS‐Net was compared with pre‐trained convolutional based models. The truncated VGG‐19 model was proposed to analyse the COVID‐19 CT scans. TheVGG‐16 model was used to extract features from the CT images using fine‐tuning. The feature selection was achieved through Principal Component Analysis (PCA) and the classification was done using four different classifier methods such as deep convolutional neural network (CNN), Bagging Ensemble with support vector machine (SVM), Extreme Learning Machine (ELM) and Online sequential ELM on 208 test images. From the different classifiers, Bagging Ensemble with SVM have achieved the following: an accuracy of 95.7%, AUC0.958 and F1‐score 95.3%. The three‐phase COVID‐19 CT image detection model is introduced in Ahuja et al. The modules are as follows: (a) Augmentation using stationary wavelets; (b) COVID‐19 detection using pre‐trained CNN model such as ResNet18, ResNet50, ResNet101, and SqueezeNet; (c) Abnormality localization in CT scan images. The experimental analysis showed that the pre‐trained ResNet18 transfer learning model has given a better classification accuracy of 99.82% for training and 99.4% for testing. A pipeline model is presented in Dey et al. with several sub‐modules to classify the segmented region from the COVID‐19 images. First, the COVID‐19 CT images are segmented using the using Social‐Group‐Optimization and Kapur's Entropy thresholding, followed by K‐means clustering and morphology‐based segmentation. Next, a classification module is introduced to classify the segmented region. Here PCA based fusion technique is used to fuse the features and then fused features are trained with different classifiers such as Random Forest, k‐Nearest Neighbours (KNN), Support Vector Machine with Radial Basis Function, and Decision Tree. Experimental results indicated an accuracy of 91% and 87% with Morphology‐based segmentation and kNN classifier. A summary of the various techniques available for COVID‐19 screening is presented in Table 1.
TABLE 1

Literature summary of related techniques for COVID‐19 screening

ReferenceTechniqueDatasetFindings
33 Aggregated residual transformations to learn a robust and expressive feature representation.110 COVID‐19 CT image of size 512 × 512 are collected from 60 patients.

The model gives 0.95 precision and 0.89 accuracy.

12.7% of accuracy and 14.5% of precision are improvement with U‐Net model.

34 Contrastive learning method.MedSeg dataset:110 CT COVID‐19 images. COVID‐19 CT dataset: 349 CT COVID‐19 images 397 Non‐COVID images.ResNet‐50 performance gives accuracy of 0.868, recall of 0.872, AUC of 0.931.
35 ResNet50, InceptionV3 and InceptionResNetV2.Chest X‐ray 100 images: 50 non‐infected image and 50 infected images. All images are 224 × 224 pixel size.ResNet50 model provides good classification performance with 98% accuracy than other two proposed models (97% accuracy for InceptionV3 and 87% accuracy for Inception‐ResNetV2).
36 An infection Size Aware Random Forest method (iSARF).COVID‐19 images are taken from 1658 patients.iSARF yielded sensitivity of 0.907, specificity of 0.833, and accuracy of 0.879.
37 A patch‐based convolutional neural network approach.COVID‐19 dataset: 180 infected image of size 224 × 224.

Global approach gives 70.7% of accuracy, 60.6% of precision, 60.1% of recall, 59% of F1‐score.

Local approach gives 88.9% of accuracy, 83.4% of precision, 85.9% of recall, 84.4% of F1‐score.

19 COVID‐19 Lung Infection Segmentation Deep Network (Inf‐Net) with a parallel partial decoder method is to aggregate the high‐level features.

Dataset consists of 100 axial CT images from different COVID‐19 patients.

45 CT images randomly selected as training, 5 CT images for validation, and the remaining 50 images for testing.

Inf‐Net achieves 0.692 of sensitivity, 0.943 of specificity, 0.725 of sensitivity, 0.960 of specificity.
40 Semi‐supervised shallow neural network model using Parallel self‐supervised neural network model (PQIS‐Net).

Variable size 2482 lung CT scans: 1252 COVID‐19 infected CT image and 1230 Non‐infected COVID image.

20 labelled COVID‐19 CT image of size 512 × 512 which includes infection region masks, lung masks (left and right) and lung‐infection pair masks.

Accuracy, precision, f1 score, and AUC are 0.931, 0.890, 0.826 and 0.982, respectively.
41

Truncated VGG‐19 with fine‐tuning method.

Classification is done with four different classifier method such as deep convolutional neural network (CNN), Bagging Ensemble with support vector machine (SVM), Extreme Learning Machine (ELM) and Online sequential ELM.

CT Scan Dataset: 344 COVID‐19 images, 358 Non‐COVID‐19 images.Best performing classifier Bagging Ensemble with SVM achieves 95.7% accuracy, 95.8% precision, 0.958 AUC and 95.3% F1‐score.
42 Transfer Learning method with argumentation: ResNet18, ResNet 50, ResNet101, SqueezeNet.COVID‐19:349 CT scan images and Non‐COVID‐19 CT: 397 CT scan images.

ResNet18 achieves the best performance model.

Training accuracy is 99.82%, validation accuracy is 97.32% and testing accuracy is 99.4%.

43 Social‐group‐optimization with Kapur's entropy thresholding and then followed by k‐means clustering and morphology‐based segmentation.COVID‐19 dataset consists of 400 numbers of greyscale lung CT image (200 Non‐infected image and 200 infected images).Morphology‐based segmentation and KNN gives more than 91% and 87% accuracy.
Literature summary of related techniques for COVID‐19 screening The model gives 0.95 precision and 0.89 accuracy. 12.7% of accuracy and 14.5% of precision are improvement with U‐Net model. Global approach gives 70.7% of accuracy, 60.6% of precision, 60.1% of recall, 59% of F1‐score. Local approach gives 88.9% of accuracy, 83.4% of precision, 85.9% of recall, 84.4% of F1‐score. Dataset consists of 100 axial CT images from different COVID‐19 patients. 45 CT images randomly selected as training, 5 CT images for validation, and the remaining 50 images for testing. Variable size 2482 lung CT scans: 1252 COVID‐19 infected CT image and 1230 Non‐infected COVID image. 20 labelled COVID‐19 CT image of size 512 × 512 which includes infection region masks, lung masks (left and right) and lung‐infection pair masks. Truncated VGG‐19 with fine‐tuning method. Classification is done with four different classifier method such as deep convolutional neural network (CNN), Bagging Ensemble with support vector machine (SVM), Extreme Learning Machine (ELM) and Online sequential ELM. ResNet18 achieves the best performance model. Training accuracy is 99.82%, validation accuracy is 97.32% and testing accuracy is 99.4%. Wang et al. have proposed the weakly‐supervised deep learning (UNet and DeCoVNet) using 3‐Dimensional CT volumes to identify COVID‐19.The literature indicates the successful evaluation of COVID‐19 infection using deep learning and traditional machine learning methods by various researchers. But the study also clarifies that some of the regions are not evaluated in the classifier layer due to lack of training data samples and inefficient feature representation. The detailed design flow of the proposed framework is explained in the next section.

METHODOLOGY

The Datasets used in the proposed framework consists of 80 images of size 512 × 512 pixels (3 channels). The segmentation process of each CT image involves (a) pre‐processing, (b) extracting the robust and sensitive features, (c) training the region‐specific region, and (d) segmenting the region of interest (infected region) from the background. The block diagram of the proposed methodology is as shown in Figure 1.
FIGURE 1

Block diagram for the flow of binary classification for computed tomography image

Block diagram for the flow of binary classification for computed tomography image

Pre‐processing and feature extraction methods

The pre‐processing is based on the standard enhancement approach called adjust image intensity values (imadjust). This approach describes the expressive information from the COVID‐19 CT image by adjusting the contrast of the GGO and pulmonary consolidation region. The infected pixels are brightened visually as shown in Figure 2. Also, the Contrast Improvement Index (CII) metrics are analysed to highlight the importance of pre‐processing on the COVID dataset. The image quality (pixel intensity) is improved by 30% to 40% when compared to the original CT image as shown in Table 2. CII is computed based on Equations (1) and (2). where C enhanced and C original are the contrast of the enhanced and original images respectively, C is the average value of the contrast in the enhanced or original image. I and I are the mean value for foreground and background contrast of the image. A higher CII value indicates improvement in image quality.
FIGURE 2

A, Original computed tomography image; B, Image after enhancement

TABLE 2

CII values for the pre‐processed COVID‐19 CT images

Number of imagesContrast for the original CT imageContrast for the pre‐processed imageCII for the pre‐processed image
10.59840.81661.3646
20.42640.73481.7234
30.16500.32791.9872
40.48080.66981.3931
50.48460.94971.9599

Note: Default CII value for the original image is 1.000.

A, Original computed tomography image; B, Image after enhancement CII values for the pre‐processed COVID‐19 CT images Note: Default CII value for the original image is 1.000. The enhanced images present significant details for extracting the features. The extraction of features is a description of each pixel of the CT image that contains descriptive information in the form of a stack of the vectors (feature vector). The feature vector helps to find out the output class (infected COVID‐19 region and background region). The feature vector considered for the work is the integration of shape and texture features. The texture features are derived from the gray level co‐occurrence matrix and shape descriptor features are extracted from the Zernike moment.

Zernike moment (ZM) based shape descriptor features and gray level co‐occurrence matrix (GLCM) for texture features

Based on the infection stage/level, the infected region of COVID‐19 CT images has dissimilarity in shapes, sizes and structures of the geometry. The shape descriptor is good to define the infected pixel information range from its background region. The shape parameters can be extracted using descriptor based on Zernike and Hu moments. , For better representation, the shape descriptor should possess the properties such as (a) provide low redundancy and higher discrimination ability. (b) Rotation, scale and translation invariance, and (c) present coarse to finer detail representation. Hu moments require extensive computation power to calculate higher order moments. Therefore, we choose the ZM based shape descriptor for representing shape features of the COVID‐19 lungs image. The ZM is calculated for every pixel of the pre‐processed image by sliding a window of size 5 × 5 with a stride of 1 along the column and row. For the computation of ZMs, different polynomial orders (n) and the corresponding repetition factor (m) were considered for the function I(x, y). Next, we compute the Zernike moments, by projecting I(x, y) onto the set of complex Zernike polynomials as given in Equation (3). where, P (x, y) is Zernike polynomial, I(x, y) is the image of size 5 × 5 pixels, n is a non‐negative integer, m is an integer such that , and n − |m| is even. The function P (x, y) is the complex conjugate of the orthogonal basis function P (x, y) defined as: where, and . The radial polynomial R (r) is defined as: The Zernike moment obtained from the Equation (3) is a complex number represented as shown in Equation (6): Equation (7) is the magnitude of the Zernike moment |Z | that denotes the ZM shape descriptor for specified n and m. Thus, a total number of 36|Z |were obtained with n ranging over 1 to 10 and matching combinations of n, m with m such that 0 ≤ |m| ≤ n, and n − |m| is even for every pixel. The magnitude of ZM calculated for every pixel to find the ZM feature maps are as displayed in Figure 3.
FIGURE 3

Zernike moment features with different combination of n and m [Color figure can be viewed at wileyonlinelibrary.com]

Zernike moment features with different combination of n and m [Color figure can be viewed at wileyonlinelibrary.com] From the observation of Zernike features, we find that all the 36 different |Z | provide unique information about the infected and background region. To analyse the Zernike robust feature (ZM ) and Zernike sensitive feature (ZM ), each point is taken from both the regions with 36 ZM for different n, m and are plotted as indicated in Figure 4 and the separation of the ZM and ZM is indicated in Equations (8) and (9).
FIGURE 4

Experimental analysis of low‐level and high‐level feature from Zernike moment (total 36 Zernike feature, red – infected point, green – background point) [Color figure can be viewed at wileyonlinelibrary.com]

Experimental analysis of low‐level and high‐level feature from Zernike moment (total 36 Zernike feature, red – infected point, green – background point) [Color figure can be viewed at wileyonlinelibrary.com] From the different combinations of ZM feature extraction image, we plot the intensity range for the set of pixels for the 1st image of the dataset. The graphical plot presented in Figure 5 represents number of pixels in x‐axis and ZM features in y‐axis. The above observation indicates that the Zernike features from background information are dense at the bottom region and the infected region information mostly occurs above a specific threshold.
FIGURE 5

Plots for Zernike moment features from specific region of the training image (Red points represent background and Green points represents infected pixels) [Color figure can be viewed at wileyonlinelibrary.com]

Plots for Zernike moment features from specific region of the training image (Red points represent background and Green points represents infected pixels) [Color figure can be viewed at wileyonlinelibrary.com] To make the binary classification more accurate, we add an extra texture feature computed using the GLCM. The texture feature is more important in medical imaging (CT image) for extracting the GGO and pulmonary consolidation features. GLCM consists of the relationship of the different angles between image pixels. If the gray level co‐occurrence matrix obtained from an image be denoted as . In this point, GLCM is used to estimate the features of r th pixel frequency with the feature of sth pixel frequency by the length (d = 1) and direction (θ = 0). The computation of the texture feature follows the same procedure as that of ZM, that is, computing the texture features for every pixel by moving a window of 5 × 5 with a stride of 1 over the pre‐processed image. From GLCM, texture features such as entropy, variance, contrast, correlation, dissimilarity, energy and homogeneity are calculated. Among all these features, we considered the GLCM variance and contrast as shown in Equations (10) and (11). These two features are combined to form GLCM features Gf that gives proper discrimination of infected region and background region from the CT image. where, i refers to the number of training images, ‘ ’ denotes mean and ‘N’ denotes the image size, Gf refers the GLCM features (2 features). The intensity for the GLCM feature images (contrast and variance) is plotted as shown in Figure 6. It represents number of pixels in x‐axis and GLCM features in y‐axis. Here the red points denote infected region and green points denote background respectively. The observation gives clear distinction between the background region and infected region.
FIGURE 6

Plots for GLCM features from specific region of the training image (Red points represents background and Green points represents infected pixels) [Color figure can be viewed at wileyonlinelibrary.com]

Plots for GLCM features from specific region of the training image (Red points represents background and Green points represents infected pixels) [Color figure can be viewed at wileyonlinelibrary.com] Now the shape features computed from the ZM and the texture features obtained from the GLCM are integrated to form a feature vector of 38 features as indicated in Equation (13). The next section discusses about the formation of training dataset from Fs features.

EXPERIMENTAL DESIGN

COVID‐19 dataset

COVID‐19 dataset (axial view) is collected from References 17 and 50 to evaluate our implementation. The previous implementations have used X‐rays and CT image datasets for segmentation and classification of the infected COVID regions. , The COVID‐19 CT dataset used in this work contains 80 images of size 512 × 512 pixels (3 channel) along with the ground truth image (binary image). Ground truth images are useful for the identification of targets from the input dataset. From the dataset, 30 images are selected for training the network, the remaining 50 images are used for testing the trained network for performance evaluation.

Training and testing dataset

COVID‐19 CT images contain additional background pixels (without infected region) other than infected pixels. Using the whole image as the training data gives testing results with less accuracy that reduce the overall performance. It should be able to classify the background region from the infected region during the testing. Hence, a balanced training dataset is required for the both infected region If and background Bf pixels along with all the features from the training images. The works , , have used the whole image as training data. The training result has delivered a performance greater than 90% as compared to the overall testing accuracy. However, as explained in Fan et al. the overall sensitivity of the test dataset is reduced by not selecting the true positive target properly (infected pixels). Hence, we chose properly infected pixels and background pixels from the extracted images.

Region‐specific for formation of training points

The dataset of training points is manually selected from the infected and background pixels from the 30 training images (after pre‐processing and Fs ) by using the region‐ specific approach. The set of features (S ) is formed by marking the specific region (infection region (If) and background region (Bf)) for each feature from the training image dataset as shown in Figure 7.
FIGURE 7

A, Specific region of background (Bf) and infection features (If) are separated out for training. B, Histogram plot for Bf (blue colour) and If (brown colour) [Color figure can be viewed at wileyonlinelibrary.com]

A, Specific region of background (Bf) and infection features (If) are separated out for training. B, Histogram plot for Bf (blue colour) and If (brown colour) [Color figure can be viewed at wileyonlinelibrary.com] We chose the If region with 62 pixels and Bf region as 456 pixels. This feature helps to classify the image pixels as infected or not from COVID‐19. The algorithm for region‐ specific approach is explained below: Region selection is helpful in getting balanced image information between background and infected pixels. The selected pixels are rearranged into the vector points for training. The total training points consist of 589 380 pixels with 519 371 background pixels and 70 009 infected pixels. In the total training points, each image contains 19 646 points of 38 features. A large number of points are taken into COVID infected region to reduce the misclassification of testing image features. During the training process, the feature vector is labelled using the ground truth as shown with T as the target vector represented in binary form. here f is the formed training dataset, S is the input features, T is target for the corresponding input features, i refers to the number of training images, k refers to the number of training points (k = 1 to 589 380 points from all features of Fs ). From the T, class 1 = ‘1’ and class 2 = ‘0’ represents an infected pixel, class 1 = ‘0’ and class 2 = ‘1’ represents the background pixel. Furthermore, to select a suitable classifier for the proposed work, the performance of different classifiers such as Support Vector Machine (SVM), Deep Neural Network (DNN), Decision Tree (DT), Logistic Regression (LR) and Gaussian Naive Bayes (GNB) are compared using the training data points. The performance of classifiers is evaluated from the confusion matrix using the standard metrics: Area Under Curve (AUC), Specificity, Sensitivity and Accuracy as presented in Table 3. From Table 3, it is observed that the DNN classifier outperforms the other classifier in terms of Accuracy, AUC and Sensitivity for the given training points and corresponding confusion matrix for the f for DNN has been presented in Table 4. AUC is estimated from the ROC graph for the different classifiers that are as plotted in Figure 8.
TABLE 3

Evaluation of training points metrics for different classifiers

Different classifiersAUCSensitivitySpecificityAccuracy (%)
SV M 0.87230.55490.898888.6
GNB 0.91070.40250.973583.3
LR 0.81870.52110.895088.2
DT 0.940.7560.959393
DNN 0.9427 0.7678 0.9285 93.8

Note: The significance of the bold numbers are highlighted results of the proposed algorithm.

TABLE 4

Confusion matrix of training points for DNN classifier

T i
Infected pixelsBackground pixelsAccuracy
Training outputInfected pixels46 469 (7.9%)12 684 (2.2%)93.8%
Background pixels23 540 (4.0%)506 687 (85.9%)
FIGURE 8

ROC for different classifiers [Color figure can be viewed at wileyonlinelibrary.com]

Evaluation of training points metrics for different classifiers Note: The significance of the bold numbers are highlighted results of the proposed algorithm. Confusion matrix of training points for DNN classifier ROC for different classifiers [Color figure can be viewed at wileyonlinelibrary.com]

Binary classification using DNN

The set of ZM and GLCM features extracted using the methodology as explained in Section 3 form the feature set. The corresponding ground truth data is considered as a target. A supervised learning method is implemented by training a Deep Neural Network (DNN). Our architecture is identical to the basic feed‐forward network with multiple hidden layers. This work considers the architecture where the size of the input layer is 38 neurons (38 features), three hidden layers with 58 neurons per layer and binary classification output layer as shown in Figure 9.
FIGURE 9

Deep neural network (DNN) architecture

Deep neural network (DNN) architecture The total number of learnable parameters (weights and biases) in the neural network is 9224. The first layer has 2262 (38 × 58 + 58), the second layer 3422 (58 × 58 + 58), the third layer 3422 (58 × 58 + 58) and the final layer with 118 (58 × 2 + 1). The loss function (cross‐entropy loss) converges by modifying the parameters using Scaled Conjugate Gradient (SCG) training function. The three hidden layers are followed with the activation function called tan‐sigmoid and the output layer involves the softmax operation that gives the probability of two outputs. The probability of the first neuron in the output layer represents the lungs infected region and the second neuron is complementary of the first probability called background region. The training required 8 minutes and 49 seconds to get the best error performance of 0.1367 at 100 epochs.

RESULTS AND DISCUSSION

The trained DNN classifier is evaluated for the test image (50 images) from the COVID‐19 dataset. After the pre‐processing and Fs , the test vector is formed for each test image of size 38 × 256 × 256. All the experimental analysis is done in MATLAB R2020a run on Intel Corei3‐2330 M CPU. The feature generation process takes around 250 minutes (about 4 hours) for the test set of 50 images and testing time for each test image is around 4 to 7 minutes. The challenges of automatic segmentation from the CT image are. CT Images are varying in terms of texture, shape, and position across different cases and consolidation is small, which may result in a false negative output from a whole CT image. Due to the low contrast and blurriness within the GGO of the infection area, there is a significant difficulty in the identification of the contours in the CT images. Artificial intelligence (AI) model requires a greater number of training data to train the model and more human/ specialist resources are required to collect and annotate the ground truths of the COVID‐19 dataset. As explained in Section 3, the feature vector of 38 features is given to the trained DNN model as input. The output of the classifier gives the probability score for two classes (infected and background). This probability score decides and falls on particular class 1 or class 0. The probability vector output is reshaped into the image size 256 × 256. Further, the image of size 256 × 256 is compared with the ground truth image to detect and segment the COVID‐19 regions and also evaluate the performance of the model. To make the output binary image effective, the morphological operation is applied to the output binary image. Different morphological structuring element produces different test image results. Hence, we find the suitable structuring element with a proper size for all the test images to achieve the overall average accuracy and selecting the structuring element that gives high accuracy. The average accuracy vs morphological structuring element graph for COVID‐19 CT image test data is shown in Figure 10. From the plot, it is determined that “imclose” disk structuring element of size ‘2’ reaches the maximum accuracy. The test dataset results are displayed in Figure 11. In Figure 11d column, GT contours are overlaid on the output image which gives visual understanding of the non‐distinguishable pixels. This arises due to the poor lightening between infected pixels and background pixels from the original image as shown in Figure 11 (3rd‐row image) and also the thin occurrence of GGO and consolidation in Figure 11 (6th‐row image). From the ablation study Table 5, 36 ZM shape descriptor training feature points alone do not achieve better results in Sensitivity and Mean IOU. Shape descriptor features do not capture the proper infected region (True Positive) information from the test image. so that, 2 GLCM texture features are added with 36 ZM features and trained the DNN model with 38 features. Evaluated metrics achieve better results with the combination of ZM and GLCM features (38 features) compared to the 36 features results. We have analysed each test image performance by standard metrics such as Accuracy (A), Sensitivity or Recall (R), Specificity, F1 score, Dice index, Precision (P), Mean IOU and score using the confusion matrix as represented in Table 6.
FIGURE 10

Test dataset plot for morphological structuring element vs average accuracy [Color figure can be viewed at wileyonlinelibrary.com]

FIGURE 11

Segmentation results for the COVID‐19 dataset. A, Column shows computed tomography axial view input image; B, Column shows ground truth image; C, Column shows the output from deep neural network model and D, Column shows overlay the GT contour on output image for visual understanding [Color figure can be viewed at wileyonlinelibrary.com]

TABLE 5

Performance evaluation for the test dataset (36 ZM features)

Number of test imagesAccuracy (A) = TP+TNTP+TN+FP+FN IOU scoreMean IOU = C0+C12 Sensitivity or Recall (R) = TPTP+FN Specificity = TNTN+FP F1 score = 2×P×RP+R Precision (P) = TPTP+FP
Class 0 (C 0) = TNTN+FP+FN Class 1 (C 1) = TPTP+FP+FN
183.50.3930.8150.6040.4140.980.5640.884
285.60.3370.8430.590.3810.9690.5050.747
388.80.4590.9360.6970.4900.9920.6290.879
491.50.1660.9280.5470.1770.9940.2850.736
589.30.3330.9250.6290.3710.9870.50.765
682.20.3820.8810.6310.4530.9670.5520.708
797.70.0580.8950.4760.0600.990.1100.583
886.30.0860.8720.4790.1010.9770.160.375
9830.660.9220.7910.7110.9840.6000.915
1083.30.2380.8240.5310.2850.9550.3850.590
TABLE 6

Confusion matrix for test image 1

GT
Infected pixelsBackground pixelsAccuracy
Testing outputInfected pixels14 025 (21.4%)2228 (3.4%)92.30%
Background pixels2818 (4.3%)46 465 (70.9%)
Test dataset plot for morphological structuring element vs average accuracy [Color figure can be viewed at wileyonlinelibrary.com] Segmentation results for the COVID‐19 dataset. A, Column shows computed tomography axial view input image; B, Column shows ground truth image; C, Column shows the output from deep neural network model and D, Column shows overlay the GT contour on output image for visual understanding [Color figure can be viewed at wileyonlinelibrary.com] Performance evaluation for the test dataset (36 ZM features) Confusion matrix for test image 1 Additionally, three more important metrics are Enhance‐alignment Measure (EMφ), Mean Absolute Error (MAE) and Structure measure (S m) are used for performance evaluation. Enhance‐alignment Measure (EMφ) is a newly proposed metric. It is useful to estimate the global and local similarities of two binary images as presented in Equation (16). where, W represents width, H represents the height of GT, F represent the final predicted output, (x, y) represents each pixel coordinate of GT and F. φ denotes the enhanced alignment matrix. Mean Absolute Error (MAE), measures the error between the F and GT of each pixel in the image as indicated in Equation (17). Structure measure (S m), measures the structural similarity between the GT mask and prediction map as represented in Equation (18). Where , the default setting α = 0.5 represents the balance factor between region‐aware similarity (S ) and object‐aware similarity (S ). The performance of each test image is evaluated by metrics as listed in Table 7.
TABLE 7

Performance evaluation for the random 10 test image from COVID‐19 dataset (36 ZM and 2 GLCM features)

Number of test imagesAccuracy (A) = TP+TNTP+TN+FP+FN IOU scoreMean IOU = C0+C12 Sensitivity or Recall (R) = TPTP+FN Specificity = TNTN+FP F1 score = 2×P×RP+R Precision (P) = TPTP+FP Dice indexMAEEMφ S m
Class 0 (C 0) = TNTN+FP+FN Class 1 (C 1) = TPTP+FP+FN
10.92300.900.730.8150.8320.950.8470.860.80280.10030.90330.8363
20.92370.9070.680.7930.8760.930.8150.760.73690.09450.92770.8431
30.96100.950.660.8050.7230.980.7950.880.81210.10250.85120.8017
40.94470.940.540.740.8350.950.7050.610.79230.06780.77780.8411
50.95530.950.560.750.6000.990.7200.9060.64790.08630.81220.7501
60.95790.950.740.8450.8820.960.8510.820.74880.07030.93660.8416
70.97880.970.30.620.3300.980.4320.660.6730.03940.96460.6475
80.93940.9380.310.6240.3460.980.4730.810.63660.08080.76270.6103
90.87710.8480.610.7290.8280.890.7590.700.66650.16240.84220.7391
100.92480.9100.670.790.8520.940.7040.760.69450.11280.89620.8212
Performance evaluation for the random 10 test image from COVID‐19 dataset (36 ZM and 2 GLCM features) We have compared our method with some of the existing works that employ deep network for the segmentation of the COVID‐19 infected region from the CT image dataset as presented in Table 8. In Ronneberger et al, standard U‐Net model is evaluated using 45 training images. The performance of the outcome is low in terms of sensitivity, specificity, MAE, Dice and EMφ. The modernized version of the U‐Net model is described in References 21, 22, 23, 24 which resulted in a poor evaluation of the Dice index, sensitivity compared to our developed method. In our method, MAE is quite higher than the Semi‐Inf‐Net and Inf‐Net model reported in Fan et al. EMφ, S m and Sensitivity are moderately 3%, 2.12%, 3.3% lesser compared to Semi‐Inf‐Net model. However, the overall average comparison of our test image result is much improved and superior with a limited number of training points as shown in Table 9. Also, Table 9 shows a performance improvement with our method in each of the metrics compared to other existing works. The proposed model improved the sensitivity by 31.2%, 10%, 6.5%, 18%, 4.3% and 1.3% compared with U‐Net model, Attention‐UNet, Gated‐UNet, Dense‐UNet, U‐Net++, and Inf‐Net respectively. Likewise, the percentage improvements of other metrics in relation to other state‐of‐the‐art is presented in Table 9.
TABLE 8

Comparative performance evaluation of other existing methods

Existing referenceMethodsSensitivitySpecificityDiceMAEEMφ S m
Ronneberger et al 20 U‐Net a 0.5340.8580.4390.1860.6250.622
Oktay et al 21 Attention‐UNet a 0.6370.7440.5830.1120.7390.744
Schlemper et al 22 Gated‐UNet a 0.6580.7250.6230.1020.8140.725
Li et al 23 Dense‐UNet a 0.5940.6550.5150.1840.6620.655
Zhou et al 24 U‐Net++ a 0.6720.7220.5810.1200.7200.722
Deng‐Ping Fan et al 19 Inf‐Net0.6920.7810.6820.0820.8380.781
Deng‐Ping Fan et al 19 Semi‐Inf‐Net0.7250.8000.7390.0640.8940.800
Our methodDNN (specific background and infected region) b 0.7010.9420.7570.0820.8670.783

Models are analysed in Reference 19 using 45 training images and 50 test images.

589 824 training point from 30 images and 20 test images, 30 test images.

TABLE 9

Percentage improvement compared between our model with existing work

ModelsSensitivitySpecificityDiceMAEEMφ S m
U‐Net31.2%9.79%72.43%44%38.7%25.8%
Attention‐UNet10.04%26.61%29%73%17.3%5.2%
Gated‐UNet6.53%29.93%21%80%6.5%8%
Dense‐UNet18.01%43.8%46%44.4%30.9%19%
U‐Net++4.31%30.47%30%68%20.4%6%
Inf‐Net1.30%20.6%10.9%3.4%0.25%
Semi‐Inf‐Net17.75%2.4%
Avg. improvement with our method10.19%25.5%30.2%44.2%16.7%9.17%

Note: —, Unavailability of data.

Comparative performance evaluation of other existing methods Models are analysed in Reference 19 using 45 training images and 50 test images. 589 824 training point from 30 images and 20 test images, 30 test images. Percentage improvement compared between our model with existing work Note: —, Unavailability of data.

CONCLUSION

In this paper, we propose a DNN model for COVID‐19 detection from the lung's CT axial view image. The proposed method has adopted the ZM for shape features, GLCM for texture feature and specific‐ region for selecting the training points, precisely to extract the unique information from the CT image. Our design gives an effective detection of the COVID‐19 infection in the lungs and an appropriate tool for radiologists to define the infection stage/percentage. Our model provided better results with limited number of training points. The average test dataset performance reached 70%, 94%, 86% and 78% in terms of Sensitivity, Specificity, EMφ and S m, respectively. With the proposed method, the average performance is improved in terms of Sensitivity, Specificity, Dice and EMφ by 10.9%, 26%, 24.7%, and 16.5%, respectively in comparison to other popular deep networks such as U‐Net, Gated‐UNet, Dense‐UNet, U‐Net++, Inf‐Net, Semi‐UNet. Moreover, these popular deep networks need a larger number of images for training dataset to maintain the model performance. From the results and higher evaluation metrics, it is evident that the proposed model performed significantly well with much smaller number of training points in the training dataset. The limitation of our model is the difficulty in detecting the GGO from the poor contrast CT images which require additional features and enhancement methods to provide more detailed information. In future, the inner structure information can be extracted using unique texture features and added to the training dataset to improve the performance of the DNN model and also to overcome the limitations.
  26 in total

1.  Automatic multiorgan segmentation in thorax CT images using U-net-GAN.

Authors:  Xue Dong; Yang Lei; Tonghe Wang; Matthew Thomas; Leonardo Tang; Walter J Curran; Tian Liu; Xiaofeng Yang
Journal:  Med Phys       Date:  2019-03-22       Impact factor: 4.071

2.  Large-scale screening of COVID-19 from community acquired pneumonia using infection size-aware classification.

Authors:  Feng Shi; Liming Xia; Fei Shan; Bin Song; Dijia Wu; Ying Wei; Huan Yuan; Huiting Jiang; Yichu He; Yaozong Gao; He Sui; Dinggang Shen
Journal:  Phys Med Biol       Date:  2021-02-19       Impact factor: 3.609

Review 3.  Review of Artificial Intelligence Techniques in Imaging Data Acquisition, Segmentation, and Diagnosis for COVID-19.

Authors:  Feng Shi; Jun Wang; Jun Shi; Ziyan Wu; Qian Wang; Zhenyu Tang; Kelei He; Yinghuan Shi; Dinggang Shen
Journal:  IEEE Rev Biomed Eng       Date:  2021-01-22

4.  Social Group Optimization-Assisted Kapur's Entropy and Morphological Segmentation for Automated Detection of COVID-19 Infection from Computed Tomography Images.

Authors:  Nilanjan Dey; V Rajinikanth; Simon James Fong; M Shamim Kaiser; Mufti Mahmud
Journal:  Cognit Comput       Date:  2020-08-15       Impact factor: 5.418

5.  Clinical features of patients infected with 2019 novel coronavirus in Wuhan, China.

Authors:  Chaolin Huang; Yeming Wang; Xingwang Li; Lili Ren; Jianping Zhao; Yi Hu; Li Zhang; Guohui Fan; Jiuyang Xu; Xiaoying Gu; Zhenshun Cheng; Ting Yu; Jiaan Xia; Yuan Wei; Wenjuan Wu; Xuelei Xie; Wen Yin; Hui Li; Min Liu; Yan Xiao; Hong Gao; Li Guo; Jungang Xie; Guangfa Wang; Rongmeng Jiang; Zhancheng Gao; Qi Jin; Jianwei Wang; Bin Cao
Journal:  Lancet       Date:  2020-01-24       Impact factor: 79.321

6.  Transfer learning-based ensemble support vector machine model for automated COVID-19 detection using lung computerized tomography scan data.

Authors:  Mukul Singh; Shrey Bansal; Sakshi Ahuja; Rahul Kumar Dubey; Bijaya Ketan Panigrahi; Nilanjan Dey
Journal:  Med Biol Eng Comput       Date:  2021-03-18       Impact factor: 2.602

7.  First detection of SARS-CoV-2 in untreated wastewaters in Italy.

Authors:  Giuseppina La Rosa; Marcello Iaconelli; Pamela Mancini; Giusy Bonanno Ferraro; Carolina Veneri; Lucia Bonadonna; Luca Lucentini; Elisabetta Suffredini
Journal:  Sci Total Environ       Date:  2020-05-23       Impact factor: 7.963

8.  Chest CT for Typical Coronavirus Disease 2019 (COVID-19) Pneumonia: Relationship to Negative RT-PCR Testing.

Authors:  Xingzhi Xie; Zheng Zhong; Wei Zhao; Chao Zheng; Fei Wang; Jun Liu
Journal:  Radiology       Date:  2020-02-12       Impact factor: 11.105

9.  A supervised blood vessel segmentation technique for digital Fundus images using Zernike Moment based features.

Authors:  Dharmateja Adapa; Alex Noel Joseph Raj; Sai Nikhil Alisetti; Zhemin Zhuang; Ganesan K; Ganesh Naik
Journal:  PLoS One       Date:  2020-03-06       Impact factor: 3.240

10.  CT image visual quantitative evaluation and clinical classification of coronavirus disease (COVID-19).

Authors:  Kunwei Li; Yijie Fang; Wenjuan Li; Cunxue Pan; Peixin Qin; Yinghua Zhong; Xueguo Liu; Mingqian Huang; Yuting Liao; Shaolin Li
Journal:  Eur Radiol       Date:  2020-03-25       Impact factor: 5.315

View more
  7 in total

1.  A fast lightweight network for the discrimination of COVID-19 and pulmonary diseases.

Authors:  Oussama Aiadi; Belal Khaldi
Journal:  Biomed Signal Process Control       Date:  2022-06-21       Impact factor: 5.076

2.  An integrated feature frame work for automated segmentation of COVID-19 infection from lung CT images.

Authors:  Deepika Selvaraj; Arunachalam Venkatesan; Vijayalakshmi G V Mahesh; Alex Noel Joseph Raj
Journal:  Int J Imaging Syst Technol       Date:  2020-11-23       Impact factor: 2.177

3.  A lightweight capsule network architecture for detection of COVID-19 from lung CT scans.

Authors:  Shamik Tiwari; Anurag Jain
Journal:  Int J Imaging Syst Technol       Date:  2022-01-29       Impact factor: 2.177

4.  DMDF-Net: Dual multiscale dilated fusion network for accurate segmentation of lesions related to COVID-19 in lung radiographic scans.

Authors:  Muhammad Owais; Na Rae Baek; Kang Ryoung Park
Journal:  Expert Syst Appl       Date:  2022-05-02       Impact factor: 8.665

5.  Head CT Image Segmentation and Three-Dimensional Reconstruction Technology Based on Human Anatomy.

Authors:  Zhenyu Wu; Lin Wang; Yifei Li; Shuhui Dai; Dongliang Zhang
Journal:  Comput Intell Neurosci       Date:  2022-06-16

6.  A modified DeepLabV3+ based semantic segmentation of chest computed tomography images for COVID-19 lung infections.

Authors:  Hasan Polat
Journal:  Int J Imaging Syst Technol       Date:  2022-06-11       Impact factor: 2.177

Review 7.  Review and classification of AI-enabled COVID-19 CT imaging models based on computer vision tasks.

Authors:  Haseeb Hassan; Zhaoyu Ren; Huishi Zhao; Shoujin Huang; Dan Li; Shaohua Xiang; Yan Kang; Sifan Chen; Bingding Huang
Journal:  Comput Biol Med       Date:  2021-12-18       Impact factor: 6.698

  7 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.