| Literature DB >> 35455654 |
Sergio Ortiz1, Fernando Rojas1, Olga Valenzuela2, Luis Javier Herrera1, Ignacio Rojas1.
Abstract
The coronavirus disease 2019 (COVID-19) has caused millions of deaths and one of the greatest health crises of all time. In this disease, one of the most important aspects is the early detection of the infection to avoid the spread. In addition to this, it is essential to know how the disease progresses in patients, to improve patient care. This contribution presents a novel method based on a hierarchical intelligent system, that analyzes the application of deep learning models to detect and classify patients with COVID-19 using both X-ray and chest computed tomography (CT). The methodology was divided into three phases, the first being the detection of whether or not a patient suffers from COVID-19, the second step being the evaluation of the percentage of infection of this disease and the final phase is to classify the patients according to their severity. Stratification of patients suffering from COVID-19 according to their severity using automatic systems based on machine learning on medical images (especially X-ray and CT of the lungs) provides a powerful tool to help medical experts in decision making. In this article, a new contribution is made to a stratification system with three severity levels (mild, moderate and severe) using a novel histogram database (which defines how the infection is in the different CT slices for a patient suffering from COVID-19). The first two phases use CNN Densenet-161 pre-trained models, and the last uses SVM with LDA supervised learning algorithms as classification models. The initial stage detects the presence of COVID-19 through X-ray multi-class (COVID-19 vs. No-Findings vs. Pneumonia) and the results obtained for accuracy, precision, recall, and F1-score values are 88%, 91%, 87%, and 89%, respectively. The following stage manifested the percentage of COVID-19 infection in the slices of the CT-scans for a patient and the results in the metrics evaluation are 0.95 in Pearson Correlation coefficient, 5.14 in MAE and 8.47 in RMSE. The last stage finally classifies a patient in three degrees of severity as a function of global infection of the lungs and the results achieved are 95% accurate.Entities:
Keywords: COVID-19; deep learning; hierarchical intelligent system; support vector machine
Year: 2022 PMID: 35455654 PMCID: PMC9027976 DOI: 10.3390/jpm12040535
Source DB: PubMed Journal: J Pers Med ISSN: 2075-4426
Comparative summary of different methodologies presented in the bibliography.
| Author, Year | Objective | Number of Images | Method | Metrics |
|---|---|---|---|---|
| Shambhu et al., 2021 [ | COVID-19 diagnosis (CT images) | COVID-19 = 349 | CNN | Accuracy = 86.9% |
| Chen Zhao et al., 2021 [ | Segmentation and automatic detection of COVID-19 (CT images) | COVID-19 = 58 | SP-V-Net | AUC = 94.7%, |
| Feng Shi et al., 2021 [ | COVID-19 Detection (CT images) | COVID-19 = 1.658 | VB-Net | sensitivity = 90.7%, |
| Shuai Wang et al., 2021 [ | Screen for Corona virus disease (CT images) | COVID-19 = 160 | Covid-net | Validation: |
| PanWar et al., 2020 [ | COVID-19 Detection (X-rays) | COVID-19 = 192 | nCOVnet | sensitivity = 97.62%, |
| Bougourzi et al., 2021 [ | COVID-19 Percentage Estimation (CT images) | COVID-19 = 3986 | ResneXt-50, | PC = 0.9365, |
| Loey et al., 2020 [ | COVID-19 Detection (X-ray images) | COVID-19 = 69 | Googlenet, | Four-classes: |
| Sethy et al., 2020 [ | COVID-19 Detection (X-ray images) | COVID-19 = 25 | ResNet50 | ResNet50-SVM: |
| Wang el al., 2020 [ | COVID-19 Detection (X-ray images) | COVID-19 = 266 | Covid-net | sensitivity = 91% and |
| Arifin et al., 2021 [ | Fast COVID-19 Detection (X-ray images) | COVID-19 = 50 | MobileNet-V1 | accuracy = 93.24% |
| Huang et al., 2020 [ | Quantitative Chest CT Assessment of COVID-19 | COVID-19 | CNN | opacification percentage |
| Shan et al., 2020 [ | Lung Infection Quantification of COVID-19 (CT Images) | COVID-19 = 300 | VB-Net | DSC = 91.6% ± 10.0% |
| Proposed | COVID-19 Detection (X-ray images) | COVID-19 = 50 | Densenet-161 | Accuracy = 88% |
| Proposed | COVID-19 Percentage Infection (CT images) | COVID-19 = 3986 | Densenet-161 | PC = 0.95, |
Figure 1Sample images from X-ray dataset. The images in the first row show 5 COVID-19 images. The images in the second row are 5 sample images of no-finding category in Non-COVID images. The images in the last row give 5 sample images from Pneumonia.
Figure 2Sample images from the CT dataset. The images in the first row give 5 sample slice images about 0–10–20–30–40% COVID-19 infection in the lungs and the second rows show 50–60–70–80–90% respectively.
Figure 3This diagram shows the architecture of the algorithm. (Stage 1): detect if a patient is sick with COVID-19 by studying their chest X-ray images. For this, a Densenet-161 network is used. (Stage 2) detects the percentage of infection in the lungs of a patient with COVID-19. In this stage, another Densenet-161 network is used. (Stage 3) classifies a patient within three degrees of gravity. For this purpose, a novel approach is proposed in which a histogram database is created with patients with different infection percentages. SVM with LDA is used to classify the patients in three degrees of severity.
Figure 4Densenet model network. In the Dense-block each deselayer takes all the preceding feature-maps as an input [38].
Figure 5Confusion matrix for Densenet-161 trained with X-ray images. The first matrix is relative to the SGD optimizer, the second is to the ADAM optimizer and the third is to the RMSprop optimizer. (testing phase).
Precision, Recall, F1-score and Accuracy of Densenet-161 model with X-ray images and with ADAM optimizer.
| Fold-1 | Precision | Recall | F1-Score | #Images |
|---|---|---|---|---|
| COVID-19 | 100 | 93 | 96 | 28 |
| No-findings | 80 | 95 | 87 | 97 |
| Pneumonia | 92 | 77 | 84 | 100 |
| Accuracy | 87 | 225 | ||
|
|
|
|
|
|
| COVID-19 | 100 | 89 | 94 | 28 |
| No-findings | 80 | 92 | 88 | 90 |
| Pneumonia | 91 | 86 | 88 | 107 |
| Accuracy | 89 | 225 | ||
|
|
|
|
|
|
| COVID-19 | 100 | 74 | 85 | 31 |
| No-findings | 80 | 93 | 90 | 102 |
| Pneumonia | 92 | 84 | 84 | 92 |
| Accuracy | 87 | 225 | ||
|
|
|
|
|
|
| COVID-19 | 100 | 83 | 91 | 18 |
| No-findings | 88 | 94 | 88 | 108 |
| Pneumonia | 92 | 80 | 85 | 99 |
| Accuracy | 87 | 225 | ||
|
|
|
|
|
|
| COVID-19 | 100 | 95 | 97 | 20 |
| No-findings | 88 | 92 | 90 | 103 |
| Pneumonia | 91 | 87 | 89 | 102 |
| Accuracy | 90 | 225 | ||
|
|
|
|
|
|
| COVID-19 | 100 | 86 | 93 | 125 |
| No-findings | 84 | 93 | 88 | 500 |
| Pneumonia | 90 | 83 | 86 | 500 |
| Accuracy | 88 | 1125 |
Figure 6Confusion matrix of 5 folds for Densenet-161 trained with X-ray images and with ADAM optimizer (testing phase).
Results of the second stage in the testing phase using CNN Densenet-161, with three different optimizers and loss functions MSE.
| SGD Optimizer | |||
|---|---|---|---|
| PC | MAE | RMSE | |
| K-fold-1 | 0.95 | 5.07 | 8.37 |
| K-fold-2 | 0.92 | 7.25 | 10.03 |
| K-fold-3 | 0.92 | 5.64 | 11.57 |
| K-fold-4 | 0.97 | 3.88 | 6.03 |
| K-fold-5 | 0.94 | 5.71 | 8.75 |
| Summary | 0.94 | 5.51 | 8.95 |
|
| |||
|
|
|
| |
| K-fold-1 | 0.96 | 4.87 | 8.10 |
| K-fold-2 | 0.95 | 5.54 | 8.34 |
| K-fold-3 | 0.92 | 6.33 | 12.16 |
| K-fold-4 | 0.97 | 3.75 | 5.69 |
| K-fold-5 | 0.95 | 5.22 | 8.05 |
| Summary | 0.95 | 5.14 | 8.47 |
|
| |||
|
|
|
| |
| K-fold-1 | 0.95 | 5.27 | 8.35 |
| K-fold-2 | 0.90 | 6.36 | 11.60 |
| K-fold-3 | 0.92 | 5.84 | 11.66 |
| K-fold-4 | 0.97 | 3.87 | 5.80 |
| K-fold-5 | 0.95 | 5.21 | 8.24 |
| Summary | 0.94 | 5.31 | 9.13 |
Figure 7Results of evaluation metric MAE for CNN Densenet-161 with optimizer RMSprop.
Figure 8Summary of evaluation metrics results PC, MAE and RMSE for CNN Densenet-161 with optimizer RMSprop (testing phase).
Figure 9Four instances of histogram dataset. Each histogram belongs to a different subject, and shows the percentage of COVID-19 infections of the different slices of lungs. In the figure on the left the histograms are represented, without normalization, by the number of slices of each patient, and on the right with the normalization applied.
Figure 10Confusion matrix of the classification in the testing phase using SVM for detecting the severity degree of the lungs of a patient (summary of 5 folds evaluated in testing phase).
Results of the classification in the testing phase using SVM for detecting the severity degree of the lungs of a patient (summary of 5 folds evaluated in the testing phase).
| SVM Results (%) | ||||
|---|---|---|---|---|
| Precision | Recall | F1-Score | #Patients | |
| Mild | 78 | 90 | 83 | 222 |
| Moderate | 85 | 71 | 77 | 267 |
| Severe | 83 | 89 | 86 | 107 |
| Accuracy | 81 | 596 | ||
Figure 11Data redistribution after applying LDA as a supervised dimensionality reduction method. (Green—Mild, Brown—Moderate and Purple—Severe).
Figure 12Data redistribution after applying LDA as a supervised dimensionality reduction method. (Green—Mild, Brown—Moderate and Purple—Severe).
Figure 13Confusion matrix of the classification in the testing phase using SVM and LDA, for detecting the severity degree of the lungs of a patient (summary of 5 folds evaluated in the testing phase).
Results of the classification in the testing phase using SVM and LDA for detecting the severity degree of the lungs of a patient (summary of 5 folds evaluated in the testing phase).
| SVM with LDA Results (%) | ||||
|---|---|---|---|---|
| Precision | Recall | F1-Score | #Patients | |
| Mild | 94 | 95 | 94 | 222 |
| Moderate | 95 | 93 | 94 | 267 |
| Severe | 96 | 97 | 97 | 107 |
| Accuracy | 95 | 596 | ||
Performance comparative of different methodologies presented in the bibliography to predict disease severity of patients with COVID-19.
| Author | Methods | Information Used | N° Patients (COVID+) | Results |
|---|---|---|---|---|
| Feng [ | Recurrent Neural Network. LesionEncoder framework | CT | 347 | Recall = 0.81; |
| Cai [ | Random Forest | CT/laboratory | 99 | AUC = 0.945 |
| Xiao [ | Convolutional Neural Network (ResNet34) | CT | 408 | AUC = 0.89; |
| Wu [ | Linear regression | CT/laboratory | 725 | Precision = (0.66–0.95); |
| Li [ | Convolutional Neural Network | CT/laboratory | 46 | Precision = 0.82; |
| Kang [ | Artificial Neural Network | CT/clinical/ | 151 | AUC = 0.95 |
| Ho [ | Convolutional Neural Network (ResNet50, Inception V3, DenseNet121 | CT | 297 | Precision = 0.78; |
| Weikert [ | Convolutional Neural Network. Multiple CT metrics | CT/clinical/ | 120 | CT metrics alone, |
| Fang [ | Deep Learning, SVM, LR and RF | CT/clinical/ | 193 (two data sets) | AUC = 0.813 (ICU) |
| Yan [ | Deep Learning, (U-Net and 3D Convolution) | CT/Expert interpretation | 221 | AUC = 0.88 (ICU) |
| Proposed methodology | DL (Densenet-161), SVM, LDA | X-ray/CT and novel histogram design | 596 | Precision = 0.95; |
Figure 14Details of the intersection between small and big data in medicine [48].