| Literature DB >> 34230837 |
Abstract
Due to the highly infectious nature of the novel coronavirus (COVID-19) disease, excessive number of patients waits in the line for chest X-ray examination, which overloads the clinicians and radiologists and negatively affects the patient's treatment, prognosis and control of the pandemic. Now that the clinical facilities such as the intensive care units and the mechanical ventilators are very limited in the face of this highly contagious disease, it becomes quite important to classify the patients according to their severity levels. This paper presents a novel implementation of convolutional neural network (CNN) approach for COVID-19 disease severity classification (assessment). An automated CNN model is designed and proposed to divide COVID-19 patients into four severity classes as mild, moderate, severe, and critical with an average accuracy of 95.52% using chest X-ray images as input. Experimental results on a sufficiently large number of chest X-ray images demonstrate the effectiveness of CNN model produced with the proposed framework. To the best of the author's knowledge, this is the first COVID-19 disease severity assessment study with four stages (mild vs. moderate vs. severe vs. critical) using a sufficiently large number of X-ray images dataset and CNN whose almost all hyper-parameters are automatically tuned by the grid search optimiser.Entities:
Year: 2021 PMID: 34230837 PMCID: PMC8251482 DOI: 10.1049/ipr2.12153
Source DB: PubMed Journal: IET Image Process ISSN: 1751-9659 Impact factor: 2.373
FIGURE 1Some sample X‐ray images of COVID‐19 patients with relevant labels and disease severity scores on them. Severity scores are obtained by summing up both lung scores computed based on ground‐glass opacity and lung involvement
Learning scheme of the proposed convolutional neural network (CNN) model
|
| |||||
|---|---|---|---|---|---|
|
|
|
|
|
|
|
|
| 1000 | 3260 | 1956 | 652 | 652 |
|
| 950 | ||||
|
| 600 | ||||
|
| 710 | ||||
FIGURE 2Architecture of the proposed convolutional neural network (CNN) model
The proposed CNN architecture details
|
|
|
|
|
| |
|---|---|---|---|---|---|
| 1 | 227 × 227 × 3 input layer | Input | 227 × 227 × 3 |
| 0 |
| 2 | 128 6 × 6 × 3 convolutions with stride [4 4] and padding [0 0 0 0] | Convolutional | 56 × 56 × 128 |
Weights: 6 × 6 × 3 × 128 Bias: 1 × 1 × 128 | 13,952 |
| 3 | ReLU‐1 | ReLU | 56 × 56 × 128 |
| 0 |
| 4 | Cross channel normalisation | Normalisation | 56 × 56 × 128 |
| 0 |
| 5 | 2 × 2 max pooling with stride [2 2] and padding [0 0 0 0] | Max pooling | 28 × 28 × 128 |
| 0 |
| 6 | 96 6 × 6 × 128 convolutions with stride [1 1] and padding [2 2 2 2] | Convolutional | 27 × 27 × 96 |
Weights: 6 × 6 × 128 × 96 Bias: 1 × 1 × 96 | 46,752 |
| 7 | ReLU‐2 | ReLU | 27 × 27 × 96 |
| 0 |
| 8 | 2 × 2 max pooling with stride [2 2] and padding [0 0 0 0] | Max pooling | 13 × 13 × 96 |
| 0 |
| 9 | 96 2 × 2 × 96 convolutions with stride [1 1] and padding [2 2 2 2] | Convolutional | 16 × 16 × 96 |
Weights: 2 × 2 × 96 × 96 Bias: 1 × 1 × 96 | 36,864 |
| 10 | ReLU‐3 | ReLU | 16 × 16 × 96 |
| 0 |
| 11 | 2 × 2 max pooling with stride [2 2] and padding [0 0 0 0] | Max pooling | 8 × 8 × 96 |
| 0 |
| 12 | 512 fully connected (FC) layer | FC | 1 × 1 × 512 |
Weights: 512 × 6144 Bias: 512 × 1 | 3,146,240 |
| 13 | 30% dropout | Dropout | 1 × 1 × 512 |
| 0 |
| 14 | Four FC layer | FC | 1 × 1 x 4 |
Weights: 4 × 512 Bias: 4 × 1 | 2052 |
| 15 | Softmax | Softmax | 1 × 1 × 4 |
| 0 |
|
|
|
| — | — |
|
Optimum hyper‐parameters results achieved by grid search
|
|
|
|
|---|---|---|
| Number of convolution and max pooling layers | [ | 3 |
| Number of FC layers | [ | 2 |
| Number of filters | [ | 128, 96, 96 |
| Filter size | [ | 6, 6, 2 |
| Activation function | [ELU, SELU, ReLU, Leaky ReLU] | ReLU |
| Mini‐batch size | [ | 64 |
| Momentum | [0.80, 0.85, 0.9, 0.95] | 0.9 |
| Learning rate | [0.0001, 0.0005, 0.001, 0.005] | 0.0001 |
| ℓ2regularisation | [0.0001, 0.0005, 0.001, 0.005] | 0.001 |
FIGURE 3Accuracy and loss curves
FIGURE 4Receiver operating characteristic curve
FIGURE 5Confusion matrix
Accuracy metrics in terms of true positive (TP), true negative (TN), false positive (FP), false negative (FN), accuracy, specificity, sensitivity and precision
|
| ||||||||||
|---|---|---|---|---|---|---|---|---|---|---|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 186 | 442 | 10 | 14 | 96.32% | 0.978 |
| 0.949 | 200 |
|
| 181 | 448 | 14 | 9 | 96.47% | 0.970 | 0.953 | 0.928 | 190 | |
|
| 139 | 507 | 3 | 3 |
| 0.994 |
|
| 142 | |
|
| 116 | 529 | 3 | 4 | 98.93% |
| 0.967 | 0.967 | 120 |
FIGURE 6Classification results and the predicted probabilities of four test images
Average classification performance and accuracy metrics for each fold
|
|
|
|
|
|
| ||
|---|---|---|---|---|---|---|---|
| Performance Metrics (%) |
| 95.72 | 96.88 | 94.86 | 97.14 | 95.51 | 96.02 |
|
| 98.40 | 99.14 | 98.86 | 98.03 | 97.71 | 98.43 | |
|
| 95.58 | 99.84 | 98.85 | 98.03 | 97.72 | 98.00 | |
|
| 95.40 | 96.01 | 94.79 | 96.93 | 94.63 |
| |
|
| 0.9976 | 0.9917 | 0.9788 | 0.9945 | 0.9738 |
|
Performance metrics of disease severity classification results under different network models
|
|
|
|
|
| |
|---|---|---|---|---|---|
| Accuracy (%) | 79.87 | 81.85 | 75.76 | 88.34 |
|
| AUC | 0.8109 | 0.8911 | 0.749 | 0.8742 |
|
| Elapsed time (s) | 1078 | 875 | 1287 | 1707 |
|
|
|
|
| 2. | Initialise CML, FCL, NF, FS, AF with default values |
|
|
|
| 4. |
|
| 5. |
|
| 6. |
|
| 7. |
|
| 8. | model = CNN_train (train, CML, FCL, NF, FS, AF) |
| 9. | score = CNN_predict (test, model) |
| 10. | cv_list.insert (score) |
| 11. | scores_list.insert (mean(cv_list), CML, FCL, NF, FS, AF) |
|
|
|
CML: Number of convolutional and max pooling layers, FCL: Number of FC layers, NF: Number of filters, FS: Filter sizes, AF: Activation function
| 1. |
|
| 2. | Initialise L2R, M, MBS, LR with default values |
| 3. |
|
| 4. |
|
| 5. |
|
| 6. |
|
| 7. | model = CNN_train (train, L2R, M, MBS, LR) |
| 8. | score = CNN_predict (test, model) |
| 9. | cv_list.insert (score) |
| 10. | scores_list.insert (mean(cv_list), L2R, M, MBS, LR) |
| 11. |
|
L2R: ℓ2 regularisation, M: Momentum, MBS: Mini‐batch size, LR: Learning rate