| Literature DB >> 32814568 |
Chenglong Liu1,2, Xiaoyang Wang3, Chenbin Liu4, Qingfeng Sun5, Wenxian Peng6.
Abstract
BACKGROUND: Chest CT screening as supplementary means is crucial in diagnosing novel coronavirus pneumonia (COVID-19) with high sensitivity and popularity. Machine learning was adept in discovering intricate structures from CT images and achieved expert-level performance in medical image analysis.Entities:
Keywords: Chest CT; General pneumonia; Machine learning; Novel coronavirus pneumonia
Mesh:
Year: 2020 PMID: 32814568 PMCID: PMC7436068 DOI: 10.1186/s12938-020-00809-9
Source DB: PubMed Journal: Biomed Eng Online ISSN: 1475-925X Impact factor: 2.819
Fig. 1Samples of COVID-19 and GP CT images. Picture a is the CT image of COVID-19 with bilateral GGOs while picture b is the CT image of GP with unilateral GGO. The red arrows point at the GGOs of COVID-19 and the blue arrow points at the GGO of GP
Relevance of each feature based on ReliefF algorithm
| Feature | Relevance | Feature | Relevance | Feature | Relevance | Feature | Relevance | Feature | Relevance |
|---|---|---|---|---|---|---|---|---|---|
| 1 | 0.0421 | 8 | 0.1329 | 15 | 0.0043 | 22 | 0.1348 | 29 | 0.1250 |
| 2 | 0.1036 | 9 | 0.1338 | 16 | 0.0687 | 23 | 0.1257 | 30 | 0.1025 |
| 3 | 0.0050 | 10 | 0.1059 | 17 | 0.0548 | 24 | 0.1170 | 31 | 0.1490 |
| 4 | 0.0178 | 11 | 0.0582 | 18 | 0.1604 | 25 | 0.1574 | 32 | 0.1250 |
| 5 | 0.0032 | 12 | 0.0360 | 19 | 0.1574 | 26 | 0.2267 | 33 | 0.0389 |
| 6 | 0.1434 | 13 | 0.0163 | 20 | 0.2561 | 27 | 0.2977 | 34 | 0.0963 |
| 7 | 0.1036 | 14 | 0.1184 | 21 | 0.1469 | 28 | 0.2094 |
Fig. 2The weight curves of 34 features based on ReliefF algorithm. The X-axis represents the numbers of features. The Y axis represents the weights of different features at different times. The algorithm run 1000 times represented by curves with different colors. The dark straight line represents weight = 0.11, which is the proposed threshold T
Selected features of four combinations
| Combination | Feature numbers | |
|---|---|---|
| 1 | 0.11a | 1,2,3,4,5,7,10,11,12,13,15,16,17,30,33,34 |
| 2 | 0b | All 34 features |
| 3 | 0.10 | 2,6,7,8,9,10,14,18,19,20,21,22,23,24,25,26,27,28,29,30,31,32 |
| 4 | 0.12 | 6,8,9,18,19,20,21,22,23,25,26,27,28,29,31,32 |
| 5 | 0.11 | 6,8,9,14,18,19,20,21,22,23,24,25,26,27,28,29,31,32 |
aSelect features with relevance smaller than the threshold T
bNo feature selection was applied
Description of extracted features
| Feature groups | Description |
|---|---|
| GLGCM | 1. Little gradient advantage; 2. Large gradient advantage; 3. Gray heterogeneity; 4. Gradient heterogeneity; 5. Energy; 6. Average gray; 7. Average gradient; 8. Gray mean square error; 9. Gradient mean square error; 10. Correlation; 11. Gray entropy; 12. Gradient entropy; 13. Hybrid entropy; 14. Inertia; 15. Inverse difference moment |
| GLCM | 16. Angular second moment; 17. Correlation; 18. Entropy; 19. Contrast; 20. Inverse difference moment; 21. Sum average; 22. Sum entropy; 23. Sum variance; 24. Variance; 25. Dissimilarity; 26. Inertia; 27. Difference variance; 28. Difference entropy |
| Histogram | 29. Entropy; 30. Uniformity; 31. Mean intensity; 32. Standard deviation; 33. Kurtosis; 34. Skewness |
Diagnosis performance based on different methods using different combinations
| Method | Combination 1 | Combination 2 | Combination 3 | Combination 4 | Combination 5 |
|---|---|---|---|---|---|
| DT | |||||
| ACC (%) | 81.98 | 88.82 | 89.41 | 89.73 | 89.82 |
| SEN (%) | 72.03 | 85.85 | 84.34 | 84.39 | 85.90 |
| SPE (%) | 80.40 | 91.95 | 95.64 | 95.55 | 95.69 |
| LR | |||||
| ACC (%) | 70.64 | 80.57 | 81.65 | 79.15 | 81.73 |
| SEN (%) | 66.50 | 76.42 | 77.89 | 75.61 | 77.89 |
| SPE (%) | 75.00 | 84.93 | 85.62 | 82.88 | 85.73 |
| SVM | |||||
| ACC (%) | 81.90 | 85.32 | 85.65 | 86.07 | 86.48 |
| SEN (%) | 76.91 | 83.58 | 82.11 | 81.30 | 83.79 |
| SPE (%) | 87.16 | 87.16 | 89.38 | 91.10 | 91.44 |
| KNN | |||||
| ACC (%) | 77.73 | 83.23 | 85.40 | 88.32 | 88.24 |
| SEN (%) | 69.43 | 73.98 | 79.51 | 81.79 | 82.33 |
| SPE (%) | 86.47 | 92.97 | 91.61 | 95.21 | 96.58 |
| EBT | |||||
| ACC (%) | 86.49 | 92.91 | 92.49 | 93.41 | 94.16 |
| SEN (%) | 78.37 | 86.18 | 85.69 | 87.32 | 88.62 |
| SPE (%) | 95.03 | 100.00 | 99.66 | 99.83 | 100.00 |
Fig. 3Accuracy comparison of five classifiers with different feature combinations
Fig. 4Sensitivity comparison of five classifiers with different feature combinations
Fig. 5Specificity comparison of five classifiers with different feature combinations
Fig. 6Comparison of receiver operating characteristic curves for the proposed classifier, KNN, SVM, LR, and DT using feature combination 5. The receiver operating characteristic curves for the proposed EBT models had an AUC that was significantly greater than that for four other models
Fig. 7The flowchart of the proposed diagnosis framework