| Literature DB >> 33513487 |
Mehmet Kivrak1, Emek Guldogan1, Cemil Colak2.
Abstract
BACKGROUND ANDEntities:
Keywords: Data Mining; Deep Learning; Extreme Gradient Boosting; Machine Learning; SARS-COV-2
Mesh:
Year: 2021 PMID: 33513487 PMCID: PMC7826038 DOI: 10.1016/j.cmpb.2021.105951
Source DB: PubMed Journal: Comput Methods Programs Biomed ISSN: 0169-2607 Impact factor: 5.428
The detailed explanation of the variables/attributes in the dataset.
| Age | Birth year (year) | Input |
| Gender | Gender (0=female, 1=male) | Input |
| Diabetes | Diabetes (1= presence, 0= absence) | Input |
| Hypertension | Hypertension (1= presence, 0= absence) | Input |
| COPD | Chronic Obstructive Pulmonary Diseases (bronchitis, pneumonia, asthma, and emphysema) | Input |
| Cancer | Cancer Diseases (1= presence, 0= absence) | Input |
| Renal | Renal Diseases (1= presence, 0= absence) | Input |
| ACE | Angiotensin-Converting Enzyme (ATC classes: C09A and C09B) | Input |
| ARBs | Angiotensin II Receptor Blockers (C09C and C09D) | Input |
| CVD | Cardiovascular disorders (heart failure, myocardial infarction, and stroke-CVD) | Input |
| COVID Severity | SARS-COV-2 Severity (0=mild, 1= severe, 2= very severe) | Input |
| Death Status (DS) | (0=alive, 1=dead) | Output |
The group comparisons of the variables.
| 50.4 ± 20.2 / | 66.4 ± 16.9 / | 76.2 ± 12.9 / | 355.1 | <0.001* | |
| 65a (34%) | 75b (38%) | 54c (28%) | 79.98 | <0.001** | |
| 207a (38%) | 207a (38%) | 129b (24%) | 186.9 | <0.001** | |
| 28a (29%) | 42b (43%) | 27a,c (28%) | 46.27 | <0.001** | |
| 49a (40%) | 46b (38%) | 27b (22%) | 23.9 | <0.001** | |
| 66a (26%) | 122b (47%) | 70a,c (27%) | 157.9 | <0.001** | |
| 23a (27%) | 40b (46%) | 23a,c (27%) | 43.67 | <0.001** | |
| 107a (43%) | 88b (35%) | 56c (22%) | 45.79 | <0.001** | |
| 86a (38%) | 90b (39%) | 52c (23%) | 59.18 | <0.001** | |
| 550a (65%) | 213b (25%) | 82b (10%) | 22.523 | <0.001 | |
| 407a (54%) | 241b (32%) | 110b (14%) | |||
| 957 (60%) | 454 (28%) | 192 (12%) | |||
The data are summarized as X̄±SD or median (min-max) and Count (Percent). Different superscripts in each row imply a significant difference between categories (Conover or Bonferroni-corrected Pearson chi-square tests for pairwise comparisons; p<0.05); *: Kruskal Wallis H test; **: Pearson chi-square test.
The baseline characteristics of the sample.
| 1603 | 957 | 454 | 192 | |
| 58.0±20.9 | 50.4± 20.2 | 66.4±16.9 | 76.2±12.9 | |
| 758 (47.3) | 407 (53.7) | 241 (31.8) | 110 (14.5) | |
| 194 (12.1) | 65 (34) | 75 (37.7) | 54 (28.3) | |
| 97 (6.0) | 28 (28.9) | 42(43.3) | 27 (27.8) | |
| 122 (7.6) | 49 (40.2) | 46 (37.7) | 27 (22.1) | |
| 258 (16.1) | 66 (25.6) | 122 (47.3) | 70 (27.1) | |
| 86 (5.4) | 23 (26.7) | 40 (46.6) | 23 (26.7) | |
| 543 (33.9) | 207 (38.1) | 207 (38.1) | 129 (23.6) | |
| 251 (15.7) | 107 (42.6) | 88 (35.1) | 56 (22.3) | |
| 228 (14.2) | 86 (37.7) | 90 (39.5) | 52 (22.8) |
Fig. 3The characteristics of the sample.
Fig. 4The pseudo-codes of the XGBoost algorithm.
Variable importance values for the deep learning and the XGBoost algorithms.
| 1.00 | 10.3 | 1.00 | 89.9 | |
| 0.98 | 10.1 | 0.004 | 0.35 | |
| 0.95 | 9.8 | 0.000 | 0.00 | |
| 0.92 | 9.5 | 0.002 | 0.15 | |
| 0.89 | 9.2 | 0.001 | 0.11 | |
| 0.89 | 9.1 | 0.004 | 0.4 | |
| 0.85 | 8.8 | 0.003 | 0.3 | |
| 0.84 | 8.7 | 0.000 | 0.00 | |
| 0.83 | 8.5 | 0.095 | 8.6 | |
| 0.81 | 8.4 | 0.000 | 0.03 | |
| 0.77 | 7.6 | 0.001 | 0.11 | |
Confusion matrix for the techniques.
| alive | 1011 | 85 | 92.24% |
| dead | 3 | 23 | 88.46% |
| 99.70% | 21.30% | ||
| alive | 999 | 17 | 98.33% |
| dead | 15 | 91 | 85.85% |
| 98.52% | 84.26% | ||
| alive | 991 | 51 | 95.11% |
| dead | 23 | 57 | 71.25% |
| 97.73% | 52.78% | ||
| alive | 1011 | 0 | 100 |
| dead | 3 | 108 | 97.3 |
| 99.7 | 100 |
Fig. 5The graphical representation of the confusion matrix for the models.
Performance metrics of the models.
| Model | ||||||
|---|---|---|---|---|---|---|
| 97.15 | 98.5 | 92.2 | 88.5 | 2.85 | 0.82 | |
| 92.15 | 99.7 | 98.3 | 85.9 | 7.85 | 0.19 | |
| 93.4 | 97.7 | 95.1 | 71.2 | 6.6 | 0.58 | |
| 99.7 | 99.7 | 99.7 | 1.00 | 0.03 | 0.91 |