| Literature DB >> 36116079 |
Serpil Ustebay1, Abdurrahman Sarmis2, Gulsum Kubra Kaya3,4, Mark Sujan5.
Abstract
ML algorithms are used to develop prognostic and diagnostic models and so to support clinical decision-making. This study uses eight supervised ML algorithms to predict the need for intensive care, intubation, and mortality risk for COVID-19 patients. The study uses two datasets: (1) patient demographics and clinical data (n = 11,712), and (2) patient demographics, clinical data, and blood test results (n = 602) for developing the prediction models, understanding the most significant features, and comparing the performances of eight different ML algorithms. Experimental findings showed that all prognostic prediction models reported an AUROC value of over 0.92, in which extra tree and CatBoost classifiers were often outperformed (AUROC over 0.94). The findings revealed that the features of C-reactive protein, the ratio of lymphocytes, lactic acid, and serum calcium have a substantial impact on COVID-19 prognostic predictions. This study provides evidence of the value of tree-based supervised ML algorithms for predicting prognosis in health care.Entities:
Keywords: COVID-19; Infectious diseases; Machine learning; Prognostic predictions; Risk factors
Year: 2022 PMID: 36116079 PMCID: PMC9483274 DOI: 10.1007/s11739-022-03101-x
Source DB: PubMed Journal: Intern Emerg Med ISSN: 1828-0447 Impact factor: 5.472
Fig. 1Datasets used to develop the prediction models
Comparison of algorithm performances
| ML algorithms | Model 1 | Model 2 | Model 3 | Model 4 | Model 5 | Model 6 |
|---|---|---|---|---|---|---|
| Extreme gradient boosting | 0.959 | 0.944 | 0.975 | 0.940 | 0.978 | |
| CatBoost classifier | 0.924 | 0.977 | 0.945 | 0.984 | ||
| Extra Trees classifier | 0.904 | 0.961 | 0.939 | |||
| Random forest classifier | 0.864 | 0.953 | 0.908 | 0.945 | 0.925 | 0.978 |
| MLP classifier | 0.859 | 0.959 | 0.922 | 0.977 | 0.930 | 0.983 |
| Logistic regression | 0.811 | 0.957 | 0.878 | 0.864 | 0.942 | 0.949 |
| Support vector machine- linear kernel | 0.808 | 0.956 | 0.874 | 0.876 | 0.932 | 0.943 |
| 0.848 | 0.945 | 0.890 | 0.904 | 0.912 | 0.932 |
The bold numbers are the AUROC values that received the highest score in each model.
Model performance for the best-performed algorithm
| Prediction model | Best algorithm | Model performance | |||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|
| Accuracy | AUROC | Sensitivity | Specificity | PPV | NPV | F1 | PLR | NLR | Scaled brier | ||
| Model 1 | XGBoost | 0.75 | 0.90 (0.89, 0.91) | 0.899 | 0.657 | 0.621 | 0.913 | 0.74 | 2.63 | 0.15 | 0.297 |
| Model 2 | CatBoost | 0.915 | 0.96 (0.95, 0.97) | 0.918 | 0.913 | 0.754 | 0.975 | 0.83 | 10.63 | 0.09 | 0.609 |
| Model 3 | CatBoost | 0.86 | 0.946 (0.94, 0.95) | 0.886 | 0.857 | 0.662 | 0.959 | 0.76 | 6.22 | 0.13 | 0.462 |
| Model 4 | ET | 0.95 | 0.987 (0.98, 1) | 0.965 | 0.946 | 0.944 | 0.967 | 0.95 | 17.96 | 0.04 | 0.754 |
| Model 5 | ET | 0.92 | 0.973 (0.95, 1) | 0.881 | 0.943 | 0.881 | 0.942 | 0.88 | 15.36 | 0.13 | 0.696 |
| Model 6 | ET | 0.95 | 0.993 (0.99, 1) | 0.945 | 0.952 | 0.896 | 0.976 | 0.92 | 19.85 | 0.06 | 0.75 |
Fig. 2Feature importance for predicting models. A Feature importance in Model 1 (the need for intensive care), B Feature importance in Model 2 (the need for intubation), and C Feature importance in Model 3 (the risk of mortality)
Fig. 3Feature importance for predicting models. D Feature importance in Model 4 (the need for intensive care), E Feature importance in Model 5 (the need for intubation), and F Feature importance in Model 6 (the risk of mortality)