| Literature DB >> 34663874 |
Seong-Hwan Kim1, Eun-Tae Jeon1, Sungwook Yu2, Kyungmi Oh3, Chi Kyung Kim3, Tae-Jin Song4, Yong-Jae Kim5, Sung Hyuk Heo6, Kwang-Yeol Park7, Jeong-Min Kim7, Jong-Ho Park8, Jay Chol Choi9, Man-Seok Park10, Joon-Tae Kim10, Kang-Ho Choi11, Yang Ha Hwang12, Bum Joon Kim13, Jong-Won Chung14, Oh Young Bang14, Gyeongmoon Kim14, Woo-Keun Seo15, Jin-Man Jung16,17.
Abstract
We aimed to develop a novel prediction model for early neurological deterioration (END) based on an interpretable machine learning (ML) algorithm for atrial fibrillation (AF)-related stroke and to evaluate the prediction accuracy and feature importance of ML models. Data from multicenter prospective stroke registries in South Korea were collected. After stepwise data preprocessing, we utilized logistic regression, support vector machine, extreme gradient boosting, light gradient boosting machine (LightGBM), and multilayer perceptron models. We used the Shapley additive explanation (SHAP) method to evaluate feature importance. Of the 3,213 stroke patients, the 2,363 who had arrived at the hospital within 24 h of symptom onset and had available information regarding END were included. Of these, 318 (13.5%) had END. The LightGBM model showed the highest area under the receiver operating characteristic curve (0.772; 95% confidence interval, 0.715-0.829). The feature importance analysis revealed that fasting glucose level and the National Institute of Health Stroke Scale score were the most influential factors. Among ML algorithms, the LightGBM model was particularly useful for predicting END, as it revealed new and diverse predictors. Additionally, the effects of the features on the predictive power of the model were individualized using the SHAP method.Entities:
Mesh:
Year: 2021 PMID: 34663874 PMCID: PMC8523653 DOI: 10.1038/s41598-021-99920-7
Source DB: PubMed Journal: Sci Rep ISSN: 2045-2322 Impact factor: 4.379
Figure 1Flowchart of included patients.
Comparison of model performance.
| Model | AUROC [95% CI] | AUPRC [95% CI] | Brier score | ACC (%) | Precision | Recall | F1 score | p value† |
|---|---|---|---|---|---|---|---|---|
| Logistic regression | 0.696 [0.636–0.755] | 0.288 [0.207–0.368] | 0.110 | 86.5 | 0.253 | 0.585 | 0.353 | |
| SVM | 0.722 [0.667–0.777] | 0.261[0.168–0.356] | 0.112 | 86.2 | 0.254 | 0.695 | 0.373 | 0.182 |
| XGBoost | 0.759 [0.700–0.817] | 0.367 [0.260–0.466] | 0.105 | 86.5 | 0.349 | 0.537 | 0.423 | 0.024 |
| LightGBM | 0.772 [0.715–0.829] | 0.385 [0.273–0.497] | 0.103 | 86.7 | 0.328 | 0.695 | 0.445 | 0.003* |
| MLP | 0.768 [0.714–0.822] | 0.374 [0.265–0.482] | 0.103 | 86.9 | 0.432 | 0.463 | 0.447 | 0.002* |
*Significant difference at p < 0.005.
†Comparison with logistic regression on AUROC.
Abbreviations: AUROC, area under the receiver operating characteristic curve; AUPRC, area under the precision-recall curve; SVM, support vector machine; XGBoost, extreme gradient boosting; LightGBM, light gradient boosting machine; MLP, multilayer perceptron.
Figure 2Model performance. (A), Solid lines and shades represent receiver operating characteristics curves and its 95% confidence intervals. An asterisk (*) indicates significant difference (P < 0.005) in comparison with logistic regression. (B), Solid lines and shades represent precision-recall curves and its 95% confidence intervals. Only the confidence intervals of the baseline model (logistic regression, “LogReg”) are represented with polka dot pattern in both plots. (C), Detailed performance analysis for the best model (LightGBM) in different discrimination thresholds. Solid lines and shades represent mean values and 95% confidence intervals in each variable. Abbreviations: AUC, area under the curve; CI, confidence interval; LogReg, Logistic regression; SVM, Support vector machine; XGBoost, Extreme gradient boosting; LightGBM, light gradient boosting machine; MLP, Multilayer perceptron.
Figure 3Matrix plots of top 23 important features. Bar plot (A) and violin plot (B). In the bar plot, the SHAP value implies the degree of contribution of a specific feature. The higher the SHAP value, the larger the model contribution of a specific feature. In the violin plot, each dot represents one patient and accumulates vertically to depict the density. The color reflects the high and low values of each feature, with the red color indicating a higher value and the blue color indicating a lower value. The X-axis of the graph represents the SHAP value, and a positive SHAP value indicates that it contributes positively to predicting the model, and that the probability of END occurring is high, and vice versa. Abbreviations: NIHSS, National Institute of Health Stroke Scale; mRS, modified Rankin scale; ALP, alkaline phosphatase; SVS, susceptibility vessel sign; ICAS, intracranial atherosclerosis; aPTT, activated partial thromboplastin time; FDP, fibrin degradation product; LA, left atrium; DBP, diastolic blood pressure; AST, aspartate aminotransferase; Hct, hematocrit; LDL, low-density lipoprotein.
Figure 4Partial SHAP dependence plot of the four representative features. Values are plotted with a scatter plot and a regression line represented with the orange line of mean and shade of SD. A red diamond represents a cut-off value of the variable. Histograms on the right and top of each plot are distributions of the SHAP and values of variables. Abbreviations: NIHSS, National Institute of Health Stroke Scale; LA, left atrium; LDL, low- density lipoprotein.