| Literature DB >> 33624945 |
Xinghua Xu1, Jiashu Zhang1, Kai Yang2, Qun Wang1, Xiaolei Chen1, Bainan Xu1.
Abstract
OBJECTIVES: Spontaneous intracerebral hemorrhage remains a major cause of death and disability throughout the world. We tried to establish accurate long-term outcome prediction models for hypertensive intracerebral hemorrhage (HICH) using CT radiomics and machine learning.Entities:
Keywords: artificial intelligence; intracerebral hemorrhage; machine learning; prognosis; radiomics
Year: 2021 PMID: 33624945 PMCID: PMC8119849 DOI: 10.1002/brb3.2085
Source DB: PubMed Journal: Brain Behav Impact factor: 2.708
FIGURE 1The image postprocessing workflow. First, hematoma on CT images was segmented. After feature extraction, feature selection, and machine learning model construction, six prognosis‐predictive models were established in the training set and were further evaluated in the validation sets
FIGURE 2Figure selection using variance threshold analysis and SelectKBest analysis. Of the 1,029 features extracted, 525 features were selected after variance threshold analysis (a) and 182 were selected after SelectKBest analysis (b)
FIGURE 3Features finally selected for prediction after the LASSO regression analysis
FIGURE 4Principal component screening and feature correlation analysis. Principal component analysis revealed the 4 features that had the greatest impact on prognosis (a). Covariance analysis showed the interdependence and relevance between the selected 18 features (b)
Sensitivity, specificity, and overall accuracy of different machine learning algorithm prediction models for training set and validation set
| Algorithms | Training set ( | Validation set ( | ||||
|---|---|---|---|---|---|---|
| Sensitivity | Specificity | Accuracy | Sensitivity | Specificity | Accuracy | |
| KNN | 86.8% | 88.3% | 87.9% | 90.0% | 82.2% | 83.6% |
| SVM | 81.8% | 87.5% | 86.0% | 90.9% | 84.1% | 85.5% |
| XGBoost | 96.0% | 89.7% | 91.2% | 92.3% | 88.1% | 89.1% |
| RF | 98.5% | 99.3% | 99.1% | 93.3% | 92.5% | 92.7% |
| LR | 75.9% | 86.6% | 83.7% | 90.9% | 84.1% | 85.5% |
| DT | 100% | 100% | 100% | 80.0% | 87.5% | 85.5% |
Abbreviations: DT, decision tree; KNN, k‐nearest neighbor; LR, logistic regression; RF, random forest; SVM, support vector machine; XGBoost, extreme gradient boosting.
FIGURE 5ROC curves of the RF model and the XGBoost model for long‐term outcome prediction of HICH in validation sets. The AUC was 0.92 for both the RF model (a) and the XGBoost model (b)