| Literature DB >> 35885568 |
Yingwei Guo1,2, Yingjian Yang1,2, Fengqiu Cao1, Wei Li2, Mingming Wang3, Yu Luo3, Jia Guo4, Asim Zaman2,5, Xueqiang Zeng2,5, Xiaoqiang Miu1,2, Longyu Li2, Weiyan Qiu2, Yan Kang1,2,5,6.
Abstract
BACKGROUND: Accurate outcome prediction is of great clinical significance in customizing personalized treatment plans, reducing the situation of poor recovery, and objectively and accurately evaluating the treatment effect. This study intended to evaluate the performance of clinical text information (CTI), radiomics features, and survival features (SurvF) for predicting functional outcomes of patients with ischemic stroke.Entities:
Keywords: clinical text information; ischemic stroke outcome; machine learning; radiomics features; survival features
Year: 2022 PMID: 35885568 PMCID: PMC9324145 DOI: 10.3390/diagnostics12071664
Source DB: PubMed Journal: Diagnostics (Basel) ISSN: 2075-4418
Patient information and scanning parameters of DSC-PWI datasets.
| Patient Information | Scanning Parameters of DSC-PWI Images | ||
|---|---|---|---|
| Numbers of patients | 56 | TE/TR | 32/1590 ms |
| Datasets (sets) | 80 | Matrix | 256 × 256 |
| Numbers of Female (%) | 15 (26.79%) | FOV | 230 × 230 mm2 |
| Age (Mean ± Std) | 71.362 ± 10.91 | Thickness | 5 mm |
| Volumes of lesions (Mean ± Std, mL) | 95.583 ± 72.304 | Number of measurements | 50 |
| Income NHISS (Mean ± Std) | 9.919 ± 6.747 | Spacing between slices | 6.5 mm |
| Outcome NHISS (Mean ± Std) | 6.275 ± 6.875 | Pixel bandwidth | 1347 Hz/pixel |
| Right limbs weakness (%) | 38 (47.5%) | Number of slices | 20 |
| Left limbs weakness (%) | 36 (45%) | ||
| Lisp out (%) | 59 (73.75%) | ||
| Confused (%) | 10 (12.5%) | ||
| Hypertension (%) | 59 (73.75%) | ||
| Diabetes (%) | 26 (32.5%) | ||
| Atrial fibrillation (%) | 28 (35%) | ||
| Intra-arterial thrombectomy (%) | 22 (27.5%) | ||
| 90-day mRS | 2.525 ± 2.326 | ||
Figure 1Flowchart of this study. (a) Process of the preprocessing of DSC-PWI datasets and making ROIs; (b) Calculating radiomics features, where the value of the feature is represented by color; (c–e) The process of feature selection, feature fusion, and stroke outcome prediction.
Figure 2Flowchart of selecting significant features. and are the ith feature in LT and NT groups, respectively.
Descriptions of the 13 feature selection methods in this study.
| Type | Method | Description | Equation |
|---|---|---|---|
| FITI | MIM | Evaluates features by correlation between features and classes measured by mutual information |
|
| MIFS/MRMR | Evaluates features by correlation between features and classes and redundancy among features |
| |
| JMI/CMIM | Evaluates features by correlation between features and classes and redundancy among features measured by conditional mutual information |
| |
| SIF | Fisher/LS | Compares features with their ratios of variance between classes and variance within classes | |
| ReliefF | Compares features with correlation between features and classes computed from ability of features to distinguish between close samples |
| |
| STF | FS | Obtains feature score with ability to distinguish positive classes and negative classes computed by average of both classes |
|
| TS | Computes feature score with average and variance of features |
| |
| SSL | MCFS | Combines cluster with feature coefficients of combinatorial classes to compute feature score |
|
| Alpha | Evaluates features by dynamically adjusting threshold on error reduction to obtain selection results |
| |
| Lasso | Uses L1 regularization to make weight of some learned features equal 0, to achieve purpose of sparse and feature selection |
|
Descriptions of the 10 models in this study.
| No. | Model | Definition in Python 3.6 |
|---|---|---|
| 1 | SVM | sklearn.svm.SVC(kernel = ‘rbf’,probability = True) |
| 2 | DT | sklearn.tree. DecisionTreeClassifier() |
| 3 | Ada | sklearn.ensemble.AdaBoostClassifier() |
| 4 | NN | sklearn.neural_network. MLPClassifier (hidden_layer_sizes = (400, 100), alpha = 0.01, max_iter = 10000) |
| 5 | RF | sklearn.ensemble.RandomForestClassifier(n_estimators = 200) |
| 6 | KNN | sklearn.neighbors. sklearn.neighbors() |
| 7 | LR | sklearn.linear_model.logisticRegressionCV(max_iter = 100,000, solver = “liblinear”) |
| 8 | DA | sklearn.discriminant_analysis.() |
| 9 | GBDT | sklearn.ensemble.GradientBoostingClassifier() |
| 10 | NB | sklearn.naive_bayes. GaussianNB() |
Figure 3Flowchart of multidimensional feature fusion.
Distributions of 90-day mRS in the three situations.
| 90-Day mRS | 0 | 1 | 2 | 3 | 4 | 5 | 6 |
|---|---|---|---|---|---|---|---|
| 7-category counts | 25 | 11 | 9 | 4 | 8 | 9 | 14 |
| 4-category counts | 25 | 20 | 12 | 23 | - | - | - |
| 2-category counts | 45 | 35 | - | - | - | - |
Figure 4Significant features in the radiomics features group. (a,b) Ratio and p-value of significant features in the eight radiomics feature groups, respectively.
Statistics of p-values of significant features in each radiomics group.
| Groups | Numbers | Mean | Std | Min | Medium | Max |
|---|---|---|---|---|---|---|
| First-order | 555 | 0.005 | 0.0105 | <0.0001 | <0.0001 | 0.0497 |
| GLDM | 419 | 0.006 | 0.0104 | <0.0001 | <0.0001 | 0.0497 |
| GLCM | 619 | 0.006 | 0.0114 | <0.0001 | <0.0001 | 0.0499 |
| GLRLM | 436 | 0.0068 | 0.0118 | <0.0001 | <0.0001 | 0.0498 |
| GLSZM | 526 | 0.0066 | 0.0104 | <0.0001 | <0.0001 | 0.0496 |
| Log-Sigma | 5551 | 0.01 | 0.0138 | <0.0001 | 0.0027 | 0.05 |
| NGTDM | 139 | 0.009 | 0.0107 | <0.0001 | 0.0063 | 0.0449 |
| Wavelet | 11612 | 0.0091 | 0.013 | <0.0001 | 0.0022 | 0.05 |
| Shape |
Statistics of extracted features from the 13 methods.
| Type | Method | Counts of Features | |
|---|---|---|---|
| TI | CMIM | 20 | 0.004 ± 0.011 |
| MIM | 20 | 0.008 ± 0.015 | |
| JMI | 20 | 0.014 ± 0.016 | |
| MRMR | 18 | 0.009 ± 0.01 | |
| MIFS | 18 | 0.006 ± 0.009 | |
| SIF | Fisher | 4 | <0.0001 |
| ReliefF | 12 | <0.0001 | |
| LS | 6 | 0.013 ± 0.018 | |
| STF | FS | 7 | <0.0001 |
| TS | 11 | <0.0001 | |
| SSL | Alpha | 11 | 0.006 ± 0.014 |
| Lasso | 16 | <0.0001 | |
| MCFS | 20 | 0.011 ± 0.012 |
Figure 5Performance of the 13 feature sets on the ten learning models. (a) Five mean indexes of 13 methods on the ten models and (b) CS of 13 methods.
Figure 6Selected mRSRF and statistics in mRS_2, mRS_4, and mRS_7. (a) mRSRF in three situations, and green color represents selected items from 128 outstanding features in Fmethod. (b) Box plot among the three groups of mRSRF. (c) C-index of the Deepsurv model in training. (d,e) Pearson correlation coefficients and p-values among the three groups of mRSRF.
Figure 7Performance of seven feature groups in the situation of mRS_2. (a–e) Auc, Pre, Acc, F1, and Recall on the ten models. (f) ROC curves of seven feature groups on the RF model.
Figure 8Performance of seven feature groups in the situation of mRS_4. (a–e) Auc, Pre, Acc, F1, and Recall on the ten models. (f) ROC curves of seven feature groups on the RF model.
Figure 9Performance of seven feature groups in the situation of mRS_7. (a–e) Auc, Pre, Acc, F1, and Recall on the ten models. (f) ROC curves of seven feature groups on the RF model.