| Literature DB >> 35388124 |
Xu Cao1, Duo Tan2, Zhi Liu3, Meng Liao1, Yubo Kan1, Rui Yao2, Liqiang Zhang4, Lisha Nie5, Ruikun Liao6, Shanxiong Chen7, Mingguo Xie8.
Abstract
This study aimed to explore the ability of radiomics derived from both MRI and 18F-fluorodeoxyglucose positron emission tomography (18F-FDG-PET) images to differentiate glioblastoma (GBM) from solitary brain metastases (SBM) and to investigate the combined application of multiple models. The imaging data of 100 patients with brain tumours (50 GBMs and 50 SBMs) were retrospectively analysed. Three model sets were built on MRI, 18F-FDG-PET, and MRI combined with 18F-FDG-PET using five feature selection methods and five classification algorithms. The model set with the highest average AUC value was selected, in which some models were selected and divided into Groups A, B, and C. Individual and joint voting predictions were performed in each group for the entire data. The model set based on MRI combined with 18F-FDG-PET had the highest average AUC compared with isolated MRI or 18F-FDG-PET. Joint voting prediction showed better performance than the individual prediction when all models reached an agreement. In conclusion, radiomics derived from MRI and 18F-FDG-PET could help differentiate GBM from SBM preoperatively. The combined application of multiple models can provide greater benefits.Entities:
Mesh:
Substances:
Year: 2022 PMID: 35388124 PMCID: PMC8986767 DOI: 10.1038/s41598-022-09803-8
Source DB: PubMed Journal: Sci Rep ISSN: 2045-2322 Impact factor: 4.379
Figure 1Workflow of current study. (1) The expert segment the region of interest on the image. (2) Radiomic features were extracted for further analysis. (3) Five feature selection methods and five classifiers combined into twenty-five models with the help of cross-validation in the training cohort. Part of the model was picked out and divided into three groups, five models in each group. (4) Combined application of the five models through voting strategies within the group.
Clinical characteristics of the entire data and the sub-dataset.
| Entire data | Training cohort | Validation cohort | |||||||
|---|---|---|---|---|---|---|---|---|---|
| GBM (n = 50) | SBM (n = 50) | GBM (n = 39) | SBM (n = 41) | GBM (n = 11) | SBM (n = 9) | ||||
| Age (years) | 59.24 | 61.31 | 0.211 a | 58.98 | 61.95 | 0.097 a | 60.20 | 58.11 | 0.644 a |
| Male | 26 | 23 | 0.689 b | 18 | 19 | 0.836 b | 8 | 4 | 0.409 b |
| Female | 24 | 27 | 21 | 22 | 3 | 5 | |||
| Supratentorial | 48 | 41 | 0.055 b | 37 | 34 | 0.182 b | 11 | 7 | 0.369 b |
| Infratentorial | 2 | 9 | 2 | 7 | 0 | 2 | |||
| Yes | 46 | 44 | 0.738 b | 36 | 36 | 0.766 b | 10 | 8 | 0.548 b |
| No | 4 | 6 | 3 | 5 | 1 | 1 | |||
| Yes | 46 | 43 | 0.522 b | 36 | 35 | 0.529 b | 10 | 8 | 0.548 b |
| No | 4 | 7 | 3 | 6 | 1 | 1 | |||
SBM solitary brain metastases, GBM glioblastoma.
aStudent ‘s t-test.
bChi-square test.
Extracted features of image data after the t-test.
| Sequence | Firstorder | GLCM | GLRLM | GLSZM | GLDM | NGTDM | Shape | ALL |
|---|---|---|---|---|---|---|---|---|
| PET | 63 | 86 | 49 | 46 | 44 | 20 | 0 | 308 |
| T1c | 96 | 111 | 124 | 107 | 101 | 9 | 1 | 549 |
| T2 | 188 | 216 | 171 | 135 | 140 | 33 | 1 | 884 |
| SUM | 347 | 413 | 344 | 288 | 285 | 62 | 2 | 1741 |
GLCM grey-level co-occurrence matrix, GLRLM grey-level run length matrix, GLSZM grey-level size zone matrix, GLDM grey-level dependence matrix, NGTDM neighboring grey tone difference matrix.
Figure 2The heat map of Fivefold mean AUC of Integration Set in validation cohort. Created by python3.8 (https://www.python.org/downloads/release/python-380).
Figure 3The heat map of Fivefold mean AUC of MRI Group in validation cohort. Created by python3.8 (https://www.python.org/downloads/release/python-380).
Figure 4The heat map of Fivefold mean AUC of PET Group in validation cohort. Created by python3.8 (https://www.python.org/downloads/release/python-380) .
The performance of the 15 selected models from Integration Set in validation cohort.
| Group | Model | AUC | ACC | Sensitivity | Specificity |
|---|---|---|---|---|---|
| A | LASSO-SVM | 0.93 | 0.83 | 0.76 | 0.92 |
| LDA-SVM | 0.92 | 0.86 | 0.84 | 0.91 | |
| LASSO-LR | 0.91 | 0.89 | 0.88 | 0.9 | |
| LDA-LR | 0.90 | 0.87 | 0.84 | 0.91 | |
| LDA-KNN | 0.90 | 0.85 | 0.80 | 0.91 | |
| B | PLS-LR | 0.86 | 0.78 | 0.76 | 0.82 |
| NCA-KNN | 0.84 | 0.83 | 0.86 | 0.81 | |
| PLS-RF | 0.83 | 0.80 | 0.80 | 0.80 | |
| PLS-SVM | 0.83 | 0.78 | 0.82 | 0.74 | |
| PLS-Adaboost | 0.83 | 0.82 | 0.82 | 0.82 | |
| C | PCA-RF | 0.80 | 0.72 | 0.74 | 0.70 |
| NCA-Adaboost | 0.79 | 0.78 | 0.82 | 0.74 | |
| PCA-LR | 0.78 | 0.77 | 0.66 | 0.88 | |
| PCA-Adaboost | 0.78 | 0.77 | 0.70 | 0.84 | |
| LASSO-Adaboost | 0.68 | 0.67 | 0.68 | 0.65 |
AUC area under curve, ACC accuracy, LASSO least absolute shrinkage and selection operator, LDA linear discriminant analysis, PLS partial least squares regression, NCA near-collar component analysis, PCA principal component analysis, SVM support vector machine, LR the logistic regression, KNN K nearest neighbors, RF random forest, Adaboost Adaptive Boosting.
Figure 5The performance of the five individual models and the combined use of each group: (a) the performance in training cohort; (b) The performance in training cohort. A, group A; B, group B; C, group C. 5A, five models reach agreement; 4A, four models reach agreement; 3A, three models reach agreement.
Figure 6The ratios of different agreement patterns in each group: (a) the ratios in training cohort; (b) The ratios in training cohort. A, group A; B, group B; C, group C.