| Literature DB >> 31270976 |
Ji Eun Park1, Seo Young Park2, Hwa Jung Kim3, Ho Sung Kim1.
Abstract
Radiomics, which involves the use of high-dimensional quantitative imaging features for predictive purposes, is a powerful tool for developing and testing medical hypotheses. Radiologic and statistical challenges in radiomics include those related to the reproducibility of imaging data, control of overfitting due to high dimensionality, and the generalizability of modeling. The aims of this review article are to clarify the distinctions between radiomics features and other omics and imaging data, to describe the challenges and potential strategies in reproducibility and feature selection, and to reveal the epidemiological background of modeling, thereby facilitating and promoting more reproducible and generalizable radiomics research.Entities:
Keywords: Generalizability; Machine learning; Radiomics; Reproducibility
Mesh:
Year: 2019 PMID: 31270976 PMCID: PMC6609433 DOI: 10.3348/kjr.2018.0070
Source DB: PubMed Journal: Korean J Radiol ISSN: 1229-6929 Impact factor: 3.500
Fig. 1Relationships among reproducibility, internal validity, and generalizability of radiomics features.
Reproducible radiomics features contribute internal validity wherein features are associated with outcome without noise or error. Generalizability refers to external validity, i.e., whether model can be transported and adopted to different populations.
Fig. 2Reproducibility in radiomics research.
Reproducibility in radiomics analysis can be obtained by pursuing imaging data reproducibility, segmentation reproducibility, computational or statistical reproducibility, and research reproducibility. ADC = apparent diffusion coefficient, CBV = cerebral blood volume
Strategies for Reproducible Radiomics Features
| Aspects | Strategy | Purpose | Utility for Feature Selection |
|---|---|---|---|
| Imaging data | Test-retest study with short time interval | Intra-individual repeatability | Yes |
| Use of same reconstruction methods on CT, MRI, and PET | Imaging data reproducibility | No | |
| Phantom or patient study | Multi-machine/center reproducibility | Yes | |
| Quantitative* maps of ADC or CBV | Multi-machine/center reproducibility | No | |
| Normalization* to contralateral side | Multi-machine/center reproducibility | No | |
| Patient-specific radiomics | |||
| Delta-radiomics* | Longitudinal data | No | |
| Patient-specific radiomics | |||
| Segmentation | Multi-reader segmentation | Segmentation reproducibility | Yes |
| Automated segmentation (possible deep learning) | Segmentation reproducibility | No | |
| Feature extraction | Use of same discretization and quantization methods across studies (standardization): Pyradiomics | Quantification reproducibility | No |
| Feature processing | Correction of batch effect from different machine andprotocols: Combat function | Quantification reproducibility | No |
*Potential, published strategies to improve reproducibility. ADC = apparent diffusion coefficient, CBV = cerebral blood volume, PET = positron emission tomography
No Golden Rule, but Possible Ensemble of Feature-Selection Methods for Radiomics Studies
| Strategy | Details |
|---|---|
| Based on reproducibility | |
| Test-retest analysis | 1. Sample size calculation |
| 2. Two or three imaging acquisitions for repeatability | |
| 3. Feature selection with high repeatability | |
| Segmentation reproducibility | 1. Segmentation by two or three readers |
| 2. Feature selection with high reproducibility | |
| Based on univariate test | |
| Filter methods such as t test, univariate logistic regression, correlation | Screen one feature at time based on strength of association with outcome |
| Based on multivariable models | |
| LASSO | Automatic feature selection |
| Selects one feature among correlated features | |
| Elastic Net | Automatic feature selection |
| Selects all features that are correlated each other, or takes them out altogether | |
| SVM, ridge regression | Use magnitude of estimated beta coefficients to select features |
| Deep learning, random forest | More appropriate when sample size is huge |
LASSO = least absolute shrinkage and selection operator, SVM = support vector machine
Fig. 3Various internal validation methods.
Split-sample, CV, CV with iterations, and nested CV methods can be applicable. Bootstrapping method can be combined to other internal validation methods. Note that CV has single AUC since each patient is tested once. AUC = area under receiver operating characteristic curve, CV = cross-validation