| Literature DB >> 35406372 |
Abdalla Ibrahim1,2,3,4, Bruno Barufaldi5, Turkey Refaee1,6, Telmo M Silva Filho7, Raymond J Acciavatti5, Zohaib Salahuddin1, Roland Hustinx3, Felix M Mottaghy2,4, Andrew D A Maidment5, Philippe Lambin1,2.
Abstract
The reproducibility of handcrafted radiomic features (HRFs) has been reported to be affected by variations in imaging parameters, which significantly affect the generalizability of developed signatures and translation to clinical practice. However, the collective effect of the variations in imaging parameters on the reproducibility of HRFs remains unclear, with no objective measure to assess it in the absence of reproducibility analysis. We assessed these effects of variations in a large number of scenarios and developed the first quantitative score to assess the reproducibility of CT-based HRFs without the need for phantom or reproducibility studies. We further assessed the potential of image resampling and ComBat harmonization for removing these effects. Our findings suggest a need for radiomics-specific harmonization methods. Our developed score should be considered as a first attempt to introduce comprehensive metrics to quantify the reproducibility of CT-based handcrafted radiomic features. More research is warranted to demonstrate its validity in clinical contexts and to further improve it, possibly by the incorporation of more realistic situations, which better reflect real patients' situations.Entities:
Keywords: ComBat; harmonization; radiomics reproducibility
Year: 2022 PMID: 35406372 PMCID: PMC8997100 DOI: 10.3390/cancers14071599
Source DB: PubMed Journal: Cancers (Basel) ISSN: 2072-6694 Impact factor: 6.639
Figure 1Explanatory diagram of our workflow. (a) The steps undertaken for data collection, analysis, and the development of the score; (b) a table showing different acquisition and reconstruction parameters of the phantom dataset; (c) a table showing the numeric value assigned to each kernel in the data analyzed.
Figure 2Boxplot of the number of reproducible HRFS (RRFs) across the different scenarios.
Figure 3Variable importance of the regression random forest model.
Performance of the score threshold for the identification of different HRFs reproducibility thresholds.
| Percentage RRFs | Score | AUC | CI 95% Lower | CI 95% Upper | Specificity | Sensitivity | False Alarm |
|---|---|---|---|---|---|---|---|
| Threshold 10% | 0.75 | 0.86 | 0.855 | 0.867 | 0.81 | 0.74 | 0.19 |
| Threshold 20% | 0.77 | 0.85 | 0.842 | 0.851 | 0.76 | 0.77 | 0.24 |
| Threshold 25% | 0.80 | 0.85 | 0.843 | 0.852 | 0.80 | 0.74 | 0.20 |
| Threshold 30% | 0.83 | 0.86 | 0.851 | 0.86 | 0.84 | 0.73 | 0.16 |
| Threshold 40% | 0.85 | 0.87 | 0.868 | 0.878 | 0.81 | 0.80 | 0.19 |
| Threshold 50% | 0.88 | 0.90 | 0.892 | 0.904 | 0.83 | 0.85 | 0.17 |
| Threshold 60% | 0.88 | 0.92 | 0.91 | 0.925 | 0.79 | 0.92 | 0.21 |
| Threshold 70% | 0.94 | 0.96 | 0.952 | 0.966 | 0.94 | 0.89 | 0.06 |
| Threshold 75% | 0.95 | 0.97 | 0.967 | 0.977 | 0.95 | 0.93 | 0.05 |
| Threshold 80% | 0.96 | 0.98 | 0.971 | 0.983 | 0.95 | 0.95 | 0.05 |
| Threshold 90% | 0.98 | 0.99 | 0.982 | 0.996 | 0.97 | 0.97 | 0.03 |
Figure 4(a) AUC distributions across 100-runs for MaasPenn radiomics reproducibility score in the training and validation datasets for each of the thresholds of percentage reproducible HRFs; (b) The sensitivity as a function of the score and threshold; and (c) The specificity as a function of the score and threshold.
Figure 5The proposed workflow of MaasPenn score for (a) planning new analyses; and (b) evaluating previously developed signatures.