| Literature DB >> 35986798 |
Jingyu Zhong1, Yangfan Hu1, Yue Xing1, Xiang Ge1, Defang Ding1, Huan Zhang2, Weiwu Yao3.
Abstract
BACKGROUND: Multiple tools have been applied to radiomics evaluation, while evidence rating tools for this field are still lacking. This study aims to assess the quality of pancreatitis radiomics research and test the feasibility of the evidence level rating tool.Entities:
Keywords: Differential diagnosis; Machine learning; Pancreatitis; Quality improvement; Systematic review
Year: 2022 PMID: 35986798 PMCID: PMC9391628 DOI: 10.1186/s13244-022-01279-4
Source DB: PubMed Journal: Insights Imaging ISSN: 1869-4101
Fig. 1Flow diagram of study inclusion
Study characteristics
| Study characteristics | Data |
|---|---|
| Sample size, mean ± standard deviation, median (range) | 137.5 ± 85.0, 111 (41–389) |
| Journal type, | |
| Imaging | 16 (53) |
| Non-imaging | 14 (47) |
| First authorship, | |
| Radiologist | 24 (80) |
| Non-radiologist | 6 (20) |
| Biomarker, | |
| Diagnostic | 24 (80) |
| Prognostic | 6 (20) |
| Imaging modality, | |
| CT | 13 (43) |
| EUS | 4 (13) |
| MRI | 9 (30) |
| PET | 4 (13) |
| Model type, | |
| Type 1a: Developed model validated with exactly the same data | 7 (23) |
| Type 1b: Developed model validated with resampling data | 10 (33) |
| Type 2a: Developed model validated with randomly splitting data | 12 (40) |
| Type 2b: Developed model validated with non-randomly splitting data | 1 (3) |
| Type 3: Developed model validated with separate data | 0 (0) |
| Type 4: Validation only | 0 (0) |
| Phase classification, | |
| Phase 0: < 100 patients; retrospective; internal validation | 16 (53) |
| Phase I: < 100 patients; retrospective; external validation | 2 (7) |
| Phase II: > 100 patients; retrospective; external validation | 12 (40) |
| Phase III: > 100 patients; prospective; external validation | 0 (0) |
| Phase IV: real-world | 0 (0) |
Fig. 2Study topics and number of studies. Three studies investigated two topics, respectively, and had been treated as two different studies in the term of topic. Therefore, there were thirty studies according to article, but thirty-three models according to topic. The bolded number with modality indicates the studies included in the meta-analysis
RQS rating of included studies
| 16 items according to 6 key domains | Range | Median (range) | Percentage of ideal score, | Adherence rate, |
|---|---|---|---|---|
| Total 16 items | − 8 to 36 | 7 (− 3 to 18) | 7.3 (20.2) | 184 (38) |
| Domain 1: protocol quality and stability in image and segmentation | 0–5 | 2 (0–2) | 1.6 (31.3) | 47 (15) |
| Protocol quality | 0–2 | 1 (0–1) | 0.9 (46.7) | 28 (93) |
| Multiple segmentations | 0–1 | 1 (0–1) | 0.6 (63.3) | 19 (63) |
| Test–retest | 0–1 | 0 (0–0) | 0 (0) | 0 (0) |
| Phantom study | 0–1 | 0 (0–0) | 0 (0) | 0 (0) |
| Domain 2: feature selection and validation | − 8 to 8 | − 2 (− 8 to 6) | 0.9 (10.8) | 42 (70) |
| Feature reduction or adjustment of multiple testing | − 3 to 3 | 3 (− 3 to 3) | 2.8 (93.3) | 29 (97) |
| Validation | − 5 to 5 | − 5 (− 5 to 3) | − 1.9 (0) | 13 (43) |
| Domain 3: biologic/clinical validation and utility | 0–6 | 1.5 (0–6) | 2.0 (33.9) | 47 (39) |
| Non-radiomics features | 0–1 | 0.5 (0–1) | 0.5 (50.0) | 15 (60) |
| Biologic correlations | 0–1 | 1 (0–1) | 0.6 (60.0) | 18 (60) |
| Comparison with “gold standard” | 0–2 | 0 (0–2) | 0.8 (40.0) | 12 (40) |
| Potential clinical utility | 0–2 | 0 (0–2) | 0.1 (6.7) | 2 (7) |
| Domain 4: model performance index | 0–5 | 2 (1–4) | 2.1 (42.7) | 34 (38) |
| Cutoff analysis | 0–1 | 0 (0–0) | 0 (0) | 0 (0) |
| Discrimination statistics | 0–2 | 2 (1–2) | 1.9 (95.0) | 30 (100) |
| Calibration statistics | 0–2 | 0 (0–2) | 0.2 (11.7) | 4 (13) |
| Domain 5: high level of evidence | 0–8 | 0 (0–7) | 0.2 (2.9) | 1 (2) |
| Prospective study | 0–7 | 0 (0–7) | 0.2 (3.3) | 1 (3) |
| Cost-effectiveness analysis | 0–1 | 0 (0–0) | 0 (0) | 0 (0) |
| Domain 6: open science and data | 0–4 | 0 (0–1) | 0.4 (10.8) | 13 (43) |
RQS Radiomics Quality Score
Fig. 3Quality assessment of included studies. a Ideal percentage of RQS; b TRIPOD adherence rate; c QUADAS-2 assessment result
TRIPOD adherence of included studies
| 35 Selected Items in 20 Criteria According to 6 Sections ( | Study, |
|---|---|
| 478 (61) | |
| Section 1: Title and Abstract | 14 (23) |
| 1. Title—identify developing/validating a model, target population, and the outcome | 2 (7) |
| 2. Abstract—provide a summary of objectives, study design, setting, participants, sample size, predictors, outcome, statistical analysis, results, and conclusions | 12 (40) |
| Section 2: Introduction | 37 (62) |
| 3a. Background—Explain the medical context and rationale for developing/validating the model | 30 (100) |
| 3b. Objective—Specify the objectives, including whether the study describes the development/validation of the model or both | 7 (23) |
| Section 3: Methods | 252 (65) |
| 4a. Source of data—describe the study design or source of data (randomized trial, cohort, or registry data) | 30 (100) |
| 4b. Source of data—specify the key dates | 26 (87) |
| 5a. Participants—specify key elements of the study setting including number and location of centers | 30 (100) |
| 5b. Participants—describe eligibility criteria for participants (inclusion and exclusion criteria) | 21 (70) |
| 5c. Participants—give details of treatment received, if relevant ( | 0 (0) |
| 6a. Outcome—clearly define the outcome, including how and when assessed | 30 (100) |
| 6b. Outcome—report any actions to blind assessment of the outcome | 0 (0) |
| 7a. Predictors—clearly define all predictors, including how and when assessed | 23 (77) |
| 7b. Predictors—report any actions to blind assessment of predictors for the outcome and other predictors | 15 (50) |
| 8. Sample size—explain how the study size was arrived at | 0 (0) |
| 9. Missing data—describe how missing data were handled with details of any imputation method | 0 (0) |
| 10a. Statistical analysis methods—describe how predictors were handled | 24 (80) |
| 10b. Statistical analysis methods—specify type of model, all model-building procedures (any predictor selection), and method for internal validation | 23 (77) |
| 10d. Statistical analysis methods—specify all measures used to assess model performance and if relevant, to compare multiple models (discrimination and calibration) | 30 (100) |
| 11. Risk groups—provide details on how risk groups were created, if done ( | 0 (0) |
| Section 4: Results | 94 (52) |
| 13a. Participants—describe the flow of participants, including the number of participants with and without the outcome. A diagram may be helpful | 16 (53) |
| 13b. Participants—describe the characteristics of the participants, including the number of participants with missing data for predictors and outcome | 24 (80) |
| 14a. Model development—specify the number of participants and outcome events in each analysis | 25 (83) |
| 14b. Model development—report the unadjusted association between each candidate predictor and outcome, if done ( | 1 (20) |
| 15a. Model specification—present the full prediction model to allow predictions for individuals (regression coefficients, intercept) | 5 (17) |
| 15b. Model specification—explain how to the use the prediction model (nomogram, calculator, etc.) | 2 (7) |
| 16. Model performance—report performance measures (with confidence intervals) for the prediction model | 22 (73) |
| Section 5: Discussion | 81 (90) |
| 18. Limitations—Discuss any limitations of the study | 30 (100) |
| 19b. Interpretation—Give an overall interpretation of the results | 30 (100) |
| 20. Implications—Discuss the potential clinical use of the model and implications for future research | 21 (70) |
| Section 6: Validation for Model type 2a, 2b, 3, and 4 ( | 9 (17) |
| 10c. Statistical analysis methods—describe how the predictions were calculated | 0 (0) |
| 10e. Statistical analysis methods—describe any model updating (recalibration), if done ( | 0 (0) |
| 12. Development versus validation—Identify any differences from the development data in setting, eligibility criteria, outcome, and predictors | 5 (38) |
| 13c. Participants (for validation)—show a comparison with the development data of the distribution of important variables | 4 (31) |
| 17. Model updating—report the results from any model updating, if done ( | 0 (0) |
| 19a. Interpretation (for validation)—discuss the results with reference to performance in the development data and any other validation data | 0 (0) |
TRIPOD Transparent Reporting of a Multivariable Prediction Model for Individual Prognosis or Diagnosis
Fig. 4IBSI preprocessing steps performed in included studies. a Adherence rate of IBSI preprocessing steps; b segmentation method; c software for radiomics feature extraction. The other software included Omni-Kinetics, Artificial Intelligent Kit, AnalysisKit, Image J, FireVoxel, and MaZda
Fig. 5Forest plots of diagnostic odds ratio for differentiation diagnosis. a Autoimmune pancreatitis versus pancreatic cancer by CT; b mass-forming focal pancreatitis versus pancreatic cancer by MRI
Diagnostic performance of meta-analyzed clinical questions
| Clinical question | AIP versus PC by CT | MFP versus PC by MRI |
|---|---|---|
| Number of studies | 6 | 4 |
| Number of available datasets | 5/8 | 5/6 |
| Events/sample size | 191/421 | 101/320 |
| Pooled analysis | ||
| DOR (95% CI) | 189.63 (79.65–451.48) | 135.70 (36.17–509.13) |
| < 0.001 | < 0.001 | |
| Sensitivity (95% CI) | 0.90 (0.84–0.94) | 0.90 (0.81–0.95) |
| Specificity (95% CI) | 0.95 (0.92–0.97) | 0.94 (0.86–0.98) |
| PLR (95% CI) | 19.01 (10.51–34.40) | 15.00 (5.94–37.92) |
| NLR (95% CI) | 0.10 (0.06–0.17) | 0.11 (0.06–0.56) |
| AUC (95% CI) | 0.97 (0.95–0.98) | 0.95 (0.93–0.96) |
| Heterogeneity | ||
| Higgins I2 test (%) | 83.26% | 97.28% |
| Cochran’s Q test ( | < 0.01 | < 0.01 |
| Publication bias | ||
| Egger’s test ( | 0.060 | 0.050 |
| Begg’s test ( | 0.221 | 0.221 |
| Deeks test ( | 0.226 | 0.538 |
| Trim and fill method | ||
| Number of missing datasets | 2 | 2 |
| Adjusted DOR (95%CI) | 135.11 (64.40–283.74) | 53.89 (15.95–182.00) |
| Level of evidence | Weak | Weak |
AIP autoimmune pancreatitis, AUC area under curve, CI confidential interval, DOR diagnostic odds ratio, MFP mass-forming pancreatitis, NLR negative likelihood ratio, n/a not applicable, PC pancreatic cancer, PLR positive likelihood ratio
Fig. 6Correlations between study characteristics and quality. Swam plots of (a) ideal percentage of RQS, (b) TRIPOD adherence rate, and (c) IBSI adherence rate. The diameter of bubbles indicates the sample size of studies. Seven studies published on journals without impact factor were excluded. The lighter color indicates the studies after the publication of RQS, TRIPOD, and IBSI; the darker color indicates those before their publications