| Literature DB >> 32275840 |
Michael O Harhay1,2, David H Au3,4, Sharon D Dell5, Michael K Gould6, Susan Redline7,8, Christopher J Ryerson9,10, Colin R Cooke11.
Abstract
Entities:
Mesh:
Year: 2020 PMID: 32275840 PMCID: PMC7258412 DOI: 10.1513/AnnalsATS.202002-141ED
Source DB: PubMed Journal: Ann Am Thorac Soc ISSN: 2325-6621
Key reporting metrics for prediction models
| Domain | Key Reporting Elements |
|---|---|
| Data source | Were data collected prospectively for this purpose or repurposed from an archival dataset? Wherever possible, the data used should be made available to readers. |
| Participants | Which patients were included in the study? Were separate populations used for model derivation and validation? How many patients were included in each of these groups? A “Table 1” describing relevant clinical features is useful. |
| Outcome | Specific details on how the outcome was defined. |
| Predictors | A specific accounting of the predictor variables included in the final model, along with the method by which these variables were selected. |
| Missing data | How much data were missing from the predictors and from the outcome? How was missing data handled? |
| Model specification | What sort of model was used (e.g., linear regression, random forest)? The final model itself should be reported with as much detail as possible, including specific equations/parameters. Whenever possible (particularly in the case of machine learning models), the code used should be provided in full such that others can replicate the analyses. |
| Model structure | The full model equation should be reported when applicable (e.g., statistical models), along with equations required to interpret results (e.g., the baseline hazard function in a time-to-event model). |
| Validation | How was the model validated (internal vs. external)? If internal validation only was performed, how was the dataset split? |
| Model performance | Performance measures should be tailored to the intended purpose of the model but generally should include a measure of discrimination (e.g., AUROC or AUPRC), a measure of calibration (e.g., Hosmer-Lemeshow, scaled Brier score), and clinically relevant performance (e.g., PPV, NPV) as indicated. |
Definition of abbreviations: AUPRC = area under the precision recall curve; AUROC = area under the receiver operating characteristic curve; NPV = negative predictive value; PPV = positive predictive value.
Reprinted by permission from Reference 1, adapted from the (Transparent Reporting of a Multivariable Prediction Model for Individual Prognosis or Diagnosis) checklist (7).
Summary of guidance for prediction models
| Recommended Practices | Cautions |
|---|---|
| Consider competing priorities of precision, parsimony, and transparency when approaching a prediction task. | Prediction frameworks should not be used to make causal inferences. |
| Think carefully about the prediction’s intended purpose and prioritize feature selection elements as appropriate. | Using |
| Report the prevalence and handling of missing data; consider steps other than case exclusion to address missing data. | The size of a dataset, as well as the number of outcomes it contains, limit the number of predictor variables that the model can accommodate. |
| Consider the expected nature of the relationships between predictors and the outcome (e.g., linear, exponential, etc.). | Categorizing continuous variables can lead to loss of information. |
| Conduct external validation to demonstrate a model can generalize to new observations. | External validation should use the same model used to report the internal performance; avoid retraining on the external dataset. |
| Seek reasonable comparators other than “no model” when evaluating model performance. | Relying on the area under receiver operator characteristics curve alone can lead to an incomplete understanding of a model’s performance. |
| Follow appropriate reporting guidelines such as TRIPOD and RECORD. |
Definition of abbreviations: RECORD = Reporting of Studies Conducted Using Observational Routinely-collected Data; TRIPOD = Transparent Reporting of a Multivariable Prediction Model for Individual Prognosis or Diagnosis.
Reprinted by permission from Reference 1.