Literature DB >> 32275840

Methodologic Guidance and Expectations for the Development and Reporting of Prediction Models and Causal Inference Studies.

Michael O Harhay1,2, David H Au3,4, Sharon D Dell5, Michael K Gould6, Susan Redline7,8, Christopher J Ryerson9,10, Colin R Cooke11.   

Abstract

Entities:  

Mesh:

Year:  2020        PMID: 32275840      PMCID: PMC7258412          DOI: 10.1513/AnnalsATS.202002-141ED

Source DB:  PubMed          Journal:  Ann Am Thorac Soc        ISSN: 2325-6621


× No keyword cloud information.
The May 2020 issue of Critical Care Medicine includes a new consensus document for developing, validating, updating, and reporting prediction models (1). This document was co-created and is co-signed by statistical editors, associate editors, and editors-in-chief at 31 pulmonary, sleep, and critical care journals, including AnnalsATS. This prediction model guidance document is the second guidance document produced by this large group of editors following the guidance on causal inference studies published in AnnalsATS in January 2019 (2). Authors submitting their work to AnnalsATS will often be directed to these (and potentially future) guidance documents, as well as the AnnalsATS detailed instructions to authors. Our goal in the present editorial is to provide the rationale and vision for how these documents and the recommendations within them can best be used by authors and reviewers.

What Kind of Studies Should Use the Guidance Documents?

AnnalsATS receives a diverse collection of manuscripts that ask a wide variety of scientific questions. Many studies are motivated by causal questions (3, 4). Some examples of such questions include: Does air pollution cause asthma? Does adherence to guidelines improve patient outcomes? Are outcomes better with a double-lung than with a single-lung transplant? The goal of the first guidance document was to provide an accessible contemporary summary and reference guide for authors to use to explore such questions using causal inference methods. Causal inference methods offer powerful and recommended conceptual and empirical tools to design studies, develop and refine statistical models, and estimate and report effect estimates (5, 6). Though causal inference methods can be used to improve the design and analysis of randomized trials, they are especially useful in guiding observational studies that seek to examine relationships between an exposure and an outcome using nonrandomized data sources. In contrast to causal inference studies, prediction modeling studies aim to develop, validate, or update a mathematical equation that calculates a specific probability or risk of a condition or future event for an individual (7). To clarify how these two study designs differ, consider the setting of lung cancer. An observational causal inference study might seek to provide an estimate of the average increase in the risk of lung cancer for each year of smoking among subjects who smoke compared with subjects without a smoking history. That is, it seeks to estimate a relationship between an exposure and an outcome, such that we can consider what would have happened if a patient had smoked less or not at all. In hopes of achieving an informative effect estimate while minimizing bias, a researcher would use causal inference methods to identify key confounders and the appropriate statistical model to generate an effect estimate for the association of smoking history with lung cancer. For a prediction model, a potential goal would be to assign an individual probability of developing lung cancer for each patient at a specific time in the future. Though the equation developed to accomplish this “prediction” goal may include variables identified from causal inference studies, such as smoking history, it does not have to include such variables, nor are effect estimates and confounding of primary interest. The primary goal of a prediction model is instead to identify the combination of variables that most accurately predicts individual outcomes, regardless of whether these variables fall along causal pathways. Prediction models familiar to readers might include the Acute Physiology and Chronic Health Evaluation IV score (8) to predict mortality at the time of intensive care unit admission or the Framingham Risk Score for general cardiovascular disease risk (9). The goal of this new document published in Critical Care Medicine (1) was to provide a unique and accessible summary of the statistical literature on the best practices for developing, validating, and reporting prediction model studies.

Why Have These Documents Been Produced?

Prediction modeling and causal inference studies are abundant in the medical literature. Although these have long been accepted research pursuits, these study designs are also frequently undertaken using methods and approaches that are both prone to bias and no longer recommended by the statistical community. Accordingly, there is logic in assembling guidance documents that summarize the leading perspectives and recommended approaches for these study designs that can be used by authors, reviewers, and readers alike. Doing so across journals provides several benefits. First, there is a general desire among editorial teams to provide greater clarity to authors regarding the statistical analysis expectations at their respective journals. Second, promoting the use of accepted and preferred methodology helps elevate the rigor and quality of research in our respective disciplines. Third, copromotion across journals helps authors design studies and draft manuscripts that are broadly acceptable for peer review at multiple venues. Finally, communicating a common set of criteria for evaluating the methodological rigor of statistical analysis approaches helps reviewers and editors avoid redundancy by providing generalized responses to common issues such as underreporting and methodological concerns. This improves the efficiency of the editorial process for everyone.

How Should Authors and Reviewers Approach These Documents?

First, we want to emphasize that these guidance documents are not a set of prescriptive rules that must be followed without deviation. Furthermore, they are not a simple recipe that, if followed, will lead to publication. They should be interpreted as strong guidance representing contemporary views and consensus recommendations on best practices for common study types, and not as immutable editorial policy. A principal goal of these documents was to push authors toward greater conceptual and reporting clarity in their manuscript submissions. Authors retain their full discretion to pursue the study design they deem most appropriate for their research questions. We invite authors to innovate; yet, when they do, they should justify the validity of their approach. Our goal is to improve the overall quality of research; it is not to stifle development or use of novel and innovative approaches.

What Is Contained in the New Prediction Recommendations?

For ease, we have reproduced two tables that summarize key elements to consider reporting in prediction model manuscripts (Table 1) and recommended prediction model practices (Table 2). We highlight a few specific topics that the editorial team at AnnalsATS will be keenly assessing as we move forward.
Table 1.

Key reporting metrics for prediction models

DomainKey Reporting Elements
Data sourceWere data collected prospectively for this purpose or repurposed from an archival dataset? Wherever possible, the data used should be made available to readers.
ParticipantsWhich patients were included in the study? Were separate populations used for model derivation and validation? How many patients were included in each of these groups? A “Table 1” describing relevant clinical features is useful.
OutcomeSpecific details on how the outcome was defined.
PredictorsA specific accounting of the predictor variables included in the final model, along with the method by which these variables were selected.
Missing dataHow much data were missing from the predictors and from the outcome? How was missing data handled?
Model specificationWhat sort of model was used (e.g., linear regression, random forest)? The final model itself should be reported with as much detail as possible, including specific equations/parameters. Whenever possible (particularly in the case of machine learning models), the code used should be provided in full such that others can replicate the analyses.
Model structureThe full model equation should be reported when applicable (e.g., statistical models), along with equations required to interpret results (e.g., the baseline hazard function in a time-to-event model).
ValidationHow was the model validated (internal vs. external)? If internal validation only was performed, how was the dataset split?
Model performancePerformance measures should be tailored to the intended purpose of the model but generally should include a measure of discrimination (e.g., AUROC or AUPRC), a measure of calibration (e.g., Hosmer-Lemeshow, scaled Brier score), and clinically relevant performance (e.g., PPV, NPV) as indicated.

Definition of abbreviations: AUPRC = area under the precision recall curve; AUROC = area under the receiver operating characteristic curve; NPV = negative predictive value; PPV = positive predictive value.

Reprinted by permission from Reference 1, adapted from the (Transparent Reporting of a Multivariable Prediction Model for Individual Prognosis or Diagnosis) checklist (7).

Table 2.

Summary of guidance for prediction models

Recommended PracticesCautions
Consider competing priorities of precision, parsimony, and transparency when approaching a prediction task.Prediction frameworks should not be used to make causal inferences.
Think carefully about the prediction’s intended purpose and prioritize feature selection elements as appropriate.Using P values from bivariable comparisons or stepwise procedures to select predictors leads to bias and overfitting.
Report the prevalence and handling of missing data; consider steps other than case exclusion to address missing data.The size of a dataset, as well as the number of outcomes it contains, limit the number of predictor variables that the model can accommodate.
Consider the expected nature of the relationships between predictors and the outcome (e.g., linear, exponential, etc.).Categorizing continuous variables can lead to loss of information.
Conduct external validation to demonstrate a model can generalize to new observations.External validation should use the same model used to report the internal performance; avoid retraining on the external dataset.
Seek reasonable comparators other than “no model” when evaluating model performance.Relying on the area under receiver operator characteristics curve alone can lead to an incomplete understanding of a model’s performance.
Follow appropriate reporting guidelines such as TRIPOD and RECORD. 

Definition of abbreviations: RECORD = Reporting of Studies Conducted Using Observational Routinely-collected Data; TRIPOD = Transparent Reporting of a Multivariable Prediction Model for Individual Prognosis or Diagnosis.

Reprinted by permission from Reference 1.

Key reporting metrics for prediction models Definition of abbreviations: AUPRC = area under the precision recall curve; AUROC = area under the receiver operating characteristic curve; NPV = negative predictive value; PPV = positive predictive value. Reprinted by permission from Reference 1, adapted from the (Transparent Reporting of a Multivariable Prediction Model for Individual Prognosis or Diagnosis) checklist (7). Summary of guidance for prediction models Definition of abbreviations: RECORD = Reporting of Studies Conducted Using Observational Routinely-collected Data; TRIPOD = Transparent Reporting of a Multivariable Prediction Model for Individual Prognosis or Diagnosis. Reprinted by permission from Reference 1. First, the guidance builds heavily on the TRIPOD (Transparent Reporting of a Multivariable Prediction Model for Individual Prognosis or Diagnosis) checklists for prediction model development and validation (www.equator-network.org/reporting-guidelines/tripod-statement/). Since 2015, many leading journals have required adherence to the TRIPOD checklists. Submissions to AnnalsATS for all prediction model studies should have completed TRIPOD checklists appropriate to the study. Second, we are interested in receiving work that goes beyond the development and internal validation of prediction tools. Articles validating a prediction tool in novel cohorts or articles showing that a prediction tool leads to changes in practice, clinical decision making, or outcomes will be prioritized. Although model development and validation for important topics is essential work that we will continue to publish, we challenge authors to send us work with real-world benefit, not just work with higher c-statistics. The medical literature is full of prediction models that perform well but are never used or fail to change clinical care. Rather than expand the number of unused models, AnnalsATS aspires to publish studies that prioritize added value to patient care. Third, we are particularly interested in the topic of prediction model “bias” or “fairness” (10). Prediction models that originate from datasets that lack representation of demographic groups traditionally overlooked in biomedical research may lead to bias in real-world applications (11, 12). As a result, these prediction models may create or reinforce biases and health disparities (11–16). For example, inclusion of black race in a prediction model could suggest that being black is associated with poor outcomes in patients with chronic obstructive pulmonary disease or low continuous positive airway pressure therapy adherence in patients with sleep apnea. However, if the poor outcomes observed in the dataset used to generate the prediction model reflect surmountable access to care issues or the outcomes of a small total number of black patients (i.e., unrelated to ancestry-linked biological factors), then perpetuating this effect in a published prediction model could negatively impact future care decisions for black patients based on this model. As a general recommendation, inclusion of such factors in prediction models should be carefully considered and justified by what information they add in the context of specific study questions (17). In closing, we reiterate that these guidance documents are better thought of as strong recommendations and not as unwavering rules and requirements for publication. Our intent is to help authors improve the rigor of their studies rather than to discourage submission. Ultimately, each submitted manuscript is evaluated on its own merits. Innovations in statistical methods, the allure of machine learning, and expanding data size and richness have led to an increase in the number of submissions related to prediction modeling as well as observational research studies. Our goal is very simply to provide tools that promote high-quality contributions to medical research in these areas. We look forward to reviewing your contributions.
  14 in total

1.  Use of race and ethnicity in biomedical publication.

Authors:  Judith B Kaplan; Trude Bennett
Journal:  JAMA       Date:  2003-05-28       Impact factor: 56.272

2.  Ensuring Fairness in Machine Learning to Advance Health Equity.

Authors:  Alvin Rajkomar; Michaela Hardt; Michael D Howell; Greg Corrado; Marshall H Chin
Journal:  Ann Intern Med       Date:  2018-12-04       Impact factor: 25.391

3.  The C-Word: Scientific Euphemisms Do Not Improve Causal Inference From Observational Data.

Authors:  Miguel A Hernán
Journal:  Am J Public Health       Date:  2018-03-22       Impact factor: 9.308

4.  The C-Word: The More We Discuss It, the Less Dirty It Sounds.

Authors:  Miguel Hernán
Journal:  Am J Public Health       Date:  2018-05       Impact factor: 9.308

5.  Dissecting racial bias in an algorithm used to manage the health of populations.

Authors:  Ziad Obermeyer; Brian Powers; Christine Vogeli; Sendhil Mullainathan
Journal:  Science       Date:  2019-10-25       Impact factor: 47.728

6.  General cardiovascular risk profile for use in primary care: the Framingham Heart Study.

Authors:  Ralph B D'Agostino; Ramachandran S Vasan; Michael J Pencina; Philip A Wolf; Mark Cobain; Joseph M Massaro; William B Kannel
Journal:  Circulation       Date:  2008-01-22       Impact factor: 29.690

7.  Control of Confounding and Reporting of Results in Causal Inference Studies. Guidance for Authors from Editors of Respiratory, Sleep, and Critical Care Journals.

Authors:  David J Lederer; Scott C Bell; Richard D Branson; James D Chalmers; Rachel Marshall; David M Maslove; David E Ost; Naresh M Punjabi; Michael Schatz; Alan R Smyth; Paul W Stewart; Samy Suissa; Alex A Adjei; Cezmi A Akdis; Élie Azoulay; Jan Bakker; Zuhair K Ballas; Philip G Bardin; Esther Barreiro; Rinaldo Bellomo; Jonathan A Bernstein; Vito Brusasco; Timothy G Buchman; Sudhansu Chokroverty; Nancy A Collop; James D Crapo; Dominic A Fitzgerald; Lauren Hale; Nicholas Hart; Felix J Herth; Theodore J Iwashyna; Gisli Jenkins; Martin Kolb; Guy B Marks; Peter Mazzone; J Randall Moorman; Thomas M Murphy; Terry L Noah; Paul Reynolds; Dieter Riemann; Richard E Russell; Aziz Sheikh; Giovanni Sotgiu; Erik R Swenson; Rhonda Szczesniak; Ronald Szymusiak; Jean-Louis Teboul; Jean-Louis Vincent
Journal:  Ann Am Thorac Soc       Date:  2019-01

8.  Transparent Reporting of a multivariable prediction model for Individual Prognosis or Diagnosis (TRIPOD): the TRIPOD statement.

Authors:  Gary S Collins; Johannes B Reitsma; Douglas G Altman; Karel G M Moons
Journal:  Ann Intern Med       Date:  2015-01-06       Impact factor: 25.391

9.  Race, ethnicity and lung function: A brief history.

Authors:  Lundy Braun
Journal:  Can J Respir Ther       Date:  2015

10.  Development and Reporting of Prediction Models: Guidance for Authors From Editors of Respiratory, Sleep, and Critical Care Journals.

Authors:  Daniel E Leisman; Michael O Harhay; David J Lederer; Michael Abramson; Alex A Adjei; Jan Bakker; Zuhair K Ballas; Esther Barreiro; Scott C Bell; Rinaldo Bellomo; Jonathan A Bernstein; Richard D Branson; Vito Brusasco; James D Chalmers; Sudhansu Chokroverty; Giuseppe Citerio; Nancy A Collop; Colin R Cooke; James D Crapo; Gavin Donaldson; Dominic A Fitzgerald; Emma Grainger; Lauren Hale; Felix J Herth; Patrick M Kochanek; Guy Marks; J Randall Moorman; David E Ost; Michael Schatz; Aziz Sheikh; Alan R Smyth; Iain Stewart; Paul W Stewart; Erik R Swenson; Ronald Szymusiak; Jean-Louis Teboul; Jean-Louis Vincent; Jadwiga A Wedzicha; David M Maslove
Journal:  Crit Care Med       Date:  2020-05       Impact factor: 7.598

View more
  2 in total

1.  Timing Is Everything. The Importance of Alignment of Time Anchors for Observational Causal Inference Research.

Authors:  Stephanie Parks Taylor; Marc A Kowalkowski; Andrew J Admon
Journal:  Ann Am Thorac Soc       Date:  2021-05

Review 2.  Algorithmic prognostication in critical care: a promising but unproven technology for supporting difficult decisions.

Authors:  Gary E Weissman; Vincent X Liu
Journal:  Curr Opin Crit Care       Date:  2021-10-01       Impact factor: 3.359

  2 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.