Literature DB >> 32703773

COVID-19 prediction models should adhere to methodological and reporting standards.

Gary S Collins1, Maarten van Smeden2, Richard D Riley3.   

Abstract

Entities:  

Mesh:

Year:  2020        PMID: 32703773      PMCID: PMC7377211          DOI: 10.1183/13993003.02643-2020

Source DB:  PubMed          Journal:  Eur Respir J        ISSN: 0903-1936            Impact factor:   16.671


× No keyword cloud information.
To the Editor: The coronavirus disease 2019 (COVID-19) pandemic has led to a proliferation of clinical prediction models to aid diagnosis, disease severity assessment and prognosis. A systematic review has identified 66 COVID-19 prediction models: concluding that all, with no exception, are at high risk of bias due to concerns surrounding the data quality, statistical analysis and reporting, and none are recommended for use [1]. Therefore, we read with interest the recent paper by Wu et al. [2] describing the development of a model to identify COVID-19 patients with severe disease on admission to facilitate triage. However, our enthusiasm was dampened by a number of concerns surrounding the design, analysis and reporting of the study which deserve highlighting to readers. Our first point relates to design. The authors randomly split their dataset into a training and test set. This been long been shown to be an inefficient use of the data [3], reducing the size of the training set (increasing the risk of model overfitting), and creating a test set too small for model evaluation. There are alternative stronger approaches that use the entire data to both develop and internally validate a model based on cross-validation or bootstrapping [3]. This naturally leads us to further elaborate on the sample size. The sample size in a prediction model study is largely influenced by the number of individuals experiencing the event to be predicted (in the study by Wu et al. [2], those with severe disease). Using published sample size formulae for developing prediction models [4, 5], based on information reported in the study by Wu et al. [2] (75 predictors, outcome prevalence of 0.237), then depending on the anticipated model R-squared, the minimum sample size in the most optimistic scenario (e.g. that the model gives the highest R-squared) would be 1285 individuals (306 events). To precisely estimate the intercept alone requires 279 individuals (66 events). After splitting their data, the authors developed their model with a sample size of 239 individuals (57 events): clearly insufficient to estimate even the model intercept, let alone develop a prediction model. The test set was then used to evaluate the performance of their model comprising 60 individuals of whom ∼14 experienced the event. To put this in perspective, current sample size recommendations to evaluate model performance suggest a minimum of 100 events [6]. The performance of the model was also evaluated separately in each of five external validation datasets where the number of events ranged from 7 to 98, none of which meet this minimum requirement. Other concerns include the handling of missing data; it is hard to believe all patients had complete information on all 75 predictors, and indeed the flow chart reveals 38 individuals with missing data were simply excluded, which can led to bias [7]. Continuous predictors were assumed to be linearly associated with the outcome, which can reduce predictive accuracy. Model overfitting (a clear concern given the small sample size) was not addressed either in adjusting the performance measures for optimism or shrinking the regression coefficients that are likely overestimated (e.g. using penalisation techniques [8]). “Synthetic sampling” was used to address imbalanced data, but this is inappropriate since artificially balancing data will produce an incorrect estimation of the model intercept (unless it is re-adjusted post-estimation), leading to incorrect model predictions (miscalibration). Model performance was poorly and inappropriately assessed, including presenting a confusion matrix (inappropriate for evaluating prediction models [8]), reporting sensitivity/specificity (where net benefit would be more informative [9]), and assessing model calibration using weak and again discredited approaches (e.g. Hosmer–Lemeshow test, rather than calibration plots with graphical loess curves [6]). We also question the arbitrary choice of risk groupings, and why individuals with a predicted risk of 0.21 are considered the same (“middle risk”) as those with a predicted risk of 0.80. Arguably the most important aspect of a prediction model article is the presentation of the model so that others can use or evaluate it in own their own setting. The authors have presented a nomogram and (prematurely) linked to a web calculator. Whilst both these formats can be used to apply the model to individual patients (though given our concerns we urge against this), for independent validation the prediction model needs to be reported in full; namely, all the regression coefficients and the intercept [10], but these are noticeably absent. Finally, the authors followed the STARD checklist for reporting their study, but this is not the correct guideline. STARD is for reporting diagnostic test accuracy studies, and not multivariable clinical prediction models. We urge the authors and other investigators developing (COVID-19) prediction models to consult the TRIPOD Statement (www.tripod-statement.org) for key information to report when describing their prediction model study, so that readers have the minimal information required to judge the quality of the study [10]. The accompanying TRIPOD explanation and elaboration paper describes the rationale of the importance of transparent reporting, but also discusses various methodological considerations [6]. This one-page PDF can be shared freely online. Shareable PDF ERJ-02643-2020.Shareable
  6 in total

1.  COVID-19 enabled co-authoring networks: a country-case analysis.

Authors:  E Sachini; K Sioumalas-Christodoulou; C Chrysomallidis; G Siganos; N Bouras; N Karampekios
Journal:  Scientometrics       Date:  2021-03-26       Impact factor: 3.801

2.  Development and validation of clinical prediction model to estimate the probability of death in hospitalized patients with COVID-19: Insights from a nationwide database.

Authors:  Ibrahim Halil Tanboğa; Uğur Canpolat; Elif Hande Özcan Çetin; Harun Kundi; Osman Çelik; Murat Çağlayan; Naim Ata; Özcan Özeke; Serkan Çay; Cihangir Kaymaz; Serkan Topaloğlu
Journal:  J Med Virol       Date:  2021-02-10       Impact factor: 20.693

3.  Development and validation of the patient history COVID-19 (PH-Covid19) scoring system: a multivariable prediction model of death in Mexican patients with COVID-19.

Authors:  J Mancilla-Galindo; J M Vera-Zertuche; A R Navarro-Cruz; O Segura-Badilla; G Reyes-Velázquez; F J Tepepa-López; P Aguilar-Alonso; J de J Vidal-Mayo; A Kammar-García
Journal:  Epidemiol Infect       Date:  2020-11-26       Impact factor: 2.451

4.  Protocol for the development of a reporting guideline for causal and counterfactual prediction models in biomedicine.

Authors:  Jie Xu; Yi Guo; Fei Wang; Hua Xu; Robert Lucero; Jiang Bian; Mattia Prosperi
Journal:  BMJ Open       Date:  2022-06-20       Impact factor: 3.006

5.  Prediction model for the spread of the COVID-19 outbreak in the global environment.

Authors:  Ron S Hirschprung; Chen Hajaj
Journal:  Heliyon       Date:  2021-06-29

6.  Machine learning in medicine: a practical introduction to natural language processing.

Authors:  Conrad J Harrison; Chris J Sidey-Gibbons
Journal:  BMC Med Res Methodol       Date:  2021-07-31       Impact factor: 4.615

  6 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.