Kym I E Snell1, Lucinda Archer2, Joie Ensor2, Laura J Bonnett3, Thomas P A Debray4, Bob Phillips5, Gary S Collins6, Richard D Riley2. 1. Centre for Prognosis Research, School of Medicine, Keele University, Keele, Staffordshire, United Kingdom. Electronic address: k.snell@keele.ac.uk. 2. Centre for Prognosis Research, School of Medicine, Keele University, Keele, Staffordshire, United Kingdom. 3. Department of Biostatistics, University of Liverpool, Liverpool, United Kingdom. 4. Julius Center for Health Sciences and Primary Care, University Medical Center Utrecht, Utrecht University, Utrecht, The Netherlands. 5. Centre for Reviews and Dissemination, University of York, York, United Kingdom. 6. Centre for Statistics in Medicine, Nuffield Department of Orthopaedics, Rheumatology and Musculoskeletal Sciences, University of Oxford, Oxford, United Kingdom; NIHR Oxford Biomedical Research Centre, John Radcliffe Hospital, Oxford, United Kingdom.
Abstract
INTRODUCTION: Sample size "rules-of-thumb" for external validation of clinical prediction models suggest at least 100 events and 100 non-events. Such blanket guidance is imprecise, and not specific to the model or validation setting. We investigate factors affecting precision of model performance estimates upon external validation, and propose a more tailored sample size approach. METHODS: Simulation of logistic regression prediction models to investigate factors associated with precision of performance estimates. Then, explanation and illustration of a simulation-based approach to calculate the minimum sample size required to precisely estimate a model's calibration, discrimination and clinical utility. RESULTS: Precision is affected by the model's linear predictor (LP) distribution, in addition to number of events and total sample size. Sample sizes of 100 (or even 200) events and non-events can give imprecise estimates, especially for calibration. The simulation-based calculation accounts for the LP distribution and (mis)calibration in the validation sample. Application identifies 2430 required participants (531 events) for external validation of a deep vein thrombosis diagnostic model. CONCLUSION: Where researchers can anticipate the distribution of the model's LP (eg, based on development sample, or a pilot study), a simulation-based approach for calculating sample size for external validation offers more flexibility and reliability than rules-of-thumb.
INTRODUCTION: Sample size "rules-of-thumb" for external validation of clinical prediction models suggest at least 100 events and 100 non-events. Such blanket guidance is imprecise, and not specific to the model or validation setting. We investigate factors affecting precision of model performance estimates upon external validation, and propose a more tailored sample size approach. METHODS: Simulation of logistic regression prediction models to investigate factors associated with precision of performance estimates. Then, explanation and illustration of a simulation-based approach to calculate the minimum sample size required to precisely estimate a model's calibration, discrimination and clinical utility. RESULTS: Precision is affected by the model's linear predictor (LP) distribution, in addition to number of events and total sample size. Sample sizes of 100 (or even 200) events and non-events can give imprecise estimates, especially for calibration. The simulation-based calculation accounts for the LP distribution and (mis)calibration in the validation sample. Application identifies 2430 required participants (531 events) for external validation of a deep vein thrombosis diagnostic model. CONCLUSION: Where researchers can anticipate the distribution of the model's LP (eg, based on development sample, or a pilot study), a simulation-based approach for calculating sample size for external validation offers more flexibility and reliability than rules-of-thumb.
Authors: Garrett S Bullock; Tom Hughes; Amelia H Arundale; Patrick Ward; Gary S Collins; Stefan Kluzek Journal: Sports Med Date: 2022-02-17 Impact factor: 11.928
Authors: Orouba Almilaji; Gwilym Webb; Alec Maynard; Thomas P Chapman; Brian S F Shine; Antony J Ellis; John Hebden; Sharon Docherty; Elizabeth J Williams; Jonathon Snook Journal: Diagn Progn Res Date: 2021-12-15
Authors: Sascha Halvachizadeh; P J Störmann; Orkun Özkurtul; Till Berk; Michel Teuben; Kai Sprengel; Hans-Christoph Pape; Rolf Lefering; Kai Oliver Jensen Journal: BMJ Open Date: 2022-04-13 Impact factor: 3.006
Authors: Eline Janssens; Eline Schillebeeckx; Kathleen Zwijsen; Jo Raskin; Joris Van Cleemput; Veerle F Surmont; Kristiaan Nackaerts; Elly Marcq; Jan P van Meerbeeck; Kevin Lamote Journal: Cancers (Basel) Date: 2022-06-29 Impact factor: 6.575
Authors: Elise Naufal; Marjan Wouthuyzen-Bakker; Sina Babazadeh; Jarrad Stevens; Peter F M Choong; Michelle M Dowsey Journal: Front Rehabil Sci Date: 2022-07-11
Authors: Menelaos Pavlou; Chen Qu; Rumana Z Omar; Shaun R Seaman; Ewout W Steyerberg; Ian R White; Gareth Ambler Journal: Stat Methods Med Res Date: 2021-04-21 Impact factor: 3.021