Lauren A May1,2, R Paul Guillerman3, Wei Zhang3, Robert C Orth3. 1. Edward B. Singleton Department of Pediatric Radiology, Texas Children's Hospital, Houston, TX, USA. Lauren.a.may10.mil@mail.mil. 2. Department of Radiology, San Antonio Military Medical Center, 3551 Roger Brooke Drive, San Antonio, TX, 78234, USA. Lauren.a.may10.mil@mail.mil. 3. Edward B. Singleton Department of Pediatric Radiology, Texas Children's Hospital, Houston, TX, USA.
Abstract
BACKGROUND: In studies of diagnostic performance that fail to detect a statistically significant difference between compared techniques, investigators often declare evidence of equivalence or similarity without having actually tested that assertion due to incorrect methodology or insufficient statistical power. OBJECTIVE: The purpose of our investigation is to measure the prevalence of unfounded assertions of equivalence or similarity in comparison studies presented at the International Pediatric Radiology (IPR) meeting of 2016 and promote awareness of this fallacy of inference to the pediatric radiology community. MATERIALS AND METHODS: Two pediatric radiologists independently reviewed the methodology and reporting quality of the 194 scientific paper abstracts from the 2016 IPR meeting. All comparison studies were identified and those failing to detect a statistically significance difference and making a claim of equivalence or similarity in the results or conclusion were assessed for a description of the study design type, statistical power and sample size estimator calculation. RESULTS: Of 194 scientific paper abstracts, 112 (58%) were comparison studies. Of these, 36/112 (32%) made unfounded inferences of equivalence or similarity in diagnostic imaging performance. No study had an equivalence or non-inferiority design. No abstract specified the statistical power of the study, and only one abstract acknowledged a small sample size as a limitation in detecting a statistically significant difference. CONCLUSION: Inadequate reporting and unfounded inferences of equivalence or similarity were common in diagnostic performance comparison studies presented at IPR 2016. Failure to recognize these limitations could have adverse consequences by leading to the adoption of unvalidated imaging techniques.
BACKGROUND: In studies of diagnostic performance that fail to detect a statistically significant difference between compared techniques, investigators often declare evidence of equivalence or similarity without having actually tested that assertion due to incorrect methodology or insufficient statistical power. OBJECTIVE: The purpose of our investigation is to measure the prevalence of unfounded assertions of equivalence or similarity in comparison studies presented at the International Pediatric Radiology (IPR) meeting of 2016 and promote awareness of this fallacy of inference to the pediatric radiology community. MATERIALS AND METHODS: Two pediatric radiologists independently reviewed the methodology and reporting quality of the 194 scientific paper abstracts from the 2016 IPR meeting. All comparison studies were identified and those failing to detect a statistically significance difference and making a claim of equivalence or similarity in the results or conclusion were assessed for a description of the study design type, statistical power and sample size estimator calculation. RESULTS: Of 194 scientific paper abstracts, 112 (58%) were comparison studies. Of these, 36/112 (32%) made unfounded inferences of equivalence or similarity in diagnostic imaging performance. No study had an equivalence or non-inferiority design. No abstract specified the statistical power of the study, and only one abstract acknowledged a small sample size as a limitation in detecting a statistically significant difference. CONCLUSION: Inadequate reporting and unfounded inferences of equivalence or similarity were common in diagnostic performance comparison studies presented at IPR 2016. Failure to recognize these limitations could have adverse consequences by leading to the adoption of unvalidated imaging techniques.
Authors: David Moher; Sally Hopewell; Kenneth F Schulz; Victor Montori; Peter C Gøtzsche; P J Devereaux; Diana Elbourne; Matthias Egger; Douglas G Altman Journal: J Clin Epidemiol Date: 2010-03-25 Impact factor: 6.437
Authors: Patrick M Bossuyt; Johannes B Reitsma; David E Bruns; Constantine A Gatsonis; Paul P Glasziou; Les Irwig; Jeroen G Lijmer; David Moher; Drummond Rennie; Henrica C W de Vet; Herbert Y Kressel; Nader Rifai; Robert M Golub; Douglas G Altman; Lotty Hooft; Daniël A Korevaar; Jérémie F Cohen Journal: Radiology Date: 2015-10-28 Impact factor: 11.105