BACKGROUND: Poor reporting of diagnostic accuracy studies impedes an objective appraisal of the clinical performance of diagnostic tests. The Standards for Reporting of Diagnostic Accuracy Studies (STARD) statement, first published in 2003, aims to improve the reporting quality of such studies. OBJECTIVE: To investigate to which extent published diagnostic accuracy studies adhere to the 25-item STARD checklist, whether the reporting quality has improved after STARD's launch and whether there are any factors associated with adherence. STUDY SELECTION: We performed a systematic review and searched MEDLINE, EMBASE and the Methodology Register of the Cochrane Library for studies that primarily aimed to examine the reporting quality of articles on diagnostic accuracy studies in humans by evaluating adherence to STARD. Study selection was performed in duplicate; data were extracted by one author and verified by the second author. FINDINGS: We included 16 studies, analysing 1496 articles in total. Three studies investigated adherence in a general sample of diagnostic accuracy studies; the others did so in a specific field of research. The overall mean number of items reported varied from 9.1 to 14.3 between 13 evaluations that evaluated all 25 STARD items. Six studies quantitatively compared post-STARD with pre-STARD articles. Combining these results in a random-effects meta-analysis revealed a modest but significant increase in adherence after STARD's introduction (mean difference 1.41 items (95% CI 0.65 to 2.18)). CONCLUSIONS: The reporting quality of diagnostic accuracy studies was consistently moderate, at least through halfway the 2000s. Our results suggest a small improvement in the years after the introduction of STARD. Adherence to STARD should be further promoted among researchers, editors and peer reviewers.
BACKGROUND: Poor reporting of diagnostic accuracy studies impedes an objective appraisal of the clinical performance of diagnostic tests. The Standards for Reporting of Diagnostic Accuracy Studies (STARD) statement, first published in 2003, aims to improve the reporting quality of such studies. OBJECTIVE: To investigate to which extent published diagnostic accuracy studies adhere to the 25-item STARD checklist, whether the reporting quality has improved after STARD's launch and whether there are any factors associated with adherence. STUDY SELECTION: We performed a systematic review and searched MEDLINE, EMBASE and the Methodology Register of the Cochrane Library for studies that primarily aimed to examine the reporting quality of articles on diagnostic accuracy studies in humans by evaluating adherence to STARD. Study selection was performed in duplicate; data were extracted by one author and verified by the second author. FINDINGS: We included 16 studies, analysing 1496 articles in total. Three studies investigated adherence in a general sample of diagnostic accuracy studies; the others did so in a specific field of research. The overall mean number of items reported varied from 9.1 to 14.3 between 13 evaluations that evaluated all 25 STARD items. Six studies quantitatively compared post-STARD with pre-STARD articles. Combining these results in a random-effects meta-analysis revealed a modest but significant increase in adherence after STARD's introduction (mean difference 1.41 items (95% CI 0.65 to 2.18)). CONCLUSIONS: The reporting quality of diagnostic accuracy studies was consistently moderate, at least through halfway the 2000s. Our results suggest a small improvement in the years after the introduction of STARD. Adherence to STARD should be further promoted among researchers, editors and peer reviewers.
Authors: Ji Eun Park; Donghyun Kim; Ho Sung Kim; Seo Young Park; Jung Youn Kim; Se Jin Cho; Jae Ho Shin; Jeong Hoon Kim Journal: Eur Radiol Date: 2019-07-26 Impact factor: 5.315
Authors: Silvia Galliazzo; Olga Nigro; Lorenza Bertù; Luigina Guasti; Anna Maria Grandi; Walter Ageno; Francesco Dentali Journal: Intern Emerg Med Date: 2018-03-05 Impact factor: 3.397
Authors: Daniël A Korevaar; Jérémie F Cohen; Maurice W J de Ronde; Gianni Virgili; Kay Dickersin; Patrick M M Bossuyt Journal: JAMA Ophthalmol Date: 2015-12 Impact factor: 7.389
Authors: Viknesh Sounderajah; Hutan Ashrafian; Robert M Golub; Shravya Shetty; Jeffrey De Fauw; Lotty Hooft; Karel Moons; Gary Collins; David Moher; Patrick M Bossuyt; Ara Darzi; Alan Karthikesalingam; Alastair K Denniston; Bilal Akhter Mateen; Daniel Ting; Darren Treanor; Dominic King; Felix Greaves; Jonathan Godwin; Jonathan Pearson-Stuttard; Leanne Harling; Matthew McInnes; Nader Rifai; Nenad Tomasev; Pasha Normahani; Penny Whiting; Ravi Aggarwal; Sebastian Vollmer; Sheraz R Markar; Trishan Panch; Xiaoxuan Liu Journal: BMJ Open Date: 2021-06-28 Impact factor: 2.692