| Literature DB >> 28137831 |
Jérémie F Cohen1,2, Daniël A Korevaar1, Douglas G Altman3, David E Bruns4, Constantine A Gatsonis5, Lotty Hooft6, Les Irwig7, Deborah Levine8,9, Johannes B Reitsma10, Henrica C W de Vet11, Patrick M M Bossuyt1.
Abstract
Diagnostic accuracy studies are, like other clinical studies, at risk of bias due to shortcomings in design and conduct, and the results of a diagnostic accuracy study may not apply to other patient groups and settings. Readers of study reports need to be informed about study design and conduct, in sufficient detail to judge the trustworthiness and applicability of the study findings. The STARD statement (Standards for Reporting of Diagnostic Accuracy Studies) was developed to improve the completeness and transparency of reports of diagnostic accuracy studies. STARD contains a list of essential items that can be used as a checklist, by authors, reviewers and other readers, to ensure that a report of a diagnostic accuracy study contains the necessary information. STARD was recently updated. All updated STARD materials, including the checklist, are available at http://www.equator-network.org/reporting-guidelines/stard Here, we present the STARD 2015 explanation and elaboration document. Through commented examples of appropriate reporting, we clarify the rationale for each of the 30 items on the STARD 2015 checklist, and describe what is expected from authors in developing sufficiently informative study reports. Published by the BMJ Publishing Group Limited. For permission to use (where not already granted under a licence) please go to http://www.bmj.com/company/products-services/rights-and-licensing/.Entities:
Keywords: Diagnostic accuracy; Medical publishing; Peer review; Reporting quality; Research waste; Sensitivity and specificity
Mesh:
Year: 2016 PMID: 28137831 PMCID: PMC5128957 DOI: 10.1136/bmjopen-2016-012799
Source DB: PubMed Journal: BMJ Open ISSN: 2044-6055 Impact factor: 2.692
Key STARD terminology
| Term | Explanation |
|---|---|
| Medical test | Any method for collecting additional information about the current or future health status of a patient |
| Index test | The test under evaluation |
| Target condition | The disease or condition that the index test is expected to detect |
| Clinical reference standard | The best available method for establishing the presence or absence of the target condition. A gold standard would be an error-free reference standard |
| Sensitivity | Proportion of those with the target condition who test positive with the index test |
| Specificity | Proportion of those without the target condition who test negative with the index test |
| Intended use of the test | Whether the index test is used for diagnosis, screening, staging, monitoring, surveillance, prediction, prognosis or other reasons |
| Role of the test | The position of the index test relative to other tests for the same condition (eg, triage, replacement, add-on, new test) |
| Indeterminate results | Results that are neither positive or negative |
The STARD 2015 list10
| Section and topic | No | Item |
|---|---|---|
| 1 | Identification as a study of diagnostic accuracy using at least one measure of accuracy (such as sensitivity, specificity, predictive values or AUC) | |
| 2 | Structured summary of study design, methods, results and conclusions (for specific guidance, see STARD for Abstracts) | |
| 3 | Scientific and clinical background, including the intended use and clinical role of the index test | |
| 4 | Study objectives and hypotheses | |
| Study design | 5 | Whether data collection was planned before the index test and reference standard were performed (prospective study) or after (retrospective study) |
| Participants | 6 | Eligibility criteria |
| 7 | On what basis potentially eligible participants were identified (such as symptoms, results from previous tests, inclusion in registry) | |
| 8 | Where and when potentially eligible participants were identified (setting, location and dates) | |
| 9 | Whether participants formed a consecutive, random or convenience series | |
| Test methods | 10a | Index test, in sufficient detail to allow replication |
| 10b | Reference standard, in sufficient detail to allow replication | |
| 11 | Rationale for choosing the reference standard (if alternatives exist) | |
| 12a | Definition of and rationale for test positivity cut-offs or result categories of the index test, distinguishing prespecified from exploratory | |
| 12b | Definition of and rationale for test positivity cut-offs or result categories of the reference standard, distinguishing prespecified from exploratory | |
| 13a | Whether clinical information and reference standard results were available to the performers or readers of the index test | |
| 13b | Whether clinical information and index test results were available to the assessors of the reference standard | |
| Analysis | 14 | Methods for estimating or comparing measures of diagnostic accuracy |
| 15 | How indeterminate index test or reference standard results were handled | |
| 16 | How missing data on the index test and reference standard were handled | |
| 17 | Any analyses of variability in diagnostic accuracy, distinguishing prespecified from exploratory | |
| 18 | Intended sample size and how it was determined | |
| Participants | 19 | Flow of participants, using a diagram |
| 20 | Baseline demographic and clinical characteristics of participants | |
| 21a | Distribution of severity of disease in those with the target condition | |
| 21b | Distribution of alternative diagnoses in those without the target condition | |
| 22 | Time interval and any clinical interventions between index test and reference standard | |
| Test results | 23 | Cross tabulation of the index test results (or their distribution) by the results of the reference standard |
| 24 | Estimates of diagnostic accuracy and their precision (such as 95% CIs) | |
| 25 | Any adverse events from performing the index test or the reference standard | |
| 26 | Study limitations, including sources of potential bias, statistical uncertainty and generalisability | |
| 27 | Implications for practice, including the intended use and clinical role of the index test | |
| 28 | Registration number and name of registry | |
| 29 | Where the full study protocol can be accessed | |
| 30 | Sources of funding and other support; role of funders | |
Figure 1Example of flow diagram from a study evaluating the accuracy of faecal immunochemical testing for diagnosis of advanced colorectal neoplasia (adapted from Collins et al,79 with permission).
Figure 2STARD 2015 flow diagram.
Example of baseline demographic and clinical characteristics of participants in a study evaluating the accuracy of point-of-care fecal tests for diagnosis of organic bowel disease (adapted from Kok et al,87 with permission)
| Patient characteristics | n (%) |
|---|---|
| Geographic region of residency in the Netherlands | |
| Central (Gelderse Vallei) | 257 (66.6) |
| South (Oostelijke Mijnstreek) | 129 (33.4) |
| Median age, years (range) | 60 (18–91) |
| Women | 211 (54.7) |
| Presenting symptoms | |
| Rectal blood loss | 141 (37.7) |
| Abdominal pain | 267 (70.6) |
| Median duration of abdominal pain (range) | 150 days (1 day to 30 years) |
| Persistent diarrhoea | 40 (16.9) |
| Diarrhoea | 131 (37.2) |
| Fever | 40 (11.0) |
| Weight loss | 62 (17.1) |
| Bloating | 195 (53.6) |
| Constipation | 169 (46.6) |
| Physical examination | |
| Pain at palpation | 117 (46.8) |
| Palpable abdominal mass | 12 (3.0) |
| Palpable rectal mass | 1 (0.3) |
Example of contingency table from a study evaluating the accuracy of pain over speed bumps for diagnosis of appendicitis (adapted from Ashdown et al,95 with permission)
| Appendicitis | |||
|---|---|---|---|
| Pain over speed bumps | Positive | Negative | Total |
| Positive | 33 | 21 | 54 |
| Negative | 1 | 9 | 10 |
| Total | 34 | 30 | 64 |