Literature DB >> 33722791

Preferred reporting items for journal and conference abstracts of systematic reviews and meta-analyses of diagnostic test accuracy studies (PRISMA-DTA for Abstracts): checklist, explanation, and elaboration.

Jérémie F Cohen1, Jonathan J Deeks2,3, Lotty Hooft4, Jean-Paul Salameh5,6, Daniël A Korevaar7, Constantine Gatsonis8, Sally Hopewell9, Harriet A Hunt10, Chris J Hyde10, Mariska M Leeflang11, Petra Macaskill12, Trevor A McGrath13, David Moher14, Johannes B Reitsma4, Anne W S Rutjes15, Yemisi Takwoingi2,3, Marcello Tonelli16, Penny Whiting17, Brian H Willis2, Brett Thombs18, Patrick M Bossuyt11, Matthew D F McInnes19.   

Abstract

Entities:  

Year:  2021        PMID: 33722791      PMCID: PMC7957862          DOI: 10.1136/bmj.n265

Source DB:  PubMed          Journal:  BMJ        ISSN: 0959-8138


× No keyword cloud information.
The PRISMA-DTA statement has become an internationally accepted reporting guideline for systematic reviews of diagnostic test accuracy studies PRISMA-DTA for Abstracts is intended to improve the completeness and informativeness of journal and conference abstracts of systematic reviews of diagnostic test accuracy studies PRISMA-DTA for Abstracts includes 12 essential items to report in journal and conference abstracts This article provides the checklist, examples of complete reporting and explanations for each item of the checklist, and abstracts of two reviews that authors can use as examples for their abstracts The abstract is often the only section read by users of biomedical articles.1 On the basis of the abstract, many readers decide whether they will read the full text. The abstract is also critical to people who do not have access to the full text, owing to pay walls or because the article is written in a language they do not understand. Therefore, abstracts should enable a quick assessment of the study’s objectives, purpose, and key design features; present an accurate picture of the validity of the main results; and allow readers to evaluate whether the study can meet their information needs.2 Informative abstracts are also key to enabling effective literature searches in electronic databases, notably in the context of systematic reviews. Several evaluations have shown deficiencies in the reporting of journal and conference abstracts across study designs and research fields.3 4 5 6 The PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) statement was developed to improve the reporting of systematic reviews, primarily for reviews of interventions.7 PRISMA for Abstracts is a checklist for reporting abstracts of systematic reviews.8 Because of the specific methods, terminology, and reporting requirements of diagnostic test accuracy (DTA) studies (table 1), our group developed the PRISMA-DTA checklist, which also includes guidance on abstracts.9 10 PRISMA-DTA for Abstracts includes 12 essential items to report in journal and conference abstracts (table 2). A recent evaluation, however, found that only half of these items were consistently reported.11 This explanation and elaboration document gives examples of complete reporting and explanations for each item of the PRISMA-DTA for Abstracts checklist and is intended to provide a useful resource for authors of DTA reviews.
Table 1

Diagnostic test accuracy terminology

TermExplanation
Index testTest under evaluation in a diagnostic accuracy study. The accuracy (eg, sensitivity and specificity) of the index test is estimated by comparing the results of the index test with those of a reference standard applied to the same participants. Multiple index tests can be evaluated within the same study
Comparative studiesStudies aiming to compare the diagnostic accuracy of two or more index tests
Reference standardThe method or combination of methods used in the study for establishing the presence or absence of the target condition
Target conditionThe disease or condition that the reference standard is expected to detect
Role of the testThe position of the index test relative to other tests in the diagnostic investigation of the same target condition (eg, triage, replacement, add-on, new test)
Intended use of the testWhether the index test is used for diagnosis, screening, staging, monitoring, surveillance, prognosis, or other purposes
SensitivityThe proportion of correctly classified participants among those with the target condition
SpecificityThe proportion of correctly classified participants among those without the target condition
QUADAS-2A tool for use in systematic reviews to assess the risk of bias and concerns about applicability of primary diagnostic accuracy studies
Table 2

PRISMA-DTA for Abstracts checklist

Section and topicItem NoDescription
Title and purpose
Title1Identify the report as a systematic review (+/– meta-analysis) of DTA studies
Objectives2Indicate the research question, including components such as participants, index test, and target conditions
Methods
Eligibility criteria3Include study characteristics used as criteria for eligibility
Information sources4List the key databases searched and the search dates
Risk of bias and applicability5Indicate the methods of assessing risk of bias and applicability
Synthesis of resultsA1Indicate the methods for the data synthesis
Results
Included studies6Indicate the number and type of included studies and the participants and relevant characteristics of the studies (including the reference standard)
Synthesis of results7Include the results for the analysis of diagnostic accuracy, preferably indicating the number of studies and participants. Describe test accuracy including variability; if meta-analysis was done, include summary results and confidence intervals
Discussion
Strengths and limitations9Provide a brief summary of the strengths and limitations of the evidence
Interpretation10Provide a general interpretation of the results and the important implications
Other
Funding11Indicate the primary source of funding for the review
Registration12Provide the registration number and the registry name

Compared with PRISMA for Abstracts, one item was deleted (item 8), and one new item was added (A1).

DTA=diagnostic test accuracy.

Diagnostic test accuracy terminology PRISMA-DTA for Abstracts checklist Compared with PRISMA for Abstracts, one item was deleted (item 8), and one new item was added (A1). DTA=diagnostic test accuracy.

Methods for developing explanation and elaboration document

During the consensus meeting to develop the PRISMA-DTA checklist,12 a first version of PRISMA-DTA for Abstracts was drafted, based on PRISMA for Abstracts and the PRISMA-DTA checklist.8 9 Compared with PRISMA for Abstracts, one item was deleted (item 8), and one new item was added (A1, about synthesis of results). We then circulated the draft PRISMA-DTA for Abstracts checklist among PRISMA-DTA Group members for review and approval. A writing committee (JFC, JJD, LH) drafted this explanation and elaboration document, which was then reviewed, edited, and approved by consensus by PRISMA-DTA Group members. Consistent with other reporting guidelines,8 13 14 we provide examples of complete reporting for each item and explanations clarifying the rationale for the item and how to incorporate it in the abstract. We have edited the examples by spelling out abbreviations. We also present the abstracts of two reviews that comply with the checklist in fewer than 300 words (box 1 and box 2).

Diagnostic accuracy of dual-energy computed tomography (DECT) to differentiate uric acid from non-uric acid calculi: systematic review and meta-analysis

Background: Uric acid stone diagnosis is done primarily with in vitro analysis of stones. Dual-energy CT (DECT) would allow earlier diagnosis and therapy. Objective: To evaluate if DECT, using stone analysis as reference standard, is sufficiently accurate to replace stone analysis for diagnosis of uric acid stones. Methods: Original studies in patients with urolithiasis examined with DECT with stone analysis as the reference standard were eligible for inclusion. MEDLINE (1946–2018), Embase (1947–2018), CENTRAL (August 2018), and multiple urology and radiology conferences were searched. QUADAS-2 was used to assess risk of bias and concerns regarding applicability. Meta-analyses were performed using a bivariate random-effects model. Results: Twenty-one studies (1105 patients, 1442 stones) were included. Fourteen studies (662 patients, 944 stones) were analyzed in the uric acid dominant target condition (majority of stone composition uric acid): summary sensitivity was 0.88 (95% CI 0.79–0.93) and specificity 0.98 (95% CI 0.96–0.99). Thirteen studies (674 patients, 760 stones) were analyzed in the uric acid-containing target condition (< majority of stone composition uric acid): summary sensitivity was 0.82 (95% CI 0.73–0.89) and specificity 0.97 (95% CI 0.94–0.98). Meta-regression identified no significant source of variability in accuracy. Two studies had one or more domains at high risk of bias and there were no concerns regarding applicability. Conclusion: DECT is an accurate replacement test for diagnosis of uric acid calculi in vivo, such that stone analysis might be replaced in the diagnostic pathway. Funding: Ontario Graduate Scholarship (OGS). Registration: CRD42018107398 (Prospero). Word count: 249. Adapted with permission of authors from McGrath TA et al. Eur Radiol 2020;30:2791-01.

Rapid antigen detection tests for group A streptococcus in children with pharyngitis: systematic review and meta-analysis of diagnostic test accuracy studies

Background: Group A streptococcus (GAS) accounts for 20% to 40% of cases of pharyngitis in children; the remaining cases are caused by viruses. Compared with throat culture, rapid antigen detection tests (RADTs) offer diagnosis at the point of care. Objectives: To evaluate the diagnostic accuracy of RADTs for diagnosing GAS in children with pharyngitis. Methods: We searched 8 databases (including MEDLINE and Embase) from 1980 through 2015. We included studies that compared RADT for GAS pharyngitis with throat culture on a blood agar plate in a microbiology laboratory in children in ambulatory care. Quality assessment was carried out using QUADAS-2. We used bivariate meta-analysis to estimate summary sensitivity and specificity, and to investigate variability in accuracy across studies. Results: We included 98 unique studies in the review (116 test evaluations; 101 121 participants). The overall methodological quality of included studies was poor, mainly because many studies were at high risk of bias regarding patient selection and the reference standard used. In our main meta-analysis (105 test evaluations; 58 244 participants; median prevalence of GAS 29.5%), RADT had a summary sensitivity of 85.6% (95% CI 83.3 to 87.6) and a summary specificity of 95.4% (95% CI 94.5 to 96.2). There was substantial variability in sensitivity across studies (range 38.6 to 100%); specificity was more stable (range 54.1 to 100%). Variability in accuracy was not explained by study-level characteristics such as age and clinical severity of participants, and GAS prevalence. Conclusions: Whether or not RADT can be used as a stand-alone test to rule out GAS will depend mainly on the epidemiological context. RADT specificity seems sufficiently high to ensure against unnecessary use of antibiotics. These results should be interpreted with caution because of high risk of bias and variability in sensitivity estimates. Funding: Association Française de Pédiatrie Ambulatoire (AFPA). Registration: CD010502 (Cochrane). Word count: 299. Adapted with permission of authors from Cohen JF et al. Cochrane Database Syst Rev 2016(7):CD010502.

PRISMA-DTA for abstracts checklist, section 1: title and purpose

Item 1: Title

Identify the report as a systematic review (+/− meta-analysis) of DTA studies.

Examples

1a: “Diagnostic accuracy of segmental enhancement inversion for diagnosis of renal oncocytoma at biphasic contrast-enhanced computed tomography: systematic review.”15 1b: “The diagnostic accuracy of serological tests for Lyme borreliosis in Europe: a systematic review and meta-analysis.”16

Explanation

To facilitate identification, the title should describe the article as a “systematic review” and as a “meta-analysis” (examples 1a-b), if appropriate. To clarify the focus of the review, the title should contain the terms “diagnostic” and “accuracy,” thereby differentiating it from other aspects of test evaluation, such as reproducibility, prognostic accuracy, optimal threshold estimation, analytical performance, clinical utility, or cost effectiveness. Alternatively, terms that refer to diagnostic accuracy measures (such as sensitivity, specificity, predictive values, or area under the curve) may be used. The title should also contain the index test, the target condition, and comparisons made between tests, if applicable. Incorporating a description of participants is encouraged.

Item 2: Objectives

Indicate the research question, including components such as participants, index test, and target conditions. 2a: “To assess the diagnostic accuracy of Xpert® MTB/RIF for pulmonary tuberculosis detection, where Xpert® MTB/RIF was used as both an initial test replacing microscopy and an add‐on test following a negative smear microscopy result.”17 2b: “To assess the diagnostic accuracy of magnetic resonance imaging for differentiating stage T1 or lower tumors from stage T2 or higher tumors and to analyse the influence of different imaging protocols in patients with bladder cancer.”18 Abstracts should include the research question for the systematic review so that readers can understand the rationale and relevance for clinical practice. This should reflect the target condition(s) for detection (example 2a) or differentiation (example 2b), index test(s) under evaluation (examples 2a-b), the population for intended use (example 2b), the setting, and the proposed role of the index test(s) (example 2a). Authors may also highlight comparative review questions here.

PRISMA-DTA for abstracts checklist, section 2: methods

Item 3: Eligibility criteria

Include the study characteristics used as criteria for eligibility. 3a: “We included randomised controlled trials, cross-sectional studies, and cohort studies using respiratory specimens that allowed for extraction of data evaluating Xpert® MTB/RIF against the reference standard. We excluded gastric fluid specimens. The reference standard for tuberculosis was culture and for rifampicin resistance was phenotypic culture-based drug susceptibility testing.”17 3b: “We included diagnostic accuracy studies that used computed tomography for diagnosis of fat-poor angiomyolipoma in patients with renal masses, using pathologic examination as the reference standard.”19 A clear description of the systematic review’s eligibility criteria allows readers to judge the applicability of findings. Eligibility criteria should include all components of the review question (item 2) plus the reference standard (examples 3a-b), along with any restrictions on study design, such as excluding studies with healthy controls. In comparative reviews, the authors may restrict studies to those in which participants underwent all tests under comparison. Additional examples of eligibility criteria may include year of publication, language, or publication status (for example, no conference abstracts). Results from older studies sometimes differ from more recent results, and studies published in non-English language journals or only in conference abstracts may report lower accuracy estimates.20 21 22 23 24

Item 4: Information sources

List the key databases searched and the search dates. 4a: “A systematic search of MEDLINE, Embase, The Cochrane Library and Science Citation Index Expanded from January 1994 to October 2014 was performed.”25 4b: “We carried out extensive literature searches including MEDLINE (1980 to 25 August 2011), Embase (1980 to 25 August 2011), BIOSIS via EDINA (1985 to 25 August 2011), CINAHL via OVID (1982 to 25 August 2011), and The Database of Abstracts of Reviews of Effects (the Cochrane Library 2011, Issue 7).”26 The abstract should report the databases searched with the date range or date of last search. This informs readers of the completeness and recency of the search and the likelihood that potentially relevant articles have been missed. Additional efforts made to identify studies (for example, searching reference lists of included studies and published reviews, contacting experts, screening trial registries) are often not reported in the abstract because of word restrictions.

Item 5: Risk of bias and applicability

Indicate the methods of assessing risk of bias and applicability.

Example

5a: “We assessed possible bias and applicability of the studies using the QUADAS-2 tool.”27 Aspects of the design and conduct of included primary DTA studies can raise questions about the validity of their findings and applicability for the review question. These include aspects that increase “risk of bias” (that is, estimates that deviate systematically from the truth) and aspects that lead to “concerns about applicability” (that is, study results that do not directly apply to the review question). Non-blinded readers of index tests, for example, can introduce bias.28 29 Recruiting a highly selected study group or using a prototype version of a test may affect applicability. Systematic reviews should evaluate risk of bias and concerns about applicability; review authors may not be able to describe this in detail in the abstract but can specify the tool or approach used. The QUADAS-2 tool 30 (example 5a) is the most frequently used tool for DTA studies.31

Item A1: Synthesis of results

Indicate the methods for the data synthesis. A1a: “We performed meta‐analyses using the bivariate and hierarchical summary receiver operating characteristic models.”32 A1b: “Variability was assessed by subgroup analyses (dual-energy computed tomography technique and risk of bias) and metaregression using test type and threshold applied as covariates.”33 A1c: “We analysed sensitivity and specificity of included studies narratively as there were insufficient studies to perform a meta‐analysis.”34 Authors are encouraged to report the approach taken to summarise study results, whether narratively or using a statistical model. Results from different approaches may diverge,35 36 and some strategies are more robust than others. Example A1a informs readers about key details of the analysis, such as whether statistical methods account for the hierarchical nature of data and the potential trade-off between sensitivity and specificity across studies. How authors evaluated variability in accuracy estimates may also be relevant (example A1b). If applicable, we encourage authors to report methods used for comparisons of multiple index tests and reasons for not pooling study results (example A1c).

PRISMA-DTA for abstracts checklist, section 3: results

Item 6: Included studies

Indicate the number and type of included studies and the participants and relevant characteristics of the studies (including the reference standard). 6a: “We included 27 unique studies […] involving 9557 participants. Sixteen studies (59%) were performed in low‐ or middle‐income countries.”17 6b: “For the diagnostic accuracy of HBsAg from dried blood spot compared to venous blood, 19 studies were included in a quantitative meta-analysis, and 23 in a narrative review.”37 6c: “Of the 40 studies that met the inclusion criteria, 33 compared rapid diagnostic test and/or enzyme immunoassays against enzyme immunoassays and 7 against nucleic-acid test as reference standards. Thirty studies assessed diagnostic accuracy of 33 brands of rapid diagnostic tests in 23,716 individuals from 23 countries using enzyme immunoassays as the reference standard.”38 6d: “All studies were at high risk of bias for the index test domain because no reported thresholds were prespecified.”39 Authors should report the number of included studies and participants (and, if possible, the number of participants with the target condition) and any other key characteristics (example 6a). Some studies may be included in the qualitative part of the review but not in the quantitative synthesis (example 6b). If the included studies use multiple reference standards, this should be reported (example 6c). This information enables readers to gauge the amount of summarised evidence and its applicability to the review question. Reviews with few included studies and a limited number of participants may produce imprecise accuracy estimates and may not add substantive value compared with the individual studies. Review authors are also invited to summarise their assessment of the quality of evidence (that is, risk of bias and concerns about applicability) and highlight their main source of concern (example 6d).

Item 7: Synthesis of results

Include the results for the analysis of diagnostic accuracy, preferably indicating the number of studies and participants. Describe test accuracy including variability; if meta-analysis was done, include summary results and confidence intervals. 7a: “In the 12 studies with the least biased estimates, sensitivity ranged from 30% to 87% and specificity ranged from 86% to 100%.”40 7b: “As an initial test replacing smear microscopy, Xpert® MTB/RIF pooled sensitivity was 89% [95% Credible Interval (CrI) 85% to 92%] and pooled specificity 99% (95% CrI 98% to 99%), (22 studies, 8998 participants: 2953 confirmed tuberculosis, 6045 non‐tuberculosis). As an add‐on test following a negative smear microscopy result, Xpert®MTB/RIF pooled sensitivity was 67% (95% CrI 60% to 74%) and pooled specificity 99% (95% CrI 98% to 99%; 21 studies, 6950 participants).”17 7c: “For HRP‐2, the meta‐analytical average sensitivity and specificity (95% CI) were 95.0% (93.5% to 96.2%) and 95.2% (93.4% to 99.4%), respectively […], for pLDH, the meta‐analytical average sensitivity and specificity (95% CI) were 93.2% (88.0% to 96.2%) and 98.5% (96.7% to 99.4%), respectively.”41 7d: “Compared to microscopy, the detection of microhaematuria on test strips had the highest sensitivity and specificity (sensitivity 75%, 95% CI 71% to 79%; specificity 87%, 95% CI 84% to 90%; 74 studies, 102,447 participants). For proteinuria, sensitivity was 61% and specificity was 82% (82,113 participants); and for leukocyturia, sensitivity was 58% and specificity 61% (1,532 participants). However, the difference in overall test accuracy between the urine reagent strips for microhaematuria and proteinuria was not found to be different when we compared separate populations (P = 0.25), or when direct comparisons within the same individuals were performed (paired studies; P = 0.21).”42 The authors should provide results for the main index test(s) evaluated in the abstract and, if relevant, thresholds defining index test positivity. If no meta-analysis was done, the abstract should describe accuracy results across included studies—for example, by describing the range of estimates (example 7a). If meta-analysis was done, authors should include summary estimates of accuracy and an expression of statistical imprecision, such as confidence intervals, prediction intervals, or Bayesian credible intervals (examples 7b-d). If space allows, authors should report the number of studies and participants used to generate summary estimates for each index test (example 7b and 7d). Measures of statistical inconsistency (“heterogeneity”) used in intervention reviews (such as I2) are usually not applicable in systematic reviews of DTA studies, and no consensus exists for alternative statistics. As such, the broader term “variability” was used in place of the term “inconsistency” in PRISMA-DTA. Variability of accuracy results could be mentioned in the abstract and may include results of main investigations of reasons for variability, such as subgroup analyses and meta-regression (example 7d).43 If the review aimed to compare tests, these results should be reported, preferably including relative or absolute differences in accuracy, with confidence intervals or tests for statistical significance (example 7d). If sensitivity analysis raised serious concerns about the robustness of the main analyses, this should be mentioned as well.

PRISMA-DTA for abstracts checklist, section 4: discussion

Item 9: Strengths and limitations

Provide a brief summary of the strengths and limitations of the evidence. 9a: “The spectrum of patients was relatively narrow in all studies, sample sizes were small, there was substantial incorporation bias, and blinding procedures were often incomplete.”44 9b: “The value of accuracy estimates is considerably undermined by the small number of included studies, and concerns about risk of bias relating to the index test and the reference standard.”34 9c: “We observed substantial variation in sensitivity and specificity of all tests, which was likely attributable to methodological differences and variations in the clinical characteristics of populations recruited.”45 The abstract should briefly highlight the main strengths and limitations of the review process and the included evidence. Review limitations might include search restrictions (for example, number of databases, language, dates) and lack of independent study selection and data extraction by more than one person, for example. Limitations of included evidence might include risk of bias (examples 9a-b), unavailability of data (examples 9a-b), variability of accuracy estimates (example 9c), imprecision (for example, due to few studies or small sample sizes; examples 9a-b), or low applicability of study findings (for example, due to patient selection within the included studies; example 9a). Such limitations may lead to summary estimates of accuracy that may not reflect the “true” performance of a test or may limit applicability in real world clinical use. Reporting all limitations in an abstract might be impossible, but the authors should mention those that they deem most important.

Item 10: Interpretation

Provide a general interpretation of the results and the important implications. 10a: “Compared with microscopy, Xpert offers better sensitivity for the diagnosis of pulmonary tuberculosis in children and its scale-up will improve access to tuberculosis diagnostics for children. Although Xpert helps to provide rapid confirmation of disease, its sensitivity remains suboptimum compared with culture tests. A negative Xpert result does not rule out tuberculosis. Good clinical acumen is still needed to decide when to start antituberculosis therapy and continued research for better diagnostics is crucial.”46 10b: “It might be too early to recommend its use because of the scarcity of reliable clinical data, heterogeneity in case definitions, and unstable accuracy estimates.”47 10c: “If the point estimates for Type 1 and Type 4 tests are applied to a hypothetical cohort of 1000 patients where 30% of those presenting with symptoms have P. falciparum, Type 1 tests will miss 16 malaria cases, and Type 4 tests will miss 26 malaria cases. The number of people wrongly diagnosed with P. falciparum malaria would be 34 with Type 1 tests, and nine with Type 4 tests.”41 “Spin,” which refers to the reporting of findings in a way that makes test performance seem better than is justified by the study results, is common in abstracts of DTA systematic reviews.48 49 The abstract’s conclusion should summarise the evidence with wording that reflects potential limitations of the review and evidence and, ideally, account for the intended use of the test (example 10a; table 1). If insufficient evidence from well conducted studies exists to allow conclusions to be drawn, this should be made clear (example 10b). If the word count permits, providing readers with the numbers of patients who would be expected to obtain correct and erroneous test results, and the likely consequences, may help with interpretation of test accuracy results and differences between tests (example 10c).

PRISMA-DTA for abstracts checklist, section 5: other

Item 11: Funding

Indicate the primary source of funding for the review. 11a: “Primary funding source: Québec Health Research Fund and BD Diagnostic Systems.”50 11b: “Funding: No external funding.”51 A conference abstract should include the main funding source(s) of the review (example 11a) or state that there was no specific funding (example 11b). Journals may require this to be reported elsewhere. This information enables readers to assess whether financial conflicts of interest occurred if, for instance, the test manufacturer provided funding for the review. Other financial conflicts of interest, such as when the inventors of a test are involved in the review,52 are relevant but not easily conveyed in the abstract. Ideally, whether financial support came from a for-profit company or a public funder should be made clear.

Item 12: Registration

Provide the registration number and the registry name. 12a: “This study was registered with PROSPERO (CRD42018089545).”47 Registration of systematic reviews is increasingly expected.53 54 Registration provides evidence that a review is being undertaken prospectively and provides a record of reviews that have been initiated, which reduces the risk of duplicated efforts and allows interested parties to contact reviewers. It also enables peer reviewers, editors, and readers to compare reported review methods against the registered record.55 As registries such as PROSPERO are typically open access, including the number and name of the registry may provide a useful additional source of information. Alternatives to citing an entry on a register include providing a link to an upload of the review protocol on a publicly available website (such as the Open Science Framework), preprint server (such as medRxiv.org), or journal publication (with the DOI).

Discussion

We developed the PRISMA-DTA for Abstracts checklist and have provided this explanation and elaboration document to help authors to improve the reporting of journal and conference abstracts of systematic reviews of DTA studies. This explanation and elaboration document is a companion to the checklist and the explanation and elaboration for PRISMA-DTA for full text reviews.9 10 It may also be useful as a pedagogical resource for people learning about DTA systematic review abstracts. PRISMA-DTA for Abstracts enriches the body of reporting guidelines for journal and conference abstracts,2 including CONSORT for Abstracts of randomised trials of interventions,56 PRISMA for Abstracts of systematic reviews,8 STROBE for Abstracts of observational studies,57 STARD for Abstracts of primary DTA studies,58 PRIO for Abstracts of overviews of systematic reviews,59 and TRIPOD for Abstracts of multivariable prediction models.60 In addition to supporting authors, these checklists can be used by editors, peer reviewers, and conference organisers to assess the completeness of abstracts submitted for publication or presentation. We also provided illustrative examples of real abstracts that comply with the checklist (box 1 and box 2). An evaluation of adherence to the PRISMA-DTA for Abstracts checklist, based on 100 published DTA reviews (2017-2018), found abstracts to be insufficiently informative.11 Items reported in less than 50% of abstracts included items 2 (participants: 49%), 4 (search dates: 42%), 5 (methods for assessing risk of bias: 38%; methods for assessing applicability: 25%), 6 (characteristics of included studies (including reference standard): 13%), 9 (strengths: 8%; limitations: 26%), 11 (funding: 3%), and 12 (registration: 5%). To ensure that abstracts of systematic reviews of DTA studies are sufficiently informative, we strongly recommend the use of structured abstracts.61 62 Recognising that many journals and conferences have their own abstract formatting requirements, we indicate the information that should be reported without specifying abstract sections. Authors should mention key features only once to make the best use of the limited space available. We encourage journals and organisers of scientific conferences to endorse the use of PRISMA-DTA for Abstracts. This may be done by implementing the checklist into instructions to authors and by inviting peer reviewers to use it when evaluating study reports. The usual 250 word limit used by many journals and conferences may be a barrier to complete reporting. We invite journal editors and conference organisers to consider increasing their word limit to at least 300 words. For example, Radiology and The BMJ now allow abstracts up to 300 and 400 words long, respectively. To enhance dissemination, all PRISMA-DTA for Abstracts material will be freely available on the EQUATOR (www.equator-network.org) and PRISMA (www.prisma-statement.org) websites. We also encourage the introduction of PRISMA-DTA and PRISMA-DTA for Abstracts in teaching programmes focusing on systematic reviews of DTA studies and the importance of transparent reporting of health research. PRISMA-DTA for Abstracts presents a minimum set of reporting criteria that should be reported in abstracts of systematic reviews of DTA studies. When possible, authors should also report other items from the full PRISMA-DTA checklist in the abstract,9 especially those deemed critical to their review question. For conferences that allow the inclusion of figures in the abstract, a chart describing the flow of study inclusion through the review is also welcomed. Other figures may include key forest and summary receiver operating characteristics plots or a test consequence graphic.63 The checklist aims at ensuring complete reporting, but it cannot guarantee that reviews adhere to principles of good research practice and research integrity. Guidance for appropriate methods to conduct systematic reviews of DTA studies can be found elsewhere (for example, the Cochrane Handbook for Systematic Reviews of Diagnostic Test Accuracy).64 The abstract should be a fair and honest summary of the full study report. As in the full text, distorted and selective reporting of findings (“spin”) should be avoided.48 49 Clinical implications should be justified by the results, and an accurate description of limitations should be provided. Abstracts are not a replacement for full text articles in informing clinical practice, policy decisions, or other research. However, they must present an accurate and trustworthy account of the research conducted and reported. The PRISMA-DTA for Abstracts checklist can guide authors in preparing an informative, complete, and fair summary of their review, thus increasing the value of the abstract to the clinical and scientific community.65 For full reports of systematic reviews of DTA studies, authors are encouraged to use the PRISMA-DTA checklist.9 10
  63 in total

1.  Empirical evidence of design-related bias in studies of diagnostic tests.

Authors:  J G Lijmer; B W Mol; S Heisterkamp; G J Bonsel; M H Prins; J H van der Meulen; P M Bossuyt
Journal:  JAMA       Date:  1999-09-15       Impact factor: 56.272

Review 2.  Accuracy of the Papanicolaou test in screening for and follow-up of cervical cytologic abnormalities: a systematic review.

Authors:  K Nanda; D C McCrory; E R Myers; L A Bastian; V Hasselblad; J D Hickey; D B Matchar
Journal:  Ann Intern Med       Date:  2000-05-16       Impact factor: 25.391

3.  Publication Bias: Association of Diagnostic Accuracy in Radiology Conference Abstracts with Full-Text Publication.

Authors:  Lindsay A Cherpak; Daniel A Korevaar; Trevor A McGrath; Wilfred Dang; Daniel Walker; Jean-Paul Salameh; Anahita Dehmoobad Sharifabadi; Matthew D F McInnes
Journal:  Radiology       Date:  2019-05-28       Impact factor: 11.105

4.  Diagnostic accuracy of serum (1,3)-beta-d-glucan for neonatal invasive candidiasis: systematic review and meta-analysis.

Authors:  J F Cohen; A Ouziel; S Matczak; J Brice; R Spijker; O Lortholary; M-E Bougnoux; J Toubiana
Journal:  Clin Microbiol Infect       Date:  2019-09-17       Impact factor: 8.067

5.  A proposal for more informative abstracts of clinical articles. Ad Hoc Working Group for Critical Appraisal of the Medical Literature.

Authors: 
Journal:  Ann Intern Med       Date:  1987-04       Impact factor: 25.391

6.  Reporting Weaknesses in Conference Abstracts of Diagnostic Accuracy Studies in Ophthalmology.

Authors:  Daniël A Korevaar; Jérémie F Cohen; Maurice W J de Ronde; Gianni Virgili; Kay Dickersin; Patrick M M Bossuyt
Journal:  JAMA Ophthalmol       Date:  2015-12       Impact factor: 7.389

Review 7.  Rapid diagnostic tests for diagnosing uncomplicated P. falciparum malaria in endemic countries.

Authors:  Katharine Abba; Jonathan J Deeks; Piero Olliaro; Cho-Min Naing; Sally M Jackson; Yemisi Takwoingi; Sarah Donegan; Paul Garner
Journal:  Cochrane Database Syst Rev       Date:  2011-07-06

Review 8.  Thromboelastography (TEG) and rotational thromboelastometry (ROTEM) for trauma induced coagulopathy in adult trauma patients with bleeding.

Authors:  Harriet Hunt; Simon Stanworth; Nicola Curry; Tom Woolley; Chris Cooper; Obioha Ukoumunne; Zhivko Zhelev; Chris Hyde
Journal:  Cochrane Database Syst Rev       Date:  2015-02-16

Review 9.  Assessing variability in results in systematic reviews of diagnostic studies.

Authors:  Christiana A Naaktgeboren; Eleanor A Ochodo; Wynanda A Van Enst; Joris A H de Groot; Lotty Hooft; Mariska M G Leeflang; Patrick M Bossuyt; Karel G M Moons; Johannes B Reitsma
Journal:  BMC Med Res Methodol       Date:  2016-01-15       Impact factor: 4.615

Review 10.  Recommendations for reporting of systematic reviews and meta-analyses of diagnostic test accuracy: a systematic review.

Authors:  Trevor A McGrath; Mostafa Alabousi; Becky Skidmore; Daniël A Korevaar; Patrick M M Bossuyt; David Moher; Brett Thombs; Matthew D F McInnes
Journal:  Syst Rev       Date:  2017-10-10
View more
  7 in total

1.  Operational definition of complementary, alternative, and integrative medicine derived from a systematic search.

Authors:  Jeremy Y Ng; Tushar Dhawan; Ekaterina Dogadova; Zhala Taghi-Zada; Alexandra Vacca; L Susan Wieland; David Moher
Journal:  BMC Complement Med Ther       Date:  2022-04-12

Review 2.  Virtual Reality (VR) Technology for Treatment of Mental Health Problems during COVID-19: A Systematic Review.

Authors:  Muhammad Hizri Hatta; Hatta Sidi; Chong Siew Koon; Nur Aishah Che Roos; Shalisah Sharip; Farah Deena Abdul Samad; Ong Wan Xi; Srijit Das; Suriati Mohamed Saini
Journal:  Int J Environ Res Public Health       Date:  2022-04-28       Impact factor: 4.614

3.  PRISMA-DTA for Abstracts: a new addition to the toolbox for test accuracy research.

Authors:  Daniël A Korevaar; Patrick M Bossuyt; Matthew D F McInnes; Jérémie F Cohen
Journal:  Diagn Progn Res       Date:  2021-04-02

4.  Guides for the Successful Conduct and Reporting of Systematic Review and Meta-Analysis of Diagnostic Test Accuracy Studies.

Authors:  Seong Ho Park
Journal:  Korean J Radiol       Date:  2022-03       Impact factor: 3.500

5.  A comprehensive search string informed by an operational definition of complementary, alternative, and integrative medicine for systematic bibliographic database search strategies.

Authors:  Jeremy Y Ng; Tushar Dhawan; Ekaterina Dogadova; Zhala Taghi-Zada; Alexandra Vacca; Renee-Gabrielle Fajardo; Hooriya A Masood; Riva Patel; Samira Sunderji; L Susan Wieland; David Moher
Journal:  BMC Complement Med Ther       Date:  2022-07-27

Review 6.  Digital Help for Substance Users (SU): A Systematic Review.

Authors:  Natale Salvatore Bonfiglio; Maria Lidia Mascia; Stefania Cataudella; Maria Pietronilla Penna
Journal:  Int J Environ Res Public Health       Date:  2022-09-08       Impact factor: 4.614

7.  Diagnostic accuracy of rapid point-of-care tests for diagnosis of current SARS-CoV-2 infections in children: a systematic review and meta-analysis.

Authors:  Naomi Fujita-Rohwerder; Lars Beckmann; Yvonne Zens; Arpana Verma
Journal:  BMJ Evid Based Med       Date:  2022-01-18
  7 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.