| Literature DB >> 30314530 |
Carole Lunny1, Sue E Brennan1, Steve McDonald1, Joanne E McKenzie2.
Abstract
BACKGROUND: Overviews of systematic reviews (SRs) attempt to systematically retrieve and summarise the results of multiple systematic reviews. This is the second of two papers from a study aiming to develop a comprehensive evidence map of the methods used in overviews. Our objectives were to (a) develop a framework of methods for conducting, interpreting and reporting overviews (stage I)-the Methods for Overviews of Reviews (MOoR) framework-and (b) to create an evidence map by mapping studies that have evaluated overview methods to the framework (stage II). In the first paper, we reported findings for the four initial steps of an overview (specification of purpose, objectives and scope; eligibility criteria; search methods; data extraction). In this paper, we report the remaining steps: assessing risk of bias; synthesis, presentation and summary of the findings; and assessing certainty of the evidence arising from the overview.Entities:
Keywords: Assessment of risk of bias in systematic reviews; Evaluation of methods; Evidence map; Meta-review; Methodology; Overview; Overview of systematic reviews; Review of reviews; Systematic review methods; Umbrella review
Mesh:
Year: 2018 PMID: 30314530 PMCID: PMC6186052 DOI: 10.1186/s13643-018-0784-8
Source DB: PubMed Journal: Syst Rev ISSN: 2046-4053
Fig. 1Summary of the research reported in each paper
Fig. 2Stages in the development of an evidence map of overview methods
Synthesis, presentation and summary of the findings
| Step Sub-step Methods/approaches | Sources (first author, year) |
|---|---|
| 1.0 Plan the approach to summarising the SR results | |
| 1.1 Determine criteria for selecting SR results/MAs, where SR/MAs include overlapping studies | |
| 1.1.1 Include all SR results/MAs | Caird 2015 [ |
| 1.1.2 Use decision rules or tools (e.g. Jadad tool [ | Caird 2015 [ |
| 1.2 Determine the summary approach | |
| 1.2.1 Describe and/or tabulate the characteristics of the included SRs in terms of PICO elements | Becker 2008 [ |
| 1.2.2 Describe and/or tabulate the results of the included SRs | Becker 2008 [ |
| 1.2.3 Describe and/or tabulate the results of the included primary studies, including new or additional primary studies a | Caird 2015 [ |
| 1.2.4 Summarise and/or tabulate RoB assessments of SRs and primary studies | Becker 2008 [ |
| 1.2.5 Summarise and/or tabulate results from any investigations of statistical heterogeneity (e.g. results from subgroup analyses / meta-regression) within the included SRs | Cooper 2012 [ |
| 1.2.6 Summarise and/or tabulate results from any investigations of reporting biases (e.g. results from statistical tests for funnel plot asymmetry) within the included SRs | Singh 2012 [ |
| 1.2.7 Determine the order of reporting the results in text and tables (e.g. by outcome domain, by effectiveness of interventions)a | Becker 2008 [ |
| 1.2.8 Determine methods for converting or standardising effect metrics (either from primary studies or meta-analyses) to the same scale (e.g. odds ratios to risk ratios)a | Becker 2008 [ |
| 1.2.9 Determine methods to group results of specific outcomes (from either primary studies or MAs) into broader outcome domainsa | Ryan 2009 [ |
| 1.3 Determine graphical approaches to present the resultsa | Becker 2008 [ |
| 2.0 Plan the approach to quantitatively synthesising the SR results | |
| 2.1 Do not conduct a new quantitative synthesis (e.g. because of lack of time or resources) | Salanti 2011 [ |
| 2.2 Specify triggers for when to conduct a new quantitative synthesis | |
| 2.2.1 Need to combine results from multiple MAs (with non-overlapping studies) for the same comparison and outcome | Robinson 2015 [ |
| 2.2.2 Need to incorporate additional primary studies; or, incorporate these studies under certain circumstances | Robinson 2015 [ |
| 2.2.3 Need to apply new meta-analysis methods, fitting a more appropriate meta-analysis method and model, or using a different effect metric | Robinson 2015 [ |
| 2.2.4 Need to limit or expand the MAs into a new MA that meets the population, intervention and comparator elements of the overview | Thomson 2010 [ |
| 2.2.5 Need to undertake a new meta-analysis because of concerns regarding the trustworthiness of the SR/MA results | Robinson 2015 [ |
| 2.2.6 Need to conduct a MA (if possible and makes sense to do so) because the SRs did not undertake MA | Inferred |
| 2.2.7 Need to conduct a MA to reconcile discordant findings of previous SRs | White 2009 [ |
| 2.3 Determine the meta-analysis approach | |
| 2.3.1 Undertake a first-order meta-analysis of effect estimates (meta-analysis of the primary study effect estimates)a | Becker 2008 [ |
| 2.3.2 Undertake a second-order meta-analysis of effect estimates (meta-analysis of meta-analyses) either ignoring the potential correlation across the meta-analysis estimates (arising from the same study included in more than one meta-analysis), or applying an adjustment to account for the potential correlation (e.g. inflating the variance of the meta-analysis) | Caird 2015 [ |
| 2.3.3 Undertake vote counting (e.g. based on direction of effect)a | Becker 2008 [ |
| 2.4 Determine the method to convert effect metrics (either from primary studies or meta-analyses) to the same scalea | Cooper 2012 [ |
| 2.5 Determine the meta-analysis model and estimation methodsa | Cooper 2012 [ |
| 2.6 Determine graphical approachesa | Becker 2008 [ |
| 3.0 Plan to assess heterogeneity | |
| 3.1 Determine summary approaches | |
| 3.1.1 Tabulate results by modifying factors (e.g. study size, quality)a | Caird 2015 [ |
| 3.1.2 Graph results by modifying factorsa | (Caird 2015 [ |
| 3.2 Determine approach to identifying and quantifying heterogeneitya | Cooper 2012 [ |
| 3.3 Determine approach to investigation of modifiers of effect in meta-analyses | |
| 3.3.1 Undertake a first-order subgroup analysis of primary study effect estimatesa | Becker 2008 [ |
| 3.3.2 Undertake a second-order subgroup analysis of meta-analysis effect estimates with moderators categorised at the level of the meta-analysis (e.g. SR quality). Issues of correlation across the meta-analysis estimates may occur (see 2.3.2) | Cooper 2012 [ |
| 3.4 Determine the meta-analysis model and estimation methodsa | Refer to 2.5 |
| 4.0 Plan the assessment of reporting biases | |
| 4.1 Determine non-statistical approaches to assess missing SRs | Pieper 2014d [ |
| 4.2 Determine non-statistical approaches to assess missing primary studies | Bolland 2014 [ |
| 4.3 Determine statistical methods for detecting and examining potential reporting biases from missing primary studies or results within studies, or selectively reported resultsa | Caird 2015 [ |
| 5.0 Plan how to deal with overlap of primary studies included in more than one SR | |
| 5.1 Determine methods for quantifying overlap | Cooper 2012 [ |
| 5.2 Determine how to visually examine and present overlap of the primary studies across SRs | Caird 2015 [ |
| 5.3 Determine methods for dealing with overlap | |
| 5.3.1 Use decision rules, or a tool, to select one (or a subset of) MAs with overlapping studies (see also 1.1.2 above) | Caird 2015 [ |
| 5.3.2 Use statistical approaches to deal with overlap | Cooper 2012 [ |
| 5.3.3 Ignore overlap among primary studies in the included SRs | Cooper 2012 [ |
| 5.3.4 Acknowledge overlap as a limitation | Caird 2015 [ |
| 6.0 Plan how to deal with discordant results, interpretations and conclusions of SRs | |
| 6.1 Determine methods for dealing with or reporting discordance across SRs | |
| 6.1.1 Examine and record discordance among SRs addressing a similar question | Caird 2015 [ |
| 6.1.2 Use decision rules or tools (e.g. Jadad 1997 [ | Bolland 2014 [ |
| 6.2 Determine tabular or graphical approaches to present discordance | Inferred |
JBI Joanna Briggs Institute; MA meta-analyses; PICOs Population (P), intervention (I), comparison (C), outcome (O), and study design (s); PROSPERO International Prospective Register of Systematic Reviews; RCT randomised controlled trial; SRs systematic reviews
aAdaptation of the step from SRs to overviews. No methods evaluation required, but special consideration needs to be given to unique issues that arise in conducting overviews
Data extracted from methods studies evaluating tools for assessing risk of bias in SRs
| Study design | Data extracted |
|---|---|
|
| |
| Study characteristics | First author, year |
| Title | |
| Primary objective | |
| Description of primary methods studies | Name of the included tools or measures |
| Type of assessment (e.g. assessment of reliability, content validity) | |
| Content validity—methods of item generation | |
| Content validity—comprehensiveness | |
| Reliability—description of reliability testing | |
| Tests of validity description of correlation coefficient testing | |
| Other assessment (feasibility, acceptability, piloting) | |
| Risk of bias criteria | Existence of a protocol |
| Method to select the sample of SRs to which the tool/measure was applied | |
| Process for selecting the raters/assessors who applied the tool/measure | |
| Pre-specified hypotheses for testing of validity | |
|
| |
| Study characteristics | First author, year |
| Title | |
| Description of SRs of methods studies | Primary objective |
| Number of included tools | |
| Number of studies reporting on the included tools | |
| Name of the included tools or measures (unnamed tools are identified by first author name and year of publication) | |
| Content validity—reported method of development (e.g. item generation, expert assessment of content) | |
| Reliability—description of reliability testing | |
| Construct validity—description of any hypothesis testing. For example, how assessments from two or more tools relate, whether assessments relate to other factors (e.g. effect estimates or findings) | |
| Other assessment (feasibility, acceptability, piloting) | |
| Risk of bias criteria (using three domains from the ROBIS tool [ | Domain 1—study eligibility criteria: concerns regarding specification of eligibility criteria (low, high or unclear concern) |
| Domain 2—identification and selection of studies: concerns regarding methods used to identify and/or select studies (low, high or unclear concern) | |
| Domain 3—data collection and study appraisal: concerns regarding methods used to collect data and appraise studies (low, high or unclear concern) | |
| Overall judgment: Interpretation addresses all concerns identified in Domains 1–3, relevance of studies was appropriately considered, reviewers avoided emphasising results based on statistical significance. | |
Fig. 3Flowchart of the main search for stages I and II studies
Characteristics of stage I studies and the extent to which each described (two ticks) or mentioned (one tick) methods pertaining to the latter steps in conducting an overview
| Citation | Type of study | Summary description of the article | Latter steps in the conduct of an overview | ||
|---|---|---|---|---|---|
| Assessment of RoB in SRs and primary studies | Synthesis, presentation and summary of the findings | Assessment of the certainty of the evidence arising from the overview | |||
| Baker [ | Article describing methods for overviews | • Describes the usefulness of overviews for decision-makers and summarises some procedural steps to be undertaken | ✓ | ✓ | |
| Becker [ | Guidance for undertaking overviews | • Early guidance providing the structure and procedural steps for the conduct of an overview | ✓✓ | ✓ | ✓✓ |
| Bolland [ | Article describing methods for overviews | • Describes criteria for explaining differences in SR/MAs addressing a similar question with discordant conclusions | ✓ | ✓ | ✓✓ |
| Brunton [ | Study examining methods used in a cohort of overviews | • Presents a tabular method of vote counting (positive effect/negative effect/no change) for each reported outcome | ✓ | ||
| Büchter [ | Study examining methods used in a cohort of overviews | • Describes potential conflicts arising from dual authorship (where an overview author includes one or more SRs they authored) and suggests ‘safeguards’ to protect against potential bias | ✓ | ||
| Caird [ | Article describing methods for overviews | • Describes the methodological challenges in the production of overviews aimed at translating the knowledge to policy makers | ✓✓ | ✓✓ | ✓✓ |
| Chen [ | Study examining methods used in a cohort of overviews | • Identifies possible aims of an overview as being to detect unintended effects, improve the precision of effect estimates or explore heterogeneity of effect across disease groups | ✓✓ | ✓✓ | |
| CMIMG [ | Guidance for undertaking overviews | • Builds on the Cochrane guidance for overviews by Becker 2008 [ | ✓ | ✓ | |
| Cooper [ | Article describing methods for overviews | • Describes steps in the conduct of an overview and methods to address challenges (e.g. dealing with overlap in primary studies across SRs) | ✓✓ | ✓✓ | ✓✓ |
| Crick [ | Article describing methods for overviews | • Examines the feasibility of using harvest plots as compared to summary tables to depict results of MAs in overviews | ✓✓ | ✓ | |
| Flodgren [ | Article describing methods for overviews | • Mentions the issue of missing or inadequately reported data | ✓✓ | ✓ | |
| Foisy [ | Article describing methods for overviews | • Mentions some challenges inherent in defining AMSTAR scoring as inclusion criteria, and inclusion of non-Cochrane reviews alongside Cochrane reviews | ✓ | ||
| Foisy [ | Study examining methods used in a cohort of overviews | • Mentions problems overview authors may encounter when applying GRADE to included SRs without going back to original data, and potential solutions | ✓ | ||
| Foisy [ | Article describing methods for overviews | • Describes differences in AMSTAR scores across Cochrane and non-Cochrane SRs, challenges in using AMSTAR to assess RoB of included SRs, and using the AMSTAR score as inclusion criterion | ✓✓ | ||
| Hartling [ | Study examining methods used in a cohort of overviews | • Describes methodological standards for SRs (PRISMA, MECIR) and their applicability to overviews | ✓✓ | ✓ | ✓✓ |
| Hartling [ | Study examining methods used in a cohort of overviews | • Identifies methodological issues when conducting overviews | ✓ | ✓ | |
| Hartling [ | Article describing methods for overviews | • Briefly defines overviews, mentions the purposes in conducting an overview, and discusses some methodological challenges | ✓ | ||
| Hemming [ | Article describing methods for overviews | • Proposes a Bayesian method for meta-analysis (hierarchical meta-analysis) which uses uninformative priors as a means of pooling effect estimates in an overview | ✓✓ | ||
| Ioannidis [ | Article describing methods for overviews | • Defines ‘umbrella reviews’ as a pre-step to network meta-analysis | ✓ | ||
| Jadad [ | Article describing methods for overviews | • Seminal paper summarising the potential sources of discordance in results in a cohort of MAs, and types of discordance | ✓ | ✓✓ | ✓✓ |
| James [ | Study examining methods used in a cohort of overviews | • Briefly describes several steps in the conduct of overviews | ✓ | ✓ | |
| Joanna Briggs Institute (JBI) [ | Guidance for undertaking overviews | • Provides guidance as to what methods should be used at which step in the conduct of an overview | ✓✓ | ✓✓ | ✓ |
| Kovacs [ | Commentary that discusses methods for overviews | • Mentions methodological challenges of overviews | ✓ | ✓ | |
| Kramer [ | Article describing methods for overviews | • Mentions the challenges encountered when the authors conducted three overviews including dealing with heterogeneity | ✓ | ✓ | |
| Li [ | Article describing methods for overviews | • Presents a pilot reporting/quality checklist | ✓ | ||
| Moja [ | Article describing methods for overviews | • Describes methods to assess discordant finding among MAs based on the Jadad 1997 tool [ | ✓ | ✓✓ | |
| O’Mara [ | Article describing methods for overviews | • Presents a ‘utility’ rating based on the SRs PICO compared to the overviews’ PICO question | ✓ | ||
| Pieper [ | Study examining methods used in a cohort of overviews | • Describes the methods used in overviews | ✓ | ✓ | ✓ |
| Pieper [ | Article describing methods for overviews | • Describes the methods recommended in 8 HTA guideline documents related to overviews | ✓✓ | ✓ | ✓ |
| Pieper [ | Study examining methods used in a cohort of overviews | • Describes the process of searching for additional primary studies in an overview | ✓ | ✓✓ | |
| Pollock [ | Article describing methods for overviews | • Adapts GRADE guidance to assess the certainty of the evidence for a specific overview | ✓ | ✓✓ | |
| Robinson [ | Article describing methods for overviews | • Describes the steps and methods to undertake a complex review that includes multiple SRs, which is similar to methods used in overviews | ✓✓ | ✓✓ | ✓✓ |
| Ryan [ | Article describing methods for overviews | • Presents tabular methods to deal with the preparation of overview evidence | ✓✓ | ✓✓ | ✓✓ |
| Salanti [ | Guidance for undertaking overviews | • Defines overviews as integrating or synthesising (rather than summarising) evidence from SRs | ✓ | ||
| Schmidt [ | Article describing methods for overviews | • Describes statistical methods for second-order meta-analysis | ✓✓ | ||
| Silva [ | Study examining methods used in a cohort of overviews | • Examines a cohort of Cochrane reviews for methods used | ✓ | ✓✓ | |
| Singh [ | Article describing methods for overviews | • Presents a pilot reporting/quality checklist | ✓✓ | ✓ | |
| Smith [ | Article describing methods for overviews | • Describes some steps and challenges in undertaking an overview, namely eligibility criteria, search methods, and RoB/quality assessment | ✓✓ | ✓✓ | ✓✓ |
| Tang [ | Article describing methods for overviews | • Describes statistical methods for second-order MA with examples | ✓✓ | ||
| Thomson [ | Article describing methods for overviews | • Describes some steps in undertaking an overview and some of the challenges in conducting an overview | ✓✓ | ✓✓ | ✓✓ |
| Thomson [ | Study examining methods used in a cohort of overviews | • Describes the process of including trials in overviews | ✓ | ||
| Wagner [ | Article describing methods for overviews | • Presents a quality assessment tool for appraisal of overviews | ✓ | ✓ | |
AHRQ’s EPC Agency for Healthcare Research and Quality’s Evidence-based Practice Center; AMSTAR A Measurement Tool to Assess Systematic Reviews; CMIMG Comparing Multiple Interventions Methods Group; GRADE Grading of Recommendations Assessment, Development, and Evaluation; HTA health technology assessment; JBI Joanna Briggs Institute; MA meta-analysis; MECIR Methodological Expectations of Cochrane Intervention Reviews; PICO Population, Intervention, Comparison, Outcome; PRISMA Preferred Reporting Items for Systematic Reviews and Meta-Analyses; RoB risk of bias; R-AMSTAR revised AMSTAR; SR systematic review
aIndicates a poster presentation
✓✓ Indicates a study describing one or more methods
✓ Indicates a study mentioning one or more methods
Methods and approaches for addressing common scenarios unique to overviews
| Scenario for which authors need to plan | Methods/approaches proposed in the literaturea | |||
|---|---|---|---|---|
| Assessment of RoB in SRs and primary studies (Table | Synthesis, presentation and summary of the findings (Table | Assessment of certainty of the evidence (Table | ||
| 1 | Reviews include | 2.1.1 | 1.1.2, 5.0 | 1.1.1–1.1.5 |
| 2 | Reviews report | 2.1.1, 2.1.2, 2.1.3 | 2.2.1, 2.2.5 | 1.1.1–1.1.5 |
| 3 | Data are | 2.1.1, 2.1.3 | 1.2.9, 2.2.1, 2.2.5 | 1.1.1–1.1.5 |
| 4 | Reviews provide incomplete coverage of the overview question (e.g. missing comparisons, populations) | 2.2.1, 2.2.4 | 1.1.1, 1.1.2, 1.1.5 | |
| 5 | Reviews are not up-to-date | 2.2.2 | 1.1.1, 1.1.2 | |
| 6 | Review methods raise concerns about bias or quality | 2.1.1, 2.1.2, 2.2.3 | 2.2.3, 2.2.5, 4.0 | 1.1.1–1.1.5 |
| 7 | Reviews report | 2.2.7, 6.0 | 1.1.1–1.1.5 | |
aThe methods/approaches could be used in combination and at several steps in the conduct of an overview. When one approach is taken, then another approach may not apply
Assessment of risk of bias in SRs and primary studies
| Step Sub-step Methods/approaches | Sources (first author, year) |
|---|---|
| 1.0 Plan to assess risk of bias (RoB) in the included SRs§ | |
| 1.1 Determine how to assess RoB in the included SRs | |
| 1.1.1 Select an existing RoB assessment tool for SRs | Baker 2014 [ |
| 1.1.2 Adapt an existing RoB tool (e.g. selecting or modifying items for the overview) | CMIMG 2012 [ |
| 1.1.3 Develop a RoB tool customised to the overview | CMIMG 2012 [ |
| 1.1.4 Use existing RoB assessments | Baker 2014 [ |
| 1.1.5 Describe characteristics of included SRs that may be associated with bias or quality without using or developing a tool | Pieper 2014a [ |
| 1.2 Determine how to summarise or score the RoB assessments for SRs | |
| 1.2.1 Report assessment for individual items or domains (with or without rationale for judgements) | Hartling 2012 [ |
| 1.2.2 Summarise assessments across items or domains by using a scoring system§§ | JBI 2014 [ |
| 1.2.3 Summarise assessments across items or domains, then use cut-off scores or thresholds to categorise RoB using qualitative descriptors (e.g. low, moderate or high quality)§§ | Crick 2013 [ |
| 1.3 Determine how to present the RoB assessments for SRs | |
| 1.3.1 Display assessments in table(s) (e.g. overall rating in summary of findings table, and another table with RoB items for each SR) | Aromataris 2015 [ |
| 1.3.2 Display assessments graphically | Crick 2015 [ |
| 1.3.3 Report assessments in text | Aromataris 2015 [ |
| 2.0 Plan how the RoB of primary studies will be assessed or re-assessed | |
| 2.1 Determine how to assess the RoB of the primary studies in the included SRs (and any additional primary studies) | |
| 2.1.1 Report RoB assessment of primary studies from the included SRs, using the approaches specified for data extraction to deal with missing, flawed assessments, or discrepant/discordant assessments of the same primary study (i.e. where two or more SRs assess the same study using different tools or report discordant judgements using the same tool) (See ‘Data extraction’ table in [ | Aromataris 2015 [ |
| 2.1.2 Report RoB assessment of primary studies from the included SRs after performing quality checks to verify that the assessment method has been applied appropriately and consistently across a sample of primary studies | Becker 2008 [ |
| 2.1.3 (Re)-assess RoB of some or all primary studiesa | CMIMG 2012 [ |
| 2.1.4 Don’t report or assess RoB of primary studies | Inferred |
| 2.2 Determine how to summarise the RoB assessments for primary studies | |
| 2.2.1 Report assessment for individual items or domains (with or without rationale for judgements)a | JBI 2014 [ |
| 2.2.2 Summarise assessments across items or domains by using a scoring system§§ | JBI 2014 [ |
| 2.2.3 Summarise assessments across items or domains, then use cut-off scores or thresholds to describe RoB (e.g. low, moderate and high quality)§§ | JBI 2014 [ |
| 2.3 Determine how to present the RoB assessments for primary studies | |
| 2.3.1 Display assessments in table(s) (e.g. overall rating in summary of findings table, and another table with RoB items for each primary study)a | Aromataris 2015 [ |
| 2.3.2 Display assessments graphicallya | Crick 2015 [ |
| 2.3.3 Report assessments in texta | Aromataris 2015 [ |
| 3.0 Plan the process for assessing RoB | |
| 3.1 Determine the number of overview authors required to assess studiesa | |
| 3.1.1 Independent assessment by 2 or more authors | Baker 2014 [ |
| 3.1.2 One author assessment | Inferred |
| 3.1.3 One assessment, 2nd confirmed | Cooper 2012 [ |
| 3.1.4 One assessment, 2nd confirms if uncertainty | Cooper 2012 [ |
| 3.2 Determine if authors (co-)authored one or several of the SRs included in the overview, and if yes, plan safeguards to avoid bias in RoB assessment | Büchter 2011 [ |
AMSTAR A Measurement Tool to Assess Systematic Reviews; CMIMG Comparing Multiple Interventions Methods Group; JBI Joanna Briggs Institute; OQAQ Overview Quality Assessment Questionnaire; RoB risk of bias; ROBIS Risk of Bias In Systematic reviews; SRs systematic reviews
§We refer to ‘risk of bias’ assessment, since assessment of SR or primary study limitations should focus on the potential of those methods to bias findings. However, the terms quality assessment and critical appraisal are common, particularly when referring to the assessment of SR methods, and hence our analysis includes all relevant literature irrespective of terminology
§§As is the case with assessment of RoB in primary studies, concerns have been raised about the validity of presenting a summary score or qualitative descriptors based on scores (e.g. low, moderate, high quality) [12, 17]
aAdaptation of the step from SRs to overviews. No methods evaluation required, but special consideration needs to be given to unique issues that arise in conducting overviews
Assessment of the certainty of the evidence arising from the overview
| Step Sub-step Methods/approaches | Sources (first author, year) |
|---|---|
| 1.0 Plan to assess certainty of the evidence | |
| 1.1 Determine how to assess the certainty of the evidence | |
| 1.1.1 Assess the certainty of the evidence using a method developed for use in overviews | Wagner 2012 [ |
| 1.1.2 Assess the certainty of the evidence using an ad hoc method developed for a specific overview | Bolland 2014 [ |
| 1.1.3 Report assessments of certainty of the evidence from the included SRs, using the approaches specified for data extraction to deal with missing data, flawed or discordant assessments (e.g. where two SRs use different methods to assess certainty of the evidence or report discordant assessments using the same method) (see ‘Data extraction’ table in [ | Becker 2008 [ |
| 1.1.4 Report assessments of certainty of the evidence from the included SRs after performing quality checks on a sample of assessments to verify that the assessment method has been applied appropriately and consistently across SRs | Becker 2008 [ |
| 1.1.5 (Re)-assess the certainty of the evidence using an existing method developed for SRs of primary studies without adapting the method for overviews | Crick 2015 [ |
| 1.1.6 Do not report or assess the certainty of the evidence | Inferred |
| 2.0 Plan the process for assessing certainty | |
| 2.1 Determine the number of overview authors required to assess the certainty of the evidencea | |
| 2.1.1 Independent assessment by 2 or more authors | Baker 2014 [ |
| 2.1.2 One author assesses | Inferred |
| 2.1.4 One assesses, 2nd confirms | Cooper 2012 [ |
| 2.1.5 One assesses, 2nd confirms if the first author is unsure | Cooper 2012 [ |
| 2.2 Determine if authors (co-)authored one or several of the SRs included in the overview, and if yes, plan safeguards to avoid bias in certainty of the evidence assessment | Büchter 2011 [ |
AHRQ Agency for Healthcare Research and Quality; CMIMG Comparing Multiple Interventions Methods Group; GRADE Grading of Recommendations Assessment, Development, and Evaluation; JBI Joanna Briggs Institute; SRs systematic reviews
aAdaptation of the step from SRs to overviews. No methods evaluation required, but special consideration needs to be given to unique issues that arise in conducting overviews
Characteristics of SRs of methods studies and assessment of risk of bias
| Study ID (first author, year) | ||
|---|---|---|
| Pieper 2014a [ | Whiting 2013 [ | |
|
| ||
| Title | Systematic review found AMSTAR, but not R(evised)-AMSTAR, to have good measurement properties | Review of existing quality assessment tools for systematic reviews (Chapter 4) |
| Primary objective | To review all empirical studies evaluating the measurement properties of AMSTAR and R-AMSTAR | To conduct a review of existing tools designed to critically appraise SRs and meta-analyses. |
| Number of included tools | 2 | 40 (5/40 tools targeted areas other than SRs of interventions, for example diagnostic test accuracy or genetic association studies) |
| Number of studies reporting on the included tools | 13 (10 reporting on AMSTAR, 2 on R-AMSTAR, 1 on both) | 43 |
| Name of the included tools or measures (unnamed tools are identified by first author name and year of publication) | AMSTAR [ | Named tools: AMSTAR [ |
| Content validity–reported method of development (e.g. item generation, expert assessment of content) | Not assessed (noted in background that AMSTAR was based on OQAQ and a checklist by Sacks 1997) | Methods of development were reported for 17/40 tools: |
| Reliability—description of reliability testing | Inter-rater reliability (IRR) assessments were reported in 11/13 studies, (9 on AMSTAR, 2 on R-AMSTAR). IRR results were reported for individual items (8 studies), the mean across all items (7 studies), and overall score (6 studies) | Inter-rater reliability assessments were reported for 5/40 tools (most reporting kappa or i |
| Tests of validity—description of correlation coefficient testing | Six studies assessed construct validity examining the correlation between total AMSTAR scale scores (summing ‘yes’ responses) and scores on OQAQ (3 studies), Sack’s list (1 study), R-AMSTAR (1 study), and expert assessment (2 studies) | No tests of validity were reported for any tools (although exploratory factor analysis was used during development of content for AMSTAR) |
| Other assessments (feasibility, acceptability, piloting) | Time taken to complete tool | The SR includes a summary of tool content (items and domains measured), tool structure (e.g. checklist, domain based), and item rating (i.e. response options) |
|
| ||
| Domain 1—study eligibility criteriab | Low | Low |
| Domain 2—identification and selection of studiesb | Low | Low |
| Domain 3—data collection and study appraisalb | High | Low |
| Overall judgementc | Low risk of bias | Low risk of bias |
AMSTAR A Measurement Tool to Assess Systematic Reviews; CASP Critical Appraisal Skills Programme; IRR Inter-rater reliability; OQAQ Overview Quality Assessment Questionnaire; MAC Meta-analysis Appraisal Checklist; NHMRC National Health and Medical Research Council; RAPiD Rapid Appraisal Protocol internet Database; ROBIS Risk of Bias In Systematic reviews; SIGN Scottish Intercollegiate Guidelines Network; SRs systematic reviews
aOQAQ [28] is also referred to as OQAC (Overview Quality Assessment Checklist), and RAPiD [88] is also referred to as RAP (Rapid Appraisal Protocol).
bLevel of concern for each domain judged as low, high or unclear
cOverall judgement is based on: interpretation address all concerns identified in domains 1–3, relevance of studies was appropriately considered, reviewers avoided emphasising results based on statistical significance
Characteristics of primary methods studies and assessment of risk of bias
| Study ID (first author, year) | ||||
|---|---|---|---|---|
| Pieper 2014e [ | Whiting 2016 [ | Parmelli 2011 [ | Pieper 2014b [ | |
|
| ||||
| Title | Impact of choice of quality appraisal tool for systematic reviews in overviews | ROBIS: A new tool to assess risk of bias in systematic reviews was developed | Using AMSTAR to assess the methodological quality of systematic reviews: An external validation study | Systematic review finds overlapping reviews were not mentioned in every other overview |
| Primary objective | To examine reliability, validity and feasibility of four quality appraisal tools in an SR and explore how the choice of tool impacts the findings of the evidence synthesis | To develop ROBIS, a new tool for assessing the risk of bias in systematic reviews (rather than in primary studies) | To measure the reliability, construct validity and feasibility of AMSTAR on a sample of SRs in different medical fields | To develop two measures to quantify the degree of overlap of primary studies across SRs and evaluate the validity of the measures |
| Name of the included tools or measures | AMSTAR [ | ROBIS [ | AMSTAR [ | CA and CCA [ |
| Type of assessment | Assess reliability/ construct validity of the tool | Assess content validity/reliability | Assess reliability/ construct validity of the tool | Construct validity testing of the measures |
| Content validity—methods of item generation | Not applicable—existing tool | Content (domains and items) was based on a reporting standard for SRs (i.e. MECIR [ | Not applicable—existing tool | Not applicable—not a tool |
| Content validity—comprehensiveness | Not applicable—existing tool | Content experts (methodologists, systematic reviewers, guideline developers) reviewed the draft ROBIS tool in a face-to-face meeting and Delphi process | Not applicable—existing tool | Not applicable—not a tool |
| Reliability—description of reliability testing | Inter-rater reliability (agreement) between two review authors who independently applied AMSTAR, OQAQ, RAPiD and AQASR to 32 SRs. A 4-week interval separated assessment with each tool. Agreement was assessed at item level for AMSTAR and OQAQ, and domain level for RAP and AQASR (Cohen’s kappa) | Inter-rater reliability (agreement) between two review authors who independently applied ROBIS to 8 SRs. Agreement was assessed at domain level (% agreement) | Inter-rater reliability (agreement) between two review authors who independently applied AMSTAR to 54 SRs. Agreement was assessed at item level (Cohen’s weighted kappa) | Not applicable |
| Tests of validity—description of correlation coefficient testing | Correlation between summary scores on OQAQ and RAPiD (not done for tools without summary scores). Qualitative assessment of whether assessment of SR quality with different tools altered overall conclusions about strength of association between volume and outcomes (where SR quality was one of four elements used to determine strength) | Not assessed | Correlation between scores on AMSTAR and scores from a similar measure, the OQAQ (Pearson’s rank correlation coefficient, results not reported in abstract) | Correlation between measures (CA, CCA) calculated on a sample of overviews (Kendall τ-b) with each other, and each measure with the number of SRs and number of primary publications. Examined whether the measures were associated with publication source (HTA or journal publication), hypothesizing that HTA reports may have more overlap |
| Other assessment (feasibility, acceptability, piloting) | Time to complete | Piloting involved three workshops on ROBIS where participants piloted the tools and provided feedback | Time to complete | Not reported |
|
| ||||
| Existence of a protocol | Not reported | Not reported | Not reported | Not reported |
| Method to select the sample of SRs to which the tool/measure was applied | Convenience: SRs were included studies in an overview that examined associations between surgery volume and outcomes (not intervention effects) | Convenience: SRs were included studies in an overview, being conducted by authors independent of the developers of ROBIS | Convenience: SRs were in two different medical fields (hypertension, colorectal cancer), and described as a convenience sample but unclear how they were selected | Census: All overviews identified from a literature search of five databases. Handsearching of websites of HTA agencies. Search restricted to articles published between 2009 and 2011. |
| Process for selecting the raters/assessors who applied the tool/measureb | Convenience: Raters were authors of an overview in which AMSTAR was used | Convenience: Raters were authors of an overview in which ROBIS was piloted, and were independent of the tool developers. Unclear how they were recruited | Unclear: No description of how raters were selected | Not applicable |
| Pre-specified hypotheses for testing of validity | No: The expected direction or magnitude of correlation was not specified. ‘The Spearman’s rank correlation coefficient was calculated to compare the CATs [critical appraisal tool]’ | Not applicable: no testing of validity | No: The expected direction or magnitude of correlation was not specified. ‘Construct validity was investigated comparing the two instruments using Pearson’s rank correlation coefficient.’ | Yes: ‘We hypothesized that the CA should have a strong (0.60–0.80) negative correlation with the number of included reviews and, compared to this, a lower negative correlation with the number of included primary publications. In contrast, we assumed that the CCA should have a very weak (0.00–0.20) or weak (0.20–0.40) negative correlation with the number of included reviews and the primary publications.’ |
AMSTAR A Measurement Tool to Assess Systematic Reviews; AQASR Assessing the Quality and Applicability of Systematic Reviews; CA Covered Area; CCA Corrected Covered Area; HTA Health Technology Assessment; MECIR Methodological Expectations of Cochrane Intervention Reviews; OQAQ Overview Quality Assessment Questionnaire; RAPiD Rapid Appraisal Protocol internet Database; RoB risk of bias; ROBIS Risk of Bias In Systematic reviews; SRs systematic reviews
aOQAQ [28] is also referred to as OQAC (Overview Quality Assessment Checklist), and RAPiD [88] is also referred to as RAP (Rapid Appraisal Protocol)
bRoB in relation to any estimates of reliability and validity