| Literature DB >> 28720141 |
Alex Pollock1, Pauline Campbell2, Ginny Brunton3, Harriet Hunt4, Lise Estcourt5.
Abstract
BACKGROUND: Overviews of systematic reviews are an increasingly popular method of evidence synthesis; there is a lack of clear guidance for completing overviews and a number of methodological challenges. At the UK Cochrane Symposium 2016, methodological challenges of five overviews were explored. Using data from these five overviews, practical implications to support methodological decision making of authors writing protocols for future overviews are proposed.Entities:
Keywords: Challenges; Methods; Overviews; Quality assessment; Synthesis
Mesh:
Year: 2017 PMID: 28720141 PMCID: PMC5516331 DOI: 10.1186/s13643-017-0534-3
Source DB: PubMed Journal: Syst Rev ISSN: 2046-4053
Methodological features of exemplar overviews
| Pollock [ | McClurg [ | Estcourt [ | Hunt [ | Brunton [ | |
|---|---|---|---|---|---|
| Type of overview | Cochrane overview of reviews | Cochrane overview of reviews | Cochrane overview of reviews | Overview of reviews of diagnostic test accuracy | Mixed method overview of reviews |
| Aim/question of overview | To synthesise systematic reviews of interventions to improve upper limb function after stroke | To synthesise Cochrane reviews of conservative interventions for the prevention or treatment of female urinary incontinence | To summarise the evidence in Cochrane reviews of the effectiveness and safety of red cell transfusions for treatment or prevention of complications experienced by people with sickle cell disease | To summarise evidence from studies of the accuracy of diagnostic tests of brief cognitive assessments for identifying dementia in primary care | Overview of workplace health promotion interventions |
| Inclusion criteria: participants, interventions, outcomes | Participants: stroke | Participants: female adults with urinary incontinence | Participants: people with sickle cell disease | Participants: all adults aged 18 years or over recruited from a primary care or general practice population | Participants: any |
| Interventions: any interventions delivered in a workplace setting | |||||
| Interventions: any | Interventions: any conservative intervention (definition provided) | Interventions: red cells transfusions | |||
| Index test: brief cognitive assessments | |||||
| Outcomes: upper limb function, impairment or activity of daily living | Outcomes: any | Outcomes: mortality, sickle-cell related complications, adverse events, quality of life, red cell transfusion requirements, and hospital admissions | Outcomes: diagnostic test result (positive/negative) | Outcomes: healthcare or wellbeing outcomes | |
| Other: diagnostic test accuracy information, other reported test outcomes | |||||
| Inclusion criteria: type of study | Systematic reviews including randomised controlled trials | Cochrane systematic reviews | Cochrane systematic reviews | Systematic reviews of diagnostic test accuracy | Systematic reviews in which the search strategy, inclusion criteria and quality assessment methods are described |
| Published from 1995 to present; in English; | |||||
| Search strategy | CDSR, DARE, PROSPERO | CDSR | CDSR | CDSR, EMBASE, MEDLINE, PsycInfo | MEDLINE, DARE, Cochrane Library, PsycInfo, DOPHER, ASSIA, ABI Inform, Scopus, Business Source Premier |
| Reported line by line for CDSR and DARE | Reported line by line | Reported line by line | Pubmed search reported line by line | ||
| Details of the search strategy are available from the reviewers | |||||
| Updating searches of out of date reviews | – | – | Searches of some included reviews were updated, in order to identify additional primary research studies, and update the original review to include these. The authors determined which reviews to update by using a process of judgement which systematically considered the search dates, presence of included or ongoing trials and relevance of the review topic | – | – |
| Selection of reviews | Two authors independently apply inclusion criteria | Two authors independently apply inclusion criteria | Two authors independently apply inclusion criteria | Two authors independently apply inclusion criteria | Two reviewers independently apply inclusion criteria |
| Data extraction | Characteristics of reviews; results of meta-analyses | Characteristics of reviews; results of meta-analyses | Characteristics of reviews; results of meta-analyses | Characteristics of reviews; results of meta analyses; diagnostic test accuracy information | Characteristics of reviews; results of meta-analyses; data from process evaluations |
| Assessment of quality of reviews | Modified AMSTAR: each of the original 11 AMSTAR questions/criteria were broken down into a series of dichotomous questions, and the answers (yes, no or unclear) to these dichotomous questions were used to determine the answer to the original AMSTAR questions | ROBIS | AMSTAR | AMSTAR and ROBIS | AMSTAR |
| (Additional comments were added to each of the AMSTAR criteria to clarify the decision making process the authors had made) | (Note: employed both tools as there was an absence of guidance relating to the optimal tool for assessing reviews of DTA and these were the available options. Both tools were reported to provide comparable assessment results, but the ROBIS tool was considered to be easier to apply, and arguably more relevant, as this tool could be specifically tailored to reviews of DTA) | (Note: the AMSTAR scores were categorised into low (score 0–3), medium (score 4–7) or high (8–11)) | |||
| (Note: this assessment was completed prior to publication and availability of the ROBIS tool) | |||||
| Assessment of quality of evidence within reviews | GRADE approach, with use of algorithm (details presented and discussed elsewhere [ | GRADE approach, with use of algorithm, supplemented with input from key stakeholders (clinicians) and statisticians | GRADE approach, applied based on assessments provided in the included reviews, with two independent overview authors applying GRADE levels of evidence and any differences resolved through discussion or third party adjudication | No quality assessment was undertaken. Quality assessment measures from included systematic reviews reported narratively. Results showed that 2/13 reviews made no reference to quality assessment; 2/13 reviews referred to STARD criteria [ | No quality assessment was undertaken: the objective of this overview was to provide a ‘systematic map’ of relevant reviews which aimed to identify characteristics associated with effectiveness |
| Data synthesis | Grouped according to GRADE, for intervention versus control, for each outcome explored. Highlighting evidence of beneficial effect, harm or no effect | Data grouped and presented according to type of urinary incontinence. Tabulated summary of systematic review evidence relating to all conservative interventions for urinary incontinence, signposting which systematic reviews address which interventions, with summary of details of the population of participants, comparisons, volume and quality of evidence | Reported using the GRADE approach, with presentation of statistical outcome data as determined by data available in the reviews | Synthesis of results across included studies using sensitivity and specificity measures, presented narratively and within tables | Framework synthesis to synthesise evidence from systematic reviews with other types of evidence (research of stakeholders’ experiences and perspectives; key policy documents). Data relating to moderator analyses to identify ‘successful’ intervention characteristics were tabulated |
| Statistical analysis | No statistical analysis was planned or completed | 1. Network maps, created using Stata software, to illustrate the number of trial and participants within trials | No statistical analysis was planned or completed | No statistical analysis was completed | No statistical analysis was planned or completed |
| (Note: additional analysis is ongoing relating to the assessment of test administration times, analysis of DTA across the most widely used brief cognitive assessments, and comparing diagnostic test performance across all systematic reviews) | |||||
| A visual plot of the effect size relating to the effect of interventions on the primary outcome was produced (reproducing data from the included reviews) | |||||
| 2. Visual plot of effect size for comparisons of conservative intervention versus control, placebo or standard care for outcomes of condition-specific quality of life and symptomatic cure or improvement | |||||
| Summary of key findings | Two key summary tables were produced; each of these summarised the available evidence relating to 19 identified different interventions, highlighting the GRADE of the evidence, the direction of the effect (where GRADE was high or moderate), and key implications for clinical practice and research (see Additional file | A ‘Summary of results’ table was planned, clearly indicating where there is evidence of an effect of conservative interventions, for each relevant intervention comparison and for both primary and secondary outcome of interest (see Additional file | ‘Overview of reviews’ table in a format similar to the ‘Summary of findings’ table as recommended in chapter 22 of the Cochrane Handbook for Systematic Reviews of Interventions [ | A summary of findings table was produced displaying review reference, component study reference, reference standard, index test, threshold, time taken, sample size and key demographic data, target condition, condition prevalence, accuracy data reported and notes. Additional summary tables were produced for comparisons at review level, non-DTA data reported, and individual brief cognitive assessments | Structured summary |
| Executive summary | |||||
| Table of ‘key characteristics identified from effectiveness reviews, views studies and policy documents’ | |||||
| A summary table in which information about each of the included reviews was provided, including the AMSTAR rating of the review. The direction of effect for different intervention types was also summarised, categorising evidence according to whether there was a statistically significant beneficial effect, non-statistically significant beneficial effect, no difference between control and intervention, non-significant detrimental effect or statistically significant detrimental effect (see Additional file | |||||
| Publication status | Published (Cochrane Library) [ | Ongoing (protocol published in Cochrane Library) [ | Ongoing (protocol published in Cochrane Library) [ | Ongoing (protocol published on PROSPERO ref. CRD42015022078) [ | Published (EPPI-Centre website) [ |
Characteristics of included reviews which are reported in tables in overviews
| Reported characteristics from included reviews | Pollock [ | McClurg [ | Estcourt [ | Hunt [ | Brunton [ |
|---|---|---|---|---|---|
| Review reference | √ | √ | √ | √ | √ |
| Date of search | √ | √ | √ | √ | |
| Country (of included studies) | √ | √ | √ | ||
| Objective of review | √ | √ | √ | √ | |
| Types of studies included in review | √ | √ | √ | √ | √ |
| Participants included in review | √ | √ | √ | √ | √ |
| Intervention included in review (including name or brief description) | √ | √ | √ | √ | |
| Reference standard(s) included in the review (including details: assessment tool used, version, assessor details, description) | √ | ||||
| Index test(s) included in review (including brief cognitive assessment details: name, creator, version, description) | √ | ||||
| Test thresholds included within the review | √ | ||||
| Comparisons included in review | √ | √ | √ | √ | |
| Outcomes included in review | √ | √ | √ | √ | |
| Target condition being addressed in the review | √ | ||||
| Population prevalence of condition addressed in the review | √ | ||||
| Setting focus of the review | √ | √ | √ | ||
| Number of studies included in review | √ | √ | √ | √ | √ |
| Number of participants included in review | √ | √ | √ | √ | |
| Other outcomes reported beyond diagnostic test accuracy in the review | √ | ||||
| Statistical data from included reviews reported in overview | |||||
| Effect size | √ | √ | √ | ||
| Confidence intervals | √ | √ | √ | √ | |
| 2 × 2 table components (TP, FP, TN, FN) | √ | ||||
| Sensitivity, specificity | √ | ||||
| Positive Predictive Value, Negative Predictive Value | √ | ||||
| Tests compared, direct/indirect comparisons | √ | ||||
| Heterogeneity (statistical measure) | √ | √ | √ | √ | |
| Direction of effect | √ | √ | √ | √ | √ |
| Results of moderator analyses | √ | ||||
Summary of the perceived key methodological challenges associated with each of the exemplar overviews, a description of what the challenge was, and examples of how this challenge was dealt with within individual overviews
| Key methodological challenges. Dealing with: | Description of challenge and why it was experienced | Complementarity with published literature on overview methods (Ballard [ | Examples of how this challenge was dealt with within our exemplar overviews |
|---|---|---|---|
| a. Overlap between reviews (studies appearing in more than one review) | In overviews which include both Cochrane and non-Cochrane reviews, multiple published reviews were often identified which had similar aims and which included the same or similar trials. | Summary of findings from Ballard [ | Pollock [ |
| • “emerging debate related to (i) overlapping systematic reviews” | |||
| • Approaches to dealing with this challenge include: | |||
|
o Calculation of degree of overlap using the “corrected cover area” (Pieper [ | |||
| o Use “a priori criteria for choosing a single systematic review for inclusion when multiple potential candidates are available ” | |||
| o Use only Cochrane reviews (avoid overlap) | |||
| In overviews which included only Cochrane reviews, some trials were found to be included in more than one review. This was particularly the case for 3 or 4 arm trials, where different arms of the trial were included in different reviews. | • Optimal approach “currently remains unresolved.” | ||
| Hunt [ | |||
| Judgement of complementarity | |||
| Agreement: | |||
| • A range of different approaches were used, and there was no consensus on an | |||
| • optimal approach | |||
| Silence (not raised by Ballard [ | |||
| • Two of our reviews (Hunt [ | |||
| McClurg [ | |||
| If studies appear in more than one review then there are risks of double counting (where results of individual studies are included more than once within meta-analysis). This could be meta-analysis completed by overview authors, or completed by review authors and synthesised within the overview. | |||
| • McClurg 2016 [ | |||
| • Brunton 2016 [ | |||
| Brunton 2016 [ | |||
| b. Reviews are out of date | Included reviews were judged to be out of date. | Summary of findings from Ballard [ | Estcourt [ |
| • “emerging debate related to (iv) updating included systematic reviews” | |||
| • Some guidelines ignore this issue, but approaches to deal with the challenge include: | |||
| o Updating included systematic reviews | |||
| o Searching for secondary and primary literature simultaneously | |||
| • Both approaches add complexity and time to the overview process. | |||
| • There is currently “no way to systematically investigate whether an update in the context of overviews is necessary .” | |||
| Judgement of complementarity | |||
| Agreement | |||
| • Estcourt [ | |||
| Other overview authors ignored this issue, only raising it during discussion. | |||
| c. Definition of “systematic review” | Published review papers were identified which were described as “literature reviews” but did not meet expected methodological standards to be classed as a “systematic review” | Summary of findings from Ballard [ | Overview authors applied clear definitions of what a systematic review was during the stage of selecting relevant reviews of inclusion. Brunton [ |
| • Ballard [ | |||
| Judgement of complementarity | |||
| Silence (not raised by Ballard [ | |||
| This challenge is raised by the overview authors but not identified in literature synthesised by Ballard [ | |||
| d. Assessment of methodological quality of reviews | Assessment of methodological quality using the AMSTAR was found to be challenging due to the multi-faceted nature of the questions within the AMSTAR tool. Assessment of methodological quality using the ROBIS was found to be challenging, with difficulties in reaching agreement between overview authors. There was no suitable tool available for assessing the methodological quality of reviews of diagnostic test accuracy. | Summary of findings from Ballard [ | The AMSTAR [ |
| • “emerging debate related to (iii) evaluating the quality and reporting of included research” | |||
| • (systematic reviews) | Pollock [ | ||
| • There is no consensus on what instrument should be used to assess methodological quality of systematic reviews | |||
| Estcourt [ | |||
| • Many overviews don’t assess methodological quality of systematic reviews (Hartling [ | |||
| McClurg [ | |||
| Judgement of complementarity | |||
| Agreement | |||
| • A range of tools were used. | |||
| Silence (not raised by Ballard [ | |||
| • Challenges in gaining agreement in ROBIS judgements between review authors were identified by overview authors. | |||
| Hunt [ | |||
| e. Quality of reporting within reviews | Methodological assessments, using either the AMSTAR or ROBIS, were limited due to the quality of reporting of the reviews. It was therefore challenging to determine whether the scores provided by AMSTAR or ROBIS reflected the quality of the methods or quality of the reporting. | Summary of findings from Ballard [ | Pollock [ |
| • “emerging debate related to (iii) evaluating the quality and reporting of included research” (systematic reviews) | |||
| • Important to differentiate between methodological quality and reporting quality. | |||
| McClurg [ | |||
| Judgement of complementarity | |||
| Agreement | Hunt [ | ||
| • Overview authors attempted to differentiate between methodological quality and reporting quality. | |||
| Silence (not raised by exemplar overviews) | |||
| • To address reporting issues Ballard [ | |||
| f. Applying GRADE | The GRADE approach has been developed specifically for judgements of quality of evidence during guideline development, and also adopted for judgement of quality of evidence within Cochrane reviews [ | Summary of findings from Ballard [ | Pollock [ |
| • “emerging debate related to (iii) evaluating the quality and reporting of included research” (“quality of the body of evidence across included systematic reviews”) | |||
| • GRADE has been described as an approach for assessing the quality of the body of evidence accrsoss systematic reviews, but there is currently a lack of guidance to ensure appropriate use and interpretation of GRADE when applied in this way | |||
| McClurg [ | |||
| Judgement of complementarity | |||
| Agreement | Estcourt [ | ||
| • Overview authors used GRADE to explore quality of evidence, identifying absence of guidance of how to apply this within the context of an overview. | |||
| g. Potential for publication bias | The potential for publication bias comes from two sources – publication bias relating to the identification and inclusion of relevant reviews, and publication bias relating to the trials which are identified and included within the reviews. | Summary of findings from Ballard [ | Other than being transparent about the potential risks of publication bias, no additional action was undertaken within our overviews. Hunt [ |
| • While the issue of publication/reporting bias is not explicitly raised as a methodological challenge within guidance on overview methods, Ballard [ | |||
| • Where systematic reviews are at high risk of reporting biases then a systematic review (rather than overview) may produce the most precise result | |||
| Judgement of complementarity | |||
| Agreement | |||
| The issue of publication bias is raised by both Ballard [ | |||
| h. Summarising key findings in brief accessible format, suitable for informing decision making | A brief, often single page, summary has insufficient space to address the factors that decision makers require to inform clinical or policy decisions. Decision makers require details of what works, with whom and in what way. | Summary of findings from Ballard [ | Pollock [ |
| • “emerging debate related to (v) synthesising and reporting the result of included systematic reviews” | |||
| • Functions of an overview can be to explore heterogeneity or summarise evidence. | |||
| Brunton [ | |||
| Judgement of complementarity | |||
| Agreement | |||
| • Our overviews aimed to summarise evidence. | Hunt [ | ||
| Within protocols, McClurg 2016 [ |
Summary of common features and differences between the exemplar overviews, and implications for development of protocols for future overviews
| Common features | Differences | Implications for development of protocols for future overviews |
|---|---|---|
| Clearly defined aim or question. | The aim or question can vary considerably depending on the focus of the overview. | 1. Be clear about the purpose of the overview and ensure clear definition of aim or question. |
| 2. Involve key stakeholders at the development stage to ensure the planned overview is relevant and useful. Resources being compiled by Cochrane Training will inform methods of involving people [ | ||
| Includes relevant reviews. | The reviews may be reviews of randomised or quasi-randomised trials, or may be reviews of other study designs, including non-randomised, qualitative or mixed methods studies. | 3. Ensure clear definition and justification of scope of overview. |
| 4. Include a clear definition of what criteria constitute a “systematic review”. These criteria may include; a pre-defined protocol detailing the scope and methods to be employed in conducting the review; a specific and detailed search strategy giving at least one example of the exact terms used within the search; a clear selection process in line with pre-specified criteria for inclusion in the systematic review; assessment of included studies for various types of bias including publication bias; synthesis of research findings in order to address the original research question; discussion of results in relation to existing evidence, limitations of the research and included studies and contribution to the field of study [ | ||
| Some Cochrane overviews only include Cochrane systematic reviews, whilst other Cochrane overviews include any systematic reviews. | ||
| Some overviews include primary research studies in addition to review. Additional searches for new primary studies have been considered when identified systematic review evidence has been judged to be out of date, or there has been known gaps within systematic review evidence. | ||
| 5. Plan, and clearly state within the overview protocol, how multiple overlapping reviews (i.e. systematic reviews which address the same, or similar, research questions, including the same, or similar, primary research studies [ | ||
| 6. Clearly state plans for action to be taken if identified systematic review evidence is out of date or gaps in the evidence (i.e. absence of reviews) are identified) [ | ||
| A clearly defined selection criteria for included reviews. | The parameters or domains which are defined reflect the aims/focus of the overview. | 7. Ensure definition of selection criteria which are relevant to the aims/focus of the overview. |
| A structured search strategy to identify relevant reviews | The search strategy may be limited to databases of reviews (e.g. CDSR, DAREa), or may involve searching a wide variety of electronic databases. | 8. Consider and justify whether overview is limited to Cochrane reviews only or whether non-Cochrane reviews are to be included, with reflection on the available time and resources, and the potential added value of including non-Cochrane reviews and searching additional databases. If non-Cochrane reviews are to be included, state databases to be searched. With the discontinuation of the DARE database, authors should consider which key databases to search, the potential added value of searching additional databases, and the use of a validated search filter for systematic reviews. Authors should remain aware of any new database options, as there is ongoing work to develop comprehensive databases of systematic reviews (e.g. EPISTEMONIKOS [ |
| The search strategy may have date restrictions. The use of date restrictions has been proposed as an approach to searching for literature within limited timescales, justification provided often relates to improvement in the quality and consistency of systematic review evidence over time (including improvements from 1999 following release of the QUality Of Reporting Of Meta-analyses (QUOROM) [ | ||
| Two independent overview authors apply inclusion criteria and select reviews. | – | 9. Plan for two independent overview authors to apply inclusion criteria and select reviews. |
| Systematic extraction and reporting of key characteristics of included reviews. | – | 10. Clearly state data to be extracted, including the characteristics of included reviews to be extracted. |
| Systematic extraction and reporting of results of meta-analyses | Some overviews extract and report key statistical data (e.g. effect size, confidence intervals), whilst others only summarise the direct of effect. | 11. Clearly state the statistical data which is to be extracted, and how this will be summarised. Detail any re-analysis if this is planned. |
| Some overviews carry out some degree of re-analysis of the results presented in the reviews. | ||
| Assessment of methodological quality of included reviews. | The tool which is used to assess methodological quality or risk of bias varies, and may be the AMSTAR or ROBIS tool. The method of applying and reporting the results from these tools can vary. | 12. Select and justify a tool to assess methodological quality of included reviews, according to up-to-date evidence relating to available tools, and with consideration of the need to assess quality of methods and/or risk of bias. Note: The issues raised in relation to the AMSTAR have been debated by other authors [ |
| 13. Plan to have two independent reviewers assessing methodological quality, and state how any disagreements will be addressed. | ||
| 14. Plan to report the results of individual questions or domains (not just a summary score). | ||
| Assessment of quality of evidence within reviews. | If quality of evidence within reviews is assessed the GRADE approach (for reviews of effectiveness) and the QUADAS-2 [ | 15. Consider best evidence and guidance relating to the application of tools to assess quality of evidence within reviews. Be aware of new developments and guidance in this field, and build on methods in previous overviews. |
| Data synthesis and summary of key findings. | While the need for clear, accessible summaries of key findings is common across all overviews, the methods of summarising key findings vary. | 16. Clearly state how key findings will be summarised, with consideration of summarising the findings in relation to the quality of evidence (e.g. GRADE), and the populations, interventions and outcomes which this evidence relates to (for evidence relating to effectiveness), and the audience to whom the overview is of relevance. A number of different templates for summarising the findings of an overview have been proposed, including simple ‘traffic light’ graphics to illustrate evidence of effective, non-effective or detrimental interventions [ |
aDARE is no longer being maintained, meaning that this database will not be up-to-date.