Literature DB >> 34249342

Development of a checklist to detect errors in meta-analyses in systematic reviews of interventions: study protocol.

Raju Kanukula1, Matthew Page1, Kerry Dwan2, Simon Turner1, Elizabeth Loder3,4, Evan Mayo-Wilson5, Tianjing Li6,7, Adya Misra8, Steve McDonald1, Andrew Forbes1, Joanne McKenzie1.   

Abstract

Background : Systematic reviews underpin clinical practice and policies that guide healthcare decisions. A core component of many systematic reviews is meta-analysis, which is a statistical synthesis of results across studies. Errors in the conduct and interpretation of meta-analysis can lead to incorrect conclusions regarding the benefits and harms of interventions; and studies have shown that these errors are common. Enabling peer reviewers to better detect errors in meta-analysis through the use of a checklist provides an opportunity for these errors to be rectified before publication. To our knowledge, no such checklist exists. Objective : To develop and evaluate a checklist to detect errors in pairwise meta-analyses in systematic reviews of interventions. Methods : We will undertake a four-step process to develop the checklist. First, we will undertake a systematic review of studies that have evaluated errors in the conduct and interpretation of meta-analysis to generate a bank of items to consider for the checklist. Second, we will undertake a survey of systematic review methodologists and statisticians to seek their views on which items, of the bank of items generated in step 1, are most important to include in the checklist. Third, we will hold a virtual meeting to agree upon which items to include in the checklist. Fourth, before finalising the checklist, we will pilot with editors and peer reviewers of journals. Conclusion : The developed checklist is intended to help journal editors and peer reviewers identify errors in the application and interpretation of meta-analyses in systematic reviews. Fewer errors in the conduct and improved interpretation will lead to more accurate review findings and conclusions to inform clinical practice. Copyright:
© 2021 Kanukula R et al.

Entities:  

Keywords:  checklist; errors; meta-analysis; pair-wise meta-analysis; reporting guideline; statistical issues; synthesis; systematic review

Year:  2021        PMID: 34249342      PMCID: PMC8258702          DOI: 10.12688/f1000research.53034.1

Source DB:  PubMed          Journal:  F1000Res        ISSN: 2046-1402


1. Introduction

Systematic reviews (SRs) frequently underpin clinical practice guidelines and policies that guide healthcare decisions. A core component of many SRs is meta-analysis, a statistical technique used to synthesise study effect estimates from studies addressing similar questions, yielding a quantitative summary. Extensions to meta-analysis (e.g. meta-regression, subgroup analysis) allow for investigation of factors that may explain variation of results across studies. These methods have the potential to provide valuable insights for healthcare decision-making; however, they are reliant on the methods being appropriately applied and interpreted. Many errors can arise when conducting meta-analysis. For example, when meta-analysing continuous outcomes, calculations may be incorrect if standard errors are confused with standard deviations. When data are included from multi-arm trials, there is the risk that participants might be counted more than once when multiple comparisons from these trials are eligible for inclusion in the same meta-analysis. For example, from a three-arm trial of paroxetine, fluoxetine, and placebo, two comparisons would be eligible for a meta-analysis of ‘antidepressants versus placebo’ (i.e., ‘paroxetine versus placebo’ and ‘fluoxetine versus placebo’). When dealing with non-standard randomized trials – such as crossover trials, cluster-randomized trials, or split-body trials – there is a risk that variances of the effect estimates in the meta-analysis do not appropriately account for the correlation in observations induced by these designs. – Such errors can lead to studies receiving the incorrect weight in the meta-analysis with potential consequent impact on the combined estimate of intervention effect and its confidence interval, and other statistics, such as the estimated heterogeneity variance and measures of inconsistency. In some circumstances, these errors will lead to a different interpretation of the findings and review conclusions. Statistical errors have been observed frequently in published SRs. For example, a study including 42 reviews from the Cochrane Cystic Fibrosis and Genetic Disorders Group found that nearly half of the SRs had at least one error (e.g., used standard error instead of standard deviation; calculated standard deviations incorrectly from the standard error given in the report; entered median instead of mean). Another study in which the authors re-extracted the data from two randomly selected trials included in each of 27 meta-analyses, found errors in how the meta-analyst entered data for at least one of the two trials in 17 (63%) of the meta-analyses. Furthermore, some published meta-analyses papers have been retracted because of errors in analyses and error in results and/or conclusions. – Research has shown that errors in the interpretation of statistical analyses in reviews are also common. For example, of 110 SRs indexed in MEDLINE® in 2014, 62 used the random-effects model, but 57 (92%) incorrectly interpreted the meta-analytic effect as the best estimate of a common intervention effect across studies, rather than as the average of the intervention effects across studies. In 42 of the 110 meta-analyses, a subgroup analysis was undertaken, but the findings were not interpreted with respect to a test for interaction in 29/42 (69%), and in 11/42 (26%), a claim of a subgroup difference was made based on a statistically significant effect in one group and not the other. Furthermore, the potential for confounding by other factors as a possible explanation for observed subgroup effects, was not raised in any of the SRs. , Many strategies have been proposed to improve the conduct of meta-analysis (thus lessening the chance of errors occurring) and the interpretation of findings. These include, for example, textbooks, – training on meta-analysis methods, connection with support systems (e.g., Cochrane’s TaskExchange), and the inclusion of statisticians on review teams. However, even with these strategies, errors will still occur. A possible additional strategy is to enable peer reviewers to better detect possible errors in meta-analyses. The peer-review process is regarded as a valuable approach for helping peer reviewers and journal editors to judge the quality, critically appraise and finally accept or reject the submitted manuscripts for publication. Researchers have explored the impact of checklists to guide peer reviewers in assessing the completeness of reporting of submitted manuscripts, and have found some evidence that these are effective. – For example, training early career researchers to use the COBPeer tool (which is an online CONSORT-based peer-review tool assessing nine domains: the eight most important CONSORT domains and a switch in primary outcome(s)) helped them detect inadequate reporting in randomized trials compared to the usual review process. , To our knowledge, no such checklist has been developed to detect statistical errors in meta-analyses. Aim: To develop and evaluate a checklist to detect conduct and interpretation errors in pairwise meta-analyses in systematic reviews of interventions.

2. Defining the concept of statistical errors

The notion of statistical conduct and interpretation errors is not simple. Brown and colleagues defined errors to be “actions or conclusions that are demonstrably and unequivocally incorrect from a logical or epistemological point of view (e.g. … mathematical mistakes, statements not supported by the data, incorrect statistical procedures …)”. In this research, we will consider statistical errors to include those arising from underlying assumptions not being met, incorrect values used in the calculations, application of incorrect statistical methods, and misinterpretation of the results and statistical tests. We plan to initially group errors into categories (Section 3.2.4) and refine and revise these based on the types of errors identified through the systematic review (Section 3.2). Our focus will be on errors where it can be reasonably expected that a trained meta-analyst should have or could have known better, recognising that there is subjectivity in making this determination.

3. Methods

3.1 Contributors

A core team (RK, MJP, KD, SLT, EL, EMW, TL, AM, ABF, JEM) will lead the development of this checklist. The core team will conduct the systematic review, develop survey content and analyse survey responses, draft the checklist, coordinate piloting of the checklist, and decide the final content of the checklist. The core team consists of individuals with experience in meta-analysis methods and SR methodology, contributors of the Cochrane Handbook for Systematic Reviews of Interventions, and editors of medical journals who frequently publish SRs (BMJ, PLOS Medicine, Cochrane Database of Systematic Reviews, American Journal of Public Health, and Systematic Reviews).

3.2 Systematic review

We will conduct an SR of studies evaluating errors in the conduct and interpretation of pairwise meta-analysis, for the purpose of identifying types of errors, their prevalence, and to generate a bank of items to potentially be included in the checklist. 3.2.1 Eligibility criteria Studies will need to meet the following eligibility criteria: Inclusion criteria: Studies evaluating types of errors (and potentially their prevalence) in the conduct and interpretation of meta-analyses (and its extensions, for example, subgroup analysis, sensitivity analysis) in SRs of interventions (irrespective of included study design); Articles presenting a checklist or tool to evaluate the conduct of meta-analyses in SRs. Exclusion criteria: Studies evaluating the methodological or reporting quality or risk of bias in SRs using a tool that does not specifically examine statistical errors (e.g. PRISMA, AMSTAR-2, ROBIS) ; Studies or checklists evaluating errors in statistical analyses in primary study designs (e.g. randomized trials and observational studies). 3.2.2 Search methods We will search MEDLINE, Embase and Scopus from inception to January 2021, without any publication type or language restrictions. The search strategies for MEDLINE and Embase combine subject heading terms and text words related to statistical errors in meta-analyses are presented in Appendix (see Extended data). The search was iteratively developed and tested by an experienced information specialist (SM) using a set of 10 methods articles relevant to the topic. The Scopus search closely replicates the MEDLINE search with respect to included terms and word adjacency but is limited to the following subject areas in Scopus: medicine, nursing, dentistry and health professions. We will also search abstracts of papers and posters presented at Cochrane Colloquia since 2011 (available at https://abstracts.cochrane.org/), including the Global Evidence Summit 2017. The search strategy will be determined by assessing the relative recall of terms from eligible studies identified from searches of MEDLINE, Embase and Scopus. We will screen the reference lists and conduct a cited reference search in Web of Science of included articles and review our personal collections of reports or studies related to statistical issues in meta-analyses. In addition, we will contact organisations that produce SRs (e.g., Cochrane, Campbell Collaboration, National Institute for Health and Care Excellence), and journals that frequently publish SRs to seek any in-house checklists they are willing to share. 3.2.3 Selection of studies Two authors will screen independently all titles and abstracts according to the aforementioned eligibility criteria and retrieve the full text of any potentially relevant articles. The same authors will screen the full texts of retrieved articles. In case of any discrepancies, a senior author will adjudicate and finalise the inclusion or exclusion of any article(s). 3.2.4 Data collection Once we finalise the studies to be included, two authors will collect data independently from each article using a standardised data collection form. For studies evaluating types of errors (and potentially their prevalence), we will collect the following information: corresponding author name, email address, year of publication, journal name, objective(s), focus of error investigation (e.g., multi-arm trials, cross-over trials, cluster randomized trials), type and prevalence of errors, and recommendations provided for conducting meta-analyses. For articles presenting a checklist or tool to evaluate conduct or reporting of meta-analyses in SRs, we will collect the following information: checklist/tool name, method of checklist/tool development, number of items included in the checklist/tool and scope of the checklist/tool. In addition, we will collect all the items and response options pertinent to meta-analysis and its interpretation identified in the tools/checklists, and these will be added to our item bank. Once we have extracted data from all articles, we will review the items in the item bank and remove any duplicate or redundant items. We will then group items into broader categories. For example, those relating to data type (e.g., continuous, binary), rare outcomes (i.e., handling of zero events in one or both arms), design of included studies (e.g. cross-over, cluster, multi-arm, non-randomised [e.g. interrupted time series, cohort]), type of analysis (meta-analysis, subgroup analysis, meta-regression, sensitivity analysis, publication bias analysis) and issues of interpretation. We will use Microsoft Excel 2016 for data management.

3.3 Survey

We will send an invitation to all SR methodologists and statisticians (identified from the Cochrane Methods community, Campbell Methods Coordinating Group, Society for Research Synthesis Methodology and other SR methodologists and statisticians known to the core team members) and SR editors and statistical editors (identified from Cochrane Review Groups, and those supporting other journals that frequently publish SRs) to complete a survey to inform the development of the checklist. The survey will ask respondents to: Provide their views on the most important items from the bank of items (generated from the systematic review) to include in the checklist. We will ask respondents to prioritise items that capture/identify the most common and consequential errors expected to occur in the conduct and interpretation of meta-analyses; Provide their views on specific signals (or ‘red flags’) that might lead them to conduct a more thorough investigation of statistical errors in reviews (e.g., size of the effect for some studies, meta-analysis methods used, I-squared value etc.). We will ask researchers to provide rationale for their responses and to suggest additional items not listed in the survey. We will calculate frequencies of each response option for each item and specific signals. For an item or specific signal to meet consensus for discussion, one of the response options for the item or signal will need to be selected by more than 70% of survey respondents; this threshold was selected according to Sumsion 1998 et al.

3.4 Virtual meeting

Following the survey, the core team will hold a virtual meeting to agree upon which items to include in the statistical errors’ checklist for editors/peer reviewers, which items might trigger further investigation (by a statistical reviewer or the authors) and discuss how best to word each item. Attendees will discuss all items exceeding the 70% threshold in the survey. We will also send the items rated as important by fewer than 70% to meeting attendees prior to the meeting, to provide them with the chance of “rescuing” items for discussion at the meeting. Following the meeting, the core team will draft the checklist and an accompanying guidance document (with examples for each item).

3.5 Piloting

In the first stage of piloting, two reviewers will read the draft checklist and guidance document before independently applying it to a random sample of 20 reviews; 10 from each of two previous methodological studies that collated systematic reviews. Specifically, the first methodological study includes 42 SRs of nutrition research (the ROBUST study) that were published between January 2018 and June 2019, and the second includes 31 SRs of interventions for arthritis, depression or anxiety (the SIM study) that were published between January 2010 and January 2012. The reviewers will record issues on whether the wording of items is ambiguous or difficult to interpret, and those items will be discussed by the core team and improved. In the second stage of piloting, we will invite associate editors and peer reviewers of journals that frequently handle SR submissions to pilot the checklist and provide feedback on its usability. After collating the feedback received from the peer reviewers and editors, we will finalise the checklist and the accompanying guidance document.

3.6 Dissemination and knowledge translation strategy

We plan to publish the developed checklist and guidance document in an open-access journal. We will disseminate the checklist via presentations and workshops at relevant conferences and workshops focused on SR methodology, health technology assessment and evidence-based medicine (e.g., Cochrane Colloquia, Evidence Live, HTAi Annual Meeting, The Global Evidence Summit, G-I-N Conference), and via social media, and in a series of international webinars. We also plan to create a user-friendly, online version of the checklist and guidance document for use by journals that publish SRs and meta-analyses. The developed checklist and related work will be published in open access journals. Associated datasets, data collection forms and analyses not included in any publication will be made publicly available via an online repository.

4. Conclusion

The developed checklist could help journal editors and peer reviewers identify errors in the conduct and interpretation of meta-analyses in SRs. Fewer errors and improved interpretation would lead to more accurate SR findings and conclusions to inform clinical practice. Study status: We have completed searching and screening for sections 3.2.1, 3.2.2 and 3.2.3 for systematic review.

Data availability

Underlying data

No underlying data are associated with this article.

Extended data

Bridges: Search strategy for systematic review of studies evaluating errors in the conduct and interpretation of pairwise meta-analysis. Monash University. DOI: https://doi.org/10.26180/14446293.v1. This project contains the following extended data: Search strategy for systematic review of studies evaluating errors in the conduct and interpretation of pairwise meta-analysis Data are available under the terms of the Creative Commons Attribution 4.0 International license (CC-BY 4.0). This work will provide an important contribution to the field, highlighting typical errors that can occur in pairwise meta-analyses of intervention effects. One question I have is the audience of the checklist. The study protocol is situated in meta-analysis for health and medical interventions. The search strategy for relevant systematic reviews for stage one of the process focuses on the health and medical literature, with only the Campbell Collaboration as a source for meta-analyses in the social sciences. If the intent of the checklist is to apply more broadly to social interventions rather than just health and medicine, then the search strategy needs to include a broader set of databases. There are relevant efforts in psychology for example (see Maasen et al. (2020 )) that could also inform the checklist. I understand if the authors wish to focus this effort only on health and medicine. In the social sciences, similar issues to those that are outlined in the protocol occur. In addition, many meta-analyses include a large number of studies, and thus key issues center on dependent effect sizes within studies and interpretation of effect size models for heterogeneity. If the authors intend the checklist to have a broader application, then these are a few issues that should be considered in the development of the checklist. Is the study design appropriate for the research question? Yes Is the rationale for, and objectives of, the study clearly described? Yes Are sufficient details of the methods provided to allow replication by others? Yes Are the datasets clearly presented in a useable and accessible format? Yes Reviewer Expertise: meta-analysis methods I confirm that I have read this submission and believe that I have an appropriate level of expertise to confirm that it is of an acceptable scientific standard. Authors have well prepared their systematic review protocol with clarity. I have following minor suggestions to improve the quality of this protocol. From inclusion criteria it is clear that authors will include SRs of interventions only. However, if possible, authors should elaborate on not including SR of proportions or SR of observational studies or SR of biomarkers etc. Each set of SRs will require a different type of statistical skills to perform meta-analysis. It also important to mention about whether only pharmacological intervention SRs or non- pharmacological interventions such as educational interventions etc. will also be included in this methodological systematic review. Did authors think that a statistician included in the author list will have any impact on error as it is possible that SRs with statistician as co-author may have less statistical errors? I think the authors have already thought about this point, however, appreciate authors reflection on this point and if they add few words in the protocol. If I am not wrong, the planned survey and consensus meeting will follow Delphi survey methodology. I am wondering that authors did not mention this word explicitly or they are following some other equally useful methodology. I believe that authors have well refined methodology to capture statistical errors. However, in many SRs, authors wrongly choose a meta-analysis model (i.e., fixed-effect vs. random-effect) to pool study estimates and possible other statistics right. How would authors capture such errors? Please elaborate if I have missed this point. Is the study design appropriate for the research question? Yes Is the rationale for, and objectives of, the study clearly described? Yes Are sufficient details of the methods provided to allow replication by others? Yes Are the datasets clearly presented in a useable and accessible format? Yes Reviewer Expertise: Meta-analysis, Statistical analysis I confirm that I have read this submission and believe that I have an appropriate level of expertise to confirm that it is of an acceptable scientific standard.
  20 in total

1.  High prevalence but low impact of data extraction and reporting errors were found in Cochrane systematic reviews.

Authors:  Ashley P Jones; Tracey Remmington; Paula R Williamson; Deborah Ashby; Rosalind L Smyth
Journal:  J Clin Epidemiol       Date:  2005-04-18       Impact factor: 6.437

2.  Unsubstantiated claims of large effects of placebo on pain: serious errors in meta-analysis of placebo analgesia mechanism studies.

Authors:  Asbjørn Hróbjartsson; Peter C Gøtzsche
Journal:  J Clin Epidemiol       Date:  2006-04       Impact factor: 6.437

3.  Guarding the guardians: research on editorial peer review. Selected proceedings from the First International Congress on Peer Review in Biomedical Publication. May 10-12, 1989, Chicago, Ill.

Authors: 
Journal:  JAMA       Date:  1990-03-09       Impact factor: 56.272

4.  Issues with data and analyses: Errors, underlying themes, and potential solutions.

Authors:  Andrew W Brown; Kathryn A Kaiser; David B Allison
Journal:  Proc Natl Acad Sci U S A       Date:  2018-03-13       Impact factor: 11.205

5.  Flaws in the application and interpretation of statistical analyses in systematic reviews of therapeutic interventions were common: a cross-sectional analysis.

Authors:  Matthew J Page; Douglas G Altman; Joanne E McKenzie; Larissa Shamseer; Nadera Ahmadzai; Dianna Wolfe; Fatemeh Yazdi; Ferrán Catalá-López; Andrea C Tricco; David Moher
Journal:  J Clin Epidemiol       Date:  2017-12-02       Impact factor: 6.437

6.  Retraction Note: Application of 4% chlorhexidine to the umbilical cord stump of newborn infants in lower income countries: a systematic review and meta-analysis.

Authors:  Aklilu Abrham Roba; Maleda Tefera; Teshager Worku; Tamirat Tesfaye Dasa; Abiy Seifu Estifanos; Nega Assefa
Journal:  Matern Health Neonatol Perinatol       Date:  2020-06-16

7.  Cluster Randomised Trials in Cochrane Reviews: Evaluation of Methodological and Reporting Practice.

Authors:  Marty Richardson; Paul Garner; Sarah Donegan
Journal:  PLoS One       Date:  2016-03-16       Impact factor: 3.240

8.  Reproducibility of individual effect sizes in meta-analyses in psychology.

Authors:  Esther Maassen; Marcel A L M van Assen; Michèle B Nuijten; Anton Olsson-Collentine; Jelte M Wicherts
Journal:  PLoS One       Date:  2020-05-27       Impact factor: 3.240

9.  A protocol of a cross-sectional study evaluating an online tool for early career peer reviewers assessing reports of randomised controlled trials.

Authors:  Anthony Chauvin; David Moher; Doug Altman; David L Schriger; Sabina Alam; Sally Hopewell; Daniel R Shanahan; Alessandro Recchioni; Philippe Ravaud; Isabelle Boutron
Journal:  BMJ Open       Date:  2017-09-15       Impact factor: 2.692

10.  Impact of a short version of the CONSORT checklist for peer reviewers to improve the reporting of randomised controlled trials published in biomedical journals: study protocol for a randomised controlled trial.

Authors:  Benjamin Speich; Sara Schroter; Matthias Briel; David Moher; Iratxe Puebla; Alejandra Clark; Michael Maia Schlüssel; Philippe Ravaud; Isabelle Boutron; Sally Hopewell
Journal:  BMJ Open       Date:  2020-03-19       Impact factor: 2.692

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.