Literature DB >> 29540417

Tools for assessing risk of reporting biases in studies and syntheses of studies: a systematic review.

Matthew J Page1,2, Joanne E McKenzie1, Julian P T Higgins2.   

Abstract

BACKGROUND: Several scales, checklists and domain-based tools for assessing risk of reporting biases exist, but it is unclear how much they vary in content and guidance. We conducted a systematic review of the content and measurement properties of such tools.
METHODS: We searched for potentially relevant articles in Ovid MEDLINE, Ovid Embase, Ovid PsycINFO and Google Scholar from inception to February 2017. One author screened all titles, abstracts and full text articles, and collected data on tool characteristics.
RESULTS: We identified 18 tools that include an assessment of the risk of reporting bias. Tools varied in regard to the type of reporting bias assessed (eg, bias due to selective publication, bias due to selective non-reporting), and the level of assessment (eg, for the study as a whole, a particular result within a study or a particular synthesis of studies). Various criteria are used across tools to designate a synthesis as being at 'high' risk of bias due to selective publication (eg, evidence of funnel plot asymmetry, use of non-comprehensive searches). However, the relative weight assigned to each criterion in the overall judgement is unclear for most of these tools. Tools for assessing risk of bias due to selective non-reporting guide users to assess a study, or an outcome within a study, as 'high' risk of bias if no results are reported for an outcome. However, assessing the corresponding risk of bias in a synthesis that is missing the non-reported outcomes is outside the scope of most of these tools. Inter-rater agreement estimates were available for five tools.
CONCLUSION: There are several limitations of existing tools for assessing risk of reporting biases, in terms of their scope, guidance for reaching risk of bias judgements and measurement properties. Development and evaluation of a new, comprehensive tool could help overcome present limitations. © Article author(s) (or their employer(s) unless otherwise stated in the text of the article) 2018. All rights reserved. No commercial use is permitted unless otherwise expressly granted.

Entities:  

Keywords:  bias (epidemiology); checklist; publication bias; review literature as topic

Mesh:

Year:  2018        PMID: 29540417      PMCID: PMC5857645          DOI: 10.1136/bmjopen-2017-019703

Source DB:  PubMed          Journal:  BMJ Open        ISSN: 2044-6055            Impact factor:   2.692


Tools for assessing risk of reporting biases, and studies evaluating their measurement properties, were identified by searching several relevant databases using a search string developed in conjunction with an information specialist. Detailed information on the content and measurement properties of existing tools was collected, providing readers with pertinent information to help decide which tools to use in evidence syntheses. Screening of articles and data collection were performed by one author only, so it is possible that some relevant articles were missed, or that errors in data collection were made. The search of grey literature was not comprehensive, so it is possible that there are other tools for assessing risk of reporting biases, and unpublished studies evaluating measurement properties, that were omitted from this review.

Background

The credibility of evidence syntheses can be compromised by reporting biases, which arise when dissemination of research findings is influenced by the nature of the results.1 For example, there may be bias due to selective publication, where a study is only published if the findings are considered interesting (also known as publication bias).2 In addition, bias due to selective non-reporting may occur, where findings (eg, estimates of intervention efficacy or an association between exposure and outcome) that are statistically non-significant are not reported or are partially reported in a paper (eg, stating only that ‘P>0.05’).3 Alternatively, there may be bias in selection of the reported result, where authors perform multiple analyses for a particular outcome/association, yet only report the result which yielded the most favourable effect estimate.4 Evidence from cohorts of clinical trials followed from inception suggest that biased dissemination is common. Specifically, on average, half of all trials are not published,1 5 trials with statistically significant results are twice as likely to be published5 and a third of trials have outcomes that are omitted, added or modified between protocol and publication.6 Audits of systematic review conduct suggest that most systematic reviewers do not assess risk of reporting biases.7–10 For example, in a cross-sectional study of 300 systematic reviews indexed in MEDLINE in February 2014,7 the risk of bias due to selective publication was not considered in 56% of reviews. A common reason for not doing so was that the small number of included studies, or inability to perform a meta-analysis, precluded the use of funnel plots. Only 19% of reviews included a search of a trial registry to identify completed but unpublished trials or prespecified but non-reported outcomes, and only 7% included a search of another source of data disseminated outside of journal articles. The risk of bias due to selective non-reporting in the included studies was assessed in only 24% of reviews.7 Another study showed that authors of Cochrane reviews routinely record whether any outcomes that were measured were not reported in the included trials, yet rarely consider if such non-reporting could have biased the results of a synthesis.11 Previous researchers have summarised the characteristics of tools designed to assess various sources of bias in randomised trials,12–14 non-randomised studies of interventions (NRSI),14 15 diagnostic test accuracy studies16 and systematic reviews.14 17 Others have summarised the performance of statistical methods developed to detect or adjust for reporting biases.18–20 However, no prior review has focused specifically on tools (ie, structured instruments such as scales, checklists or domain-based tools) for assessing the risk of reporting biases. A particular challenge when assessing risk of reporting biases is that existing tools vary in their level of assessment. For example, tools for assessing risk of bias due to selective publication direct assessments at the level of the synthesis, whereas tools for assessing risk of bias due to selective non-reporting within studies can direct assessments at the level of the individual study, at the level of the synthesis or at both levels. It is unclear how many tools are available to assess different types of reporting bias, and what level they direct assessments at. It is also unclear whether criteria for reaching risk of bias judgements are consistent across existing tools. Therefore, the aim of this research was to conduct a systematic review of the content and measurement properties of such tools.

Methods

Protocol

Methods for this systematic review were prespecified in a protocol which was uploaded to the Open Science Framework in February 2017 (https://osf.io/9ea22/).

Eligibility criteria

Papers were included if the authors described a tool that was designed for use by individuals performing evidence syntheses to assess risk of reporting biases in the included studies or in their synthesis of studies. Tools could assess any type of reporting bias, including bias due to selective publication, bias due to selective non-reporting or bias in selection of the reported result. Tools could assess the risk of reporting biases in any type of study (eg, randomised trial of intervention, diagnostic test accuracy study, observational study estimating prevalence of an exposure) and in any type of result (eg, estimate of intervention efficacy or harm, estimate of diagnostic accuracy, association between exposure and outcome). Eligible tools could take any form, including scales, checklists and domain-based tools. To be considered a scale, each item had to have a numeric score attached to it, so that an overall summary score could be calculated.12 To be considered a checklist, the tool had to include multiple questions, but the developers’ intention was not to attach a numerical score to each response, or to calculate an overall score.13 Domain-based tools were those that required users to judge risk of bias or quality within specific domains, and to record the information on which each judgement was based.21 Tools with a broad scope, for example, to assess multiple sources of bias or the overall quality of the body of evidence, were eligible if one of the items covered risk of reporting bias. Multidimensional tools with a statistical component were also eligible (eg, those that require users to respond to a set of questions about the comprehensiveness of the search, as well as to perform statistical tests for funnel plot asymmetry). In addition, any studies that evaluated the measurement properties of existing tools (eg, construct validity, inter-rater agreement, time taken to complete assessments) were eligible for inclusion. Papers were eligible regardless of the date or format of publication, but were limited to those written in English. The following were ineligible: articles or book chapters providing guidance on how to address reporting biases, but which do not include a structured tool that can be applied by users (eg, the 2011 Cochrane Handbook chapter on reporting biases22); tools developed or modified for use in one particular systematic review; tools designed to appraise published systematic reviews, such as the Risk Of Bias In Systematic reviews (ROBIS) tool23 or A MeaSurement Tool to Assess systematic Reviews (AMSTAR)24; articles that focus on the development or evaluation of statistical methods to detect or adjust for reporting biases, as these have been reviewed elsewhere.18–20

Search methods

On 9 February 2017, one author (MJP) searched for potentially relevant records in Ovid MEDLINE (January 1946 to February 2017), Ovid Embase (January 1980 to February 2017) and Ovid PsycINFO (January 1806 to February 2017). The search strategies included terms relating to reporting bias which were combined with a search string used previously by Whiting et al to identify risk of bias/quality assessment tools17 (see full Boolean search strategies in online supplementary table S1). To capture any tools not published by formal academic publishers, we searched Google Scholar using the phrase ‘reporting bias tool OR risk of bias’. One author (MJP) screened the titles of the first 300 records, as recommended by Haddaway et al.25 To capture any papers that may have been missed by all searches, one author (MJP) screened the references of included articles. In April 2017, the same author emailed the list of included tools to 15 individuals with expertise in reporting biases and risk of bias assessment, and asked if they were aware of any other tools we had not identified.

Study selection and data collection

One author (MJP) screened all titles and abstracts retrieved by the searches. The same author screened any full-text articles retrieved. One author (MJP) collected data from included papers using a standardised data-collection form. The following data on included tools were collected: type of tool (scale, checklist or domain-based tool); types of reporting bias addressed by the tool; level of assessment (ie, whether users direct assessments at the synthesis or at the individual studies included in the synthesis); whether the tool is designed for general use (generic) or targets specific study designs or topic areas (specific); items included in the tool; how items within the tool are rated; methods used to develop the tool (eg, Delphi study, expert consensus meeting); availability of guidance to assist with completion of the tool (eg, guidance manual). The following data from studies evaluating measurement properties of an included tool were collected: tool evaluated measurement properties evaluated (eg, inter-rater agreement) number of syntheses/studies evaluated publication year of syntheses/studies evaluated areas of healthcare addressed by syntheses/studies evaluated number of assessors estimate (and precision) of psychometric statistics (eg, weighted kappa; κ).

Data analysis

We summarised the characteristics of included tools in tables. We calculated the median (IQR) number of items across all tools, and tabulated the frequency of different criteria used in tools to denote a judgement of ‘high’ risk of reporting bias. We summarised estimates of psychometric statistics, such as weighted κ to estimate inter-rater agreement,26 by reporting the range of values across studies. For studies reporting weighted κ, we categorised agreement according to the system proposed by Landis and Koch,27 as poor (0.00), slight (0.01–0.20), fair (0.21–0.40), moderate (0.41–0.60), substantial (0.61–0.80) or almost perfect (0.81–1.00).

Results

In total, 5554 records were identified from the searches, of which we retrieved 165 for full-text screening (figure 1). The inclusion criteria were met by 42 reports summarising 18 tools (table 1) and 17 studies evaluating the measurement properties of tools.3 4 21 28–66 A list of excluded papers is presented in online supplementary table S2. No additional tools were identified by the 15 experts contacted.
Figure 1

Flow diagram of identification, screening and inclusion of studies. aRecords identified from Ovid MEDLINE, Ovid Embase, Ovid PsycINFO and Google Scholar. bRecords identified from screening references of included articles. SR, systematic review.

Table 1

List of included tools

Article IDToolScope of toolTypes of reporting biases assessedLevel of assessment*
Selective publicationSelective non-reportingSelection of the reported result
Balshem et al 28 Agency for Healthcare Research and Quality (AHRQ) outcome and analysis reporting bias frameworkReporting bias onlySpecific outcome/result in a study
Berkman et al 29 AHRQ tool for evaluating the risk of reporting biasReporting bias onlySpecific synthesis of studies
Downes et al 30 Appraisal tool for Cross-Sectional Studies (AXIS) toolMultiple sources of biasStudy
Downs and Black31 Downs-Black toolMultiple sources of biasStudy
Guyatt et al 33–37 Grading of Recommendations Assessment, Development and Evaluation (GRADE)Multiple sources of biasSpecific synthesis of studies
Hayden et al 38 Quality In Prognosis Studies (QUIPS) toolMultiple sources of biasStudy
Higgins et al 21 39 40 Cochrane risk of bias tool for randomised trials (RoB 1.0)Multiple sources of biasStudy
Higgins et al 41 42 RoB 2.0 revised tool for assessing risk of bias in randomised trialsMultiple sources of biasSpecific outcome/result in a study
Hoojimans et al 43 SYstematic Review Centre for Laboratory animal Experimentation (SYRCLE) RoB toolMultiple sources of biasStudy
Kim et al 44 Risk of Bias Assessment Tool for Nonrandomized Studies (RoBANS)Multiple sources of biasStudy
Kirkham et al 3 32 Outcome Reporting Bias In Trials I (ORBIT-I) classification system for benefit outcomesReporting bias onlySpecific outcome/result in a study
Meader et al 45 46 Semi-Automated Quality Assessment Tool (SAQAT)Multiple sources of biasSpecific synthesis of studies
Reid et al 47 Selective reporting bias algorithmReporting bias onlyStudy
Saini et al 48 ORBIT-II classification system for harm outcomesReporting bias onlySpecific outcome/result in a study
Salanti et al 49 50 Framework for evaluating the quality of evidence from a network meta-analysisMultiple sources of biasSpecific synthesis of studies
Sterne et al 4 Risk Of Bias In Non-randomized Studies of Interventions I (ROBINS-I) toolMultiple sources of biasSpecific outcome/result in a study
Viswanathan and Berkman51 Research Triangle Institute (RTI) item bank for assessment of risk of bias and precision for observational studies of interventions or exposuresMultiple sources of biasStudy
Viswanathan et al52 RTI item bank for assessing risk of bias and confounding for observational studies of interventions or exposuresMultiple sources of biasStudy

*Level of assessment classified as: ‘study’ when assessments are directed at a study as a whole (eg, tool used to assess whether any outcomes in a study were not reported); ‘specific outcome/result in a study’ when assessments are directed at a specific outcome or result within a study (eg, tools used to assess whether a particular outcome, such as pain, was not reported) or ‘specific synthesis of studies’ when assessments are directed at a specific synthesis (eg, tool used to assess whether a particular synthesis, such as a meta-analysis of pain, is missing unpublished studies).

List of included tools *Level of assessment classified as: ‘study’ when assessments are directed at a study as a whole (eg, tool used to assess whether any outcomes in a study were not reported); ‘specific outcome/result in a study’ when assessments are directed at a specific outcome or result within a study (eg, tools used to assess whether a particular outcome, such as pain, was not reported) or ‘specific synthesis of studies’ when assessments are directed at a specific synthesis (eg, tool used to assess whether a particular synthesis, such as a meta-analysis of pain, is missing unpublished studies). Flow diagram of identification, screening and inclusion of studies. aRecords identified from Ovid MEDLINE, Ovid Embase, Ovid PsycINFO and Google Scholar. bRecords identified from screening references of included articles. SR, systematic review.

General characteristics of included tools

Nearly all of the included tools (16/18; 89%) were domain-based, where users judge risk of bias or quality within specific domains (table 2; individual characteristics of each tool are presented in online supplementary table S3). All tools were designed for generic rather than specific use. Five tools focused solely on the risk of reporting biases3 28 29 47 48; the remainder addressed reporting biases and other sources of bias/methodological quality (eg, problems with randomisation, lack of blinding). Half of the tools (9/18; 50%) addressed only one type of reporting bias (eg, bias due to selective non-reporting only). Tools varied in regard to the study design that they assessed (ie, randomised trial, non-randomised study of an intervention, laboratory animal experiment). The publication year of the tools ranged from 1998 to 2016 (the earliest was the Downs-Black tool,31 a 27-item tool assessing multiple sources of bias, one of which focuses on risk of bias in the selection of the reported result).
Table 2

Summary of general characteristics of included tools

CharacteristicSummary data (n=18 tools)
Type of tool
 Domain-based16 (89%)
 Checklist1 (6%)
 Scale1 (6%)
Scope of tool
 Assessment of reporting bias only5 (28%)
 Assessment of multiple sources of bias/quality13 (72%)
Types of reporting bias assessed
 Bias due to selective publication only0 (0%)
 Bias due to selective non-reporting only6 (33%)
 Bias in selection of the reported result only3 (17%)
 Bias due to selective publication and bias due to selective non-reporting4 (22%)
 Bias due to selective non-reporting and bias in selection of the reported result5 (28%)
Total number of items in the tool7 (5–13)
Number of items relevant to risk of reporting bias1 (1–2)
Number of response options for risk of reporting bias judgement3 (3–3)
Types of study designs to which the tool applies
 Randomised trials only5 (28%)
 Systematic reviews only3 (17%)
 Non-randomised studies of interventions only2 (11%)
 Randomised trials and non-randomised studies of interventions2 (11%)
 Non-randomised studies of interventions or exposures2 (11%)
 Other (cross-sectional studies, animal studies, network meta-analyses, prognosis studies)4 (22%)
Level of assessment of risk of reporting bias
 Study as a whole9 (50%)
 Specific outcome/result in a study5 (28%)
 Specific synthesis of studies4 (22%)
Data sources used to inform tool content*
 Literature review (eg, of items in existing tools or empirical evidence)9 (50%)
 Ideas generated at expert consensus meeting8 (44%)
 Pilot feedback on preliminary version of the tool7 (39%)
 Data from psychometric or cognitive testing†5 (28%)
 Other (eg, adaptation of existing tool)5 (28%)
 Delphi study responses2 (11%)
 No methods stated2 (11%)
Guidance available
 Brief annotation per item/response option9 (50%)
 Detailed guidance manual4 (22%)
 Worked example for each response option2 (11%)
 Detailed annotation per item/response option1 (6%)
 None2 (11%)

Summary data given as number (%) or median (IQR).

*The percentages in this category do not sum to 100% since the development of some tools was informed by multiple data sources.

†Psychometric testing includes any evaluation of the measurement properties (eg, construct validity, inter-rater reliability, test–retest reliability) of a draft version of the tool. Cognitive testing includes use of qualitative methods (eg, interview) to explore whether assessors who are using the tool for the first time were interpreting the tool and guidance as intended.

Summary of general characteristics of included tools Summary data given as number (%) or median (IQR). *The percentages in this category do not sum to 100% since the development of some tools was informed by multiple data sources. †Psychometric testing includes any evaluation of the measurement properties (eg, construct validity, inter-rater reliability, test–retest reliability) of a draft version of the tool. Cognitive testing includes use of qualitative methods (eg, interview) to explore whether assessors who are using the tool for the first time were interpreting the tool and guidance as intended. Assessments for half of the tools (9/18; 50%) are directed at an individual study (eg, tool is used to assess whether any outcomes in a study were not reported). In 5/18 (28%) tools, assessments are directed at a specific outcome or result within a study (eg, tool is used to assess whether a particular outcome in a study, such as pain, was not reported). In a few tools (4/18; 22%), assessments are directed at a specific synthesis (eg, tool is used to assess whether a particular synthesis, such as a meta-analysis of studies examining pain as an outcome, is missing unpublished studies). The content of the included tools was informed by various sources of data. The most common included a literature review of items used in existing tools or a literature review of empirical evidence of bias (9/18; 50%), ideas generated at an expert consensus meeting (8/18; 44%) and pilot feedback on a preliminary version of the tool (7/18; 39%). The most common type of guidance available for the tools was a brief annotation per item/response option (9/18; 50%). A detailed guidance manual is available for four (22%) tools.

Tool content

Four tools include items for assessing risk of bias due to both selective publication and selective non-reporting.29 33 45 49 One of these tools (the AHRQ tool for evaluating the risk of reporting bias29) directs users to assess a particular synthesis, where a single risk of bias judgement is made based on information about unpublished studies and under-reported outcomes. In the other three tools (the GRADE framework, and two others which are based on GRADE),33 45 49 the different sources of reporting bias are assessed in separate domains (bias due to selective non-reporting is considered in a ‘study limitations (risk of bias)’ domain, while bias due to selective publication is considered in a ‘publication bias’ domain). Five tools21 28 43 44 47 guide users to assess risk of bias due to both selective non-reporting and selection of the reported result (ie, problems with outcomes/results that are not reported and those that are reported, respectively). Four of these tools, which include the Cochrane risk of bias tool for randomised trials21 and three others which are based on the Cochrane tool,43 44 47 direct assessments at the study level. That is, a whole study is rated at ‘high’ risk of reporting bias if any outcome/result in the study has been omitted, or fully reported, on the basis of the findings. Some of the tools designed to assess the risk of bias due to selective non-reporting ask users to assess, for particular outcomes of interest, whether the outcome was not reported or only partially reported in the study on the basis of its results (eg, Outcome Reporting Bias In Trials (ORBIT) tools,3 48 the AHRQ outcome reporting bias framework,28 and GRADE.34 This allows users to perform multiple outcome-level assessments of the risk of reporting bias (rather than one assessment for the study as a whole). In total, 15 tools include a mechanism for assessing risk of bias due to selective non-reporting in studies, but assessing the corresponding risk of bias in a synthesis that is missing the non-reported outcomes is not within the scope of 11 of these tools.3 21 28 30 38 43 44 47 48 51 52 A variety of criteria are used in existing tools to inform a judgement of ‘high’ risk of bias due to selective publication (table 3), selective non-reporting (table 4), and selection of the reported result (table 5; more detail is provided in online supplementary table S4). In the four tools with an assessment of risk of bias due to selective publication, ‘high’ risk criteria include evidence of funnel plot asymmetry, discrepancies between published and unpublished studies, use of non-comprehensive searches and presence of small, ‘positive’ studies with for-profit interest (table 3). However, not all of these criteria appear in all tools (only evidence of funnel plot asymmetry does), and the relative weight assigned to each criterion in the overall risk of reporting bias judgement is clear for only one tool (the Semi-Automated Quality Assessment Tool; SAQAT).45 46
Table 3

Criteria used in existing tools to inform a judgement of ‘high’ risk of bias due to selective publication

‘High’ risk of bias criteria proposed in existing toolsAHRQ RRBGRADESAQATNMA-QualityTotal, n (%)
Assessment directed at a specific synthesis (eg, meta-analysis)
 Evidence of funnel plot asymmetry (based on visual inspection of funnel plot or statistical test for funnel plot asymmetry)4 (100)
 Smaller studies tend to demonstrate more favourable results (based on visual assessment, without funnel plot)1 (25)
 Clinical decision would differ for estimates from a fixed-effect versus a random-effects model because the findings from a fixed-effect model are closer to the null1 (25)
 Substantial heterogeneity in the meta-analysis cannot be explained by some clinical or methodological factor1 (25)
 At least one study is affected by non-publication or non-accessibility1 (25)
 Presence of small (often ‘positive’) studies with for-profit interest in the synthesis2 (50)
 Presence of early studies (ie, set of small, ‘positive’ trials addressing a novel therapy) in the synthesis2 (50)
 Discrepancy in findings between published and unpublished trials3 (75)
 Search strategies were not comprehensive3 (75)
 Methods to identify all available evidence were not comprehensive2 (50)
 Grey literature were not searched1 (25)
 Restrictions to study selection on the basis of language were applied1 (25)
 Industry influence may apply to studies included in the synthesis1 (25)

AHRQ RRB, AHRQ tool for evaluating the risk of reporting bias29; GRADE, GRADE rating of quality of evidence34–37; NMA-Quality, Framework for evaluating the quality of evidence from a network meta-analysis49; SAQAT, Semi-Automated Quality Assessment Tool.45 46

Table 4

Criteria used in existing tools to inform a judgement of ‘high’ risk of bias due to selective non-reporting

‘High’ risk of bias criteria proposed in existing toolsAHRQ ORBAHRQ RRBAXISGRADEQUIPSRoB 1.0SYRCLE RoBRoBANSORBIT-ISAQATReidORBIT-IINMA-QualityRTI 2012RTI 2013Total, n (%)
Assessment directed at study as a whole
 One or more outcomes of interest were clearly measured, but no results were reported8 (53)
 One or more outcomes of interest were reported incompletely so that they could not be entered in a meta-analysis2 (13)
 The study report fails to include results for a key outcome that would be expected to have been reported for such a study5 (33)
Assessment directed at a specific outcome
 Particular outcome clearly measured but no results were reported6 (40)
 Particular outcome of interest is reported incompletely so that it cannot be entered in a meta-analysis (typically stating only that P>0.05)6 (40)
 Judgement says particular outcome is likely to have been measured and analysed but not reported on the basis of its results6 (40)
 Composite outcomes are presented without the individual component outcomes2 (13)
 Result reported globally across all groups1 (7)
 Result reported for some groups only1 (7)
 Data were not reported consistently for the outcome of interest1 (7)
Assessment directed at a specific synthesis
 Selective non-reporting suspected in a number of included studies4 (27)

AHRQ ORB, AHRQ outcome and analysis reporting bias framework28; AHRQ RRB, AHRQ tool for evaluating the risk of reporting bias29; AXIS, Appraisal tool for Cross-Sectional Studies30; GRADE, GRADE rating of quality of evidence34–37; NMA-Quality, Framework for evaluating the quality of evidence from a network meta-analysis49; ORBIT-I, Outcome Reporting Bias In Trials classification system for benefit outcomes3 32; ORBIT-II, Outcome Reporting Bias In Trials classification system for harm outcomes48; QUIPS, Quality In Prognosis Studies tool38; Reid, Reid et al selective reporting bias algorithm47; RoB 1.0, Cochrane risk of bias tool for randomised trials21 39 40; RoBANS, Risk of Bias Assessment Tool for Nonrandomized Studies44; RTI 2012, RTI Item Bank for Assessment of Risk of Bias and Precision for Observational Studies of Interventions or Exposures51; RTI 2013, RTI Item Bank for Assessing Risk of Bias and Confounding for Observational Studies of Interventions or Exposures52; SAQAT, Semi-Automated Quality Assessment Tool45 46; SYRCLE RoB, SYstematic Review Centre for Laboratory animal Experimentation risk of bias tool.43

Table 5

Criteria used in existing tools to inform a judgement of ‘high’ risk of bias in selection of the reported result

‘High’ risk of bias criteria proposed in existing toolsAHRQ ORBDowns-BlackRoB 1.0RoB 2.0SYRCLE RoBRoBANSReidROBINS-ITotal, n (%)
Assessment directed at study as a whole
 One or more reported outcomes were not prespecified (unless clear justification for their reporting is provided, such as an unexpected adverse event)4 (50)
 One or more outcomes were reported using measurements, analysis methods or subsets of the data (eg, subscales) that were not prespecified2 (15)
 One or more retrospective, unplanned, subgroup analyses were reported1 (13)
 Any analyses that had not been planned at the outset of the study were not clearly indicated1 (13)
Assessment directed at a specific outcome/result
 Particular outcome was not prespecified but results were reported1 (13)
 Reported result for a particular outcome is likely to have been selected, on the basis of the findings, from multiple outcome measurements (eg, scales, definitions, time points) within the outcome domain2 (25)
 Reported result for a particular outcome is likely to have been selected, on the basis of the findings, from multiple analyses of the data2 (25)
 Reported result for a particular outcome is likely to have been selected, on the basis of the findings, from different subgroups1 (13)

AHRQ ORB, AHRQ outcome and analysis reporting bias framework28; Downs-Black, Downs Black tool31; Reid, Reid et al selective reporting bias algorithm47; RoB 1.0, Cochrane risk of bias tool for randomised trials21 39 40; RoB 2.0, Revised tool for assessing risk of bias in randomised trials41 42; RoBANS, Risk of Bias Assessment Tool for Nonrandomized Studies44; ROBINS-I, Risk Of Bias In Non-randomized Studies of Interventions tool4; SYRCLE RoB, SYstematic Review Centre for Laboratory animal Experimentation risk of bias tool.43

Criteria used in existing tools to inform a judgement of ‘high’ risk of bias due to selective publication AHRQ RRB, AHRQ tool for evaluating the risk of reporting bias29; GRADE, GRADE rating of quality of evidence34–37; NMA-Quality, Framework for evaluating the quality of evidence from a network meta-analysis49; SAQAT, Semi-Automated Quality Assessment Tool.45 46 Criteria used in existing tools to inform a judgement of ‘high’ risk of bias due to selective non-reporting AHRQ ORB, AHRQ outcome and analysis reporting bias framework28; AHRQ RRB, AHRQ tool for evaluating the risk of reporting bias29; AXIS, Appraisal tool for Cross-Sectional Studies30; GRADE, GRADE rating of quality of evidence34–37; NMA-Quality, Framework for evaluating the quality of evidence from a network meta-analysis49; ORBIT-I, Outcome Reporting Bias In Trials classification system for benefit outcomes3 32; ORBIT-II, Outcome Reporting Bias In Trials classification system for harm outcomes48; QUIPS, Quality In Prognosis Studies tool38; Reid, Reid et al selective reporting bias algorithm47; RoB 1.0, Cochrane risk of bias tool for randomised trials21 39 40; RoBANS, Risk of Bias Assessment Tool for Nonrandomized Studies44; RTI 2012, RTI Item Bank for Assessment of Risk of Bias and Precision for Observational Studies of Interventions or Exposures51; RTI 2013, RTI Item Bank for Assessing Risk of Bias and Confounding for Observational Studies of Interventions or Exposures52; SAQAT, Semi-Automated Quality Assessment Tool45 46; SYRCLE RoB, SYstematic Review Centre for Laboratory animal Experimentation risk of bias tool.43 Criteria used in existing tools to inform a judgement of ‘high’ risk of bias in selection of the reported result AHRQ ORB, AHRQ outcome and analysis reporting bias framework28; Downs-Black, Downs Black tool31; Reid, Reid et al selective reporting bias algorithm47; RoB 1.0, Cochrane risk of bias tool for randomised trials21 39 40; RoB 2.0, Revised tool for assessing risk of bias in randomised trials41 42; RoBANS, Risk of Bias Assessment Tool for Nonrandomized Studies44; ROBINS-I, Risk Of Bias In Non-randomized Studies of Interventions tool4; SYRCLE RoB, SYstematic Review Centre for Laboratory animal Experimentation risk of bias tool.43 All 15 tools with an assessment of the risk of bias due to selective non-reporting suggest that the risk of bias is ‘high’ when it is clear that an outcome was measured but no results were reported (table 4). Fewer of these tools (n=8; 53%) also recommend a ‘high’ risk judgement when results for an outcome are partially reported (eg, it is stated that the result was non-significant, but no effect estimate or summary statistics are presented). The eight tools that include an assessment of the risk of bias in selection of the reported result recommend various criteria for a ‘high’ risk judgement (table 5). These include when some outcomes that were not prespecified are added post hoc (in 4 (50%) tools), or when it is likely that the reported result for a particular outcome has been selected, on the basis of the findings, from among multiple outcome measurements or analyses within the outcome domain (in 2 (25%) tools).

General characteristics of studies evaluating measurement properties of included tools

Despite identifying 17 studies that evaluated measurement properties of an included tool, psychometric statistics for the risk of reporting bias component were available only from 12 studies43 44 54–60 62 64 66 (the other five studies include only data on properties of the multidimensional tool as a whole31 53 61 63 65; online supplementary table S5). Nearly all 12 studies (11; 92%) evaluated inter-rater agreement between two assessors; eight of these studies reported weighted κ values, but only two described the weighting scheme.55 62 Eleven studies43 44 54–60 64 66 evaluated the measurement properties of tools for assessing risk of bias in a study due to selective non-reporting or risk of bias in selection of the reported result; in these 11 studies, a median of 40 (IQR 32–109) studies were assessed. One study62 evaluated a tool for assessing risk of bias in a synthesis due to selective publication, in which 44 syntheses were assessed. In the studies evaluating inter-rater agreement, all involved two assessors.

Results of evaluation studies

Five studies54 56–58 60 included data on the inter-rater agreement of assessments of risk of bias due to selective non-reporting using the Cochrane risk of bias tool for randomised trials21 (table 6). Weighted κ values in four studies54 56–58 ranged from 0.13 to 0.50 (sample size ranged from 87 to 163 studies), suggesting slight to moderate agreement.27 In the other study,60 the per cent agreement in selective non-reporting assessments in trials that were included in two different Cochrane reviews was low (43% of judgements were in agreement). Two other studies found that inter-rater agreement of selective non-reporting assessments were substantial for SYRCLE’s RoB tool (κ=0.62, n=32),43 but poor for the RoBANS tool (κ=0, n=39).44 There was substantial agreement between raters in the assessment of risk of bias due to selective publication using the SAQAT (κ=0.63, n=29).62 The inter-rater agreement of assessments of risk of bias in selection of the reported result using the ROBINS-I tool4 was moderate for NRSI included in a review of the effect of cyclooxygenase-2 inhibitors on cardiovascular events (κ=0.45, n=21), and substantial for NRSI included in a review of the effect of thiazolidinediones on cardiovascular events (κ=0.78, n=16).55
Table 6

Reported measurement properties of tools with an assessment of the risk of reporting bias

Study IDToolMeasurement propertySample sizeAreas of healthcare addressedWeighted kappa (95% CI)Weighting schemeInterpretation of kappa*
Armijo-Olivo et al 54 RoB 1.0Inter-rater agreement of assessments of risk of bias due to selective non-reporting (between two external reviewers)87Musculoskeletal, cardiorespiratory, neurological and gynaecological conditions.0.5 (CI not reported)Not describedModerate agreement
Armijo-Olivo et al 54 RoB 1.0Inter-rater agreement of assessments of risk of bias due to selective non-reporting (between two external reviewers and Cochrane reviewers)87See above0.13 (CI not reported)Not describedSlight agreement
Hartling et al 56 RoB 1.0Inter-rater agreement of assessments of risk of bias due to selective non-reporting163Child health0.13 (95% CI −0.05 to 0.31)Not describedSlight agreement
Hartling et al 57 RoB 1.0Inter-rater agreement of assessments of risk of bias due to selective non-reporting107Asthma0.4 (95% CI 0.14 to 0.67)Not describedFair agreement
Hartling et al 58 59 RoB 1.0Inter-rater agreement of assessments of risk of bias due to selective non-reporting (between two reviewers, all trials)124Varied0.27 (95% CI 0.06 to 0.49)Not describedFair agreement
Hartling et al 58 59 RoB 1.0Inter-rater agreement of assessments of risk of bias due to selective non-reporting (between pairs of reviewers across different centres, all trials)30Varied0.08 (95% CI −0.09 to 0.26)Not describedSlight agreement
Jordan et al 60 RoB 1.0Inter-rater agreement of assessments of risk of bias due to selective non-reporting (between judgements of trials appearing in two SRs)28SubfertilityNot reported†Not applicableNot applicable
Vale et al 66 RoB 1.0Agreement between selective non-reporting assessments performed using published article only versus published article and data collected during the individual participant data process95Cancer painNot reported†Not applicableNot applicable
Hoojimans et al 43 SYRCLE RoBInter-rater agreement of assessments of risk of bias due to selective non-reporting32Animal studies (not specified)0.62 (CI not reported)Not describedSubstantial agreement
Kim et al 44 RoBANSInter-rater agreement of assessments of risk of bias due to selective non-reporting39Depression, myocardial infarction, postpartum haemorrhage, chronic non-cancer pain0 (CI not reported)Not describedPoor agreement
Llewellyn et al 62 SAQATInter-rater agreement of assessments of risk of bias due to selective publication (between two SAQAT raters)29Varied0.63 (95% CI 0.17 to 1)QuadraticSubstantial agreement
Llewellyn et al 62 SAQATInter-rater agreement of assessments of risk of bias due to selective publication (between one rater using SAQAT and one using the standard GRADE approach)15VariedNot reported†Not applicableNot applicable
Norris et al 64 ORBIT-IInter-rater agreement of ORBIT-I classifications of risk of bias due to selective non-reporting40VariedNot calculated, as too little variation in judgementsNot applicableNot applicable
Bilandzic et al 55 ROBINS-IInter-rater agreement of assessments of risk of bias in selection of the reported result16Thiazolidinediones and cardiovascular events0.78 (CI not reported)LinearSubstantial agreement
Bilandzic et al 55 ROBINS-IInter-rater agreement of assessments of risk of bias in selection of the reported result21COX-2 inhibitors and cardiovascular events0.45 (CI not reported)LinearModerate agreement

*Interpretation of kappa based on categorisation system defined by Landis et al.27

†Data presented as per cent agreement, not weighted kappa.

COX-2, cyclooxygenase-2; ORBIT-I, Outcome Reporting Bias In Trials classification system for benefit outcomes3 32; RoB 1.0, Cochrane risk of bias tool for randomised trials21 39 40; RoBANS, Risk of Bias Assessment Tool for Nonrandomized Studies44; ROBINS-I, Risk Of Bias In Non-randomized Studies of Interventions tool4; SAQAT, Semi-Automated Quality Assessment Tool45 46; SRs, systematic reviews; SYRCLE RoB, SYstematic Review Centre for Laboratory animal Experimentation risk of bias tool.43

Reported measurement properties of tools with an assessment of the risk of reporting bias *Interpretation of kappa based on categorisation system defined by Landis et al.27 †Data presented as per cent agreement, not weighted kappa. COX-2, cyclooxygenase-2; ORBIT-I, Outcome Reporting Bias In Trials classification system for benefit outcomes3 32; RoB 1.0, Cochrane risk of bias tool for randomised trials21 39 40; RoBANS, Risk of Bias Assessment Tool for Nonrandomized Studies44; ROBINS-I, Risk Of Bias In Non-randomized Studies of Interventions tool4; SAQAT, Semi-Automated Quality Assessment Tool45 46; SRs, systematic reviews; SYRCLE RoB, SYstematic Review Centre for Laboratory animal Experimentation risk of bias tool.43

Discussion

From a systematic search of the literature, we identified 18 tools designed for use by individuals performing evidence syntheses to assess risk of reporting biases in the included studies or in their synthesis of studies. The tools varied with regard to the type of reporting bias assessed (eg, bias due to selective publication, bias due to selective non-reporting), and the level of assessment (eg, for the study as a whole, a particular outcome within a study or a particular synthesis of studies). Various criteria are used across tools to designate a synthesis as being at ‘high’ risk of bias due to selective publication (eg, evidence of funnel plot asymmetry, use of non-comprehensive searches). However, the relative weight assigned to each criterion in the overall judgement is not clear for most of these tools. Tools for assessing risk of bias due to selective non-reporting guide users to assess a study, or an outcome within a study, as ‘high’ risk of bias if no results are reported for an outcome. However, assessing the corresponding risk of bias in a synthesis that is missing the non-reported outcomes is outside the scope of most of these tools. Inter-rater agreement estimates were available for five tools,4 21 43 44 62 and ranged from poor to substantial; however, the sample sizes of most evaluations were small, and few described the weighting scheme used to calculate κ.

Strengths and limitations

There are several strengths of this research. Methods were conducted in accordance with a systematic review protocol (https://osf.io/9ea22/). Published articles were identified by searching several relevant databases using a search string developed in conjunction with an information specialist,17 and by contacting experts to identify tools missed by the search. Detailed information on the content and measurement properties of existing tools was collected, providing readers with pertinent information to help decide which tools to use in future reviews. However, the findings need to be considered in light of some limitations. Screening of articles and data collection were performed by one author only. It is therefore possible that some relevant articles were missed, or that errors in data collection were made. The search for unpublished tools was not comprehensive (only Google Scholar was searched), so it is possible that other tools for assessing risk of reporting biases exist. Further, restricting the search to articles in English was done to expedite the review process, but may have resulted in loss of information about tools written in other languages, and additional evidence on measurement properties of tools.

Comparison with other studies

Other systematic reviews of risk of bias tools12–17 have restricted inclusion to tools developed for particular study designs (eg, randomised trials, diagnostic test accuracy studies), where the authors recorded all the sources of bias addressed. A different approach was taken in the current review, where all tools (regardless of study design) that address a particular source of bias were examined. By focusing on one source of bias only, the analysis of included items and criteria for risk of bias judgements was more detailed than that recorded previously. Some of the existing reviews of tools15 considered tools that were developed or modified in the context of a specific systematic review. However, such tools were excluded from the current review as they are unlikely to have been developed systematically,15 67 and are difficult to find (all systematic reviews conducted during a particular period would need to have been examined for the search to be considered exhaustive).

Explanations and implications

Of the 18 tools identified, only four (22%) included a mechanism for assessing risk of bias due to selective publication, which is the type of reporting bias that has been investigated by methodologists most often.2 This is perhaps unsurprising given that hundreds of statistical methods to ‘detect’ or ‘adjust’ for bias due to selective publication have been developed.18 These statistical methods may be considered by methodologists and systematic reviewers as the tools of choice for assessing this type of bias. However, application of these statistical methods without considering other factors (eg, existence of registered but unpublished studies, conflicts of interest that may influence investigators to not disseminate studies with unfavourable results) is not sufficiently comprehensive, and could lead to incorrect conclusions about the risk of bias due to selective publication. Further, there are many limitations of these statistical approaches, in terms of their underlying assumptions, statistical power, which is often low because most meta-analyses include few studies,7 and the need for specialist statistical software to apply them.19 68 These factors may have limited their use in practice and potentially explain why a large number of systematic reviewers currently ignore the risk of bias due to selective publication.7–9 69 Our analysis suggests that the factors that need to be considered to assess risk of reporting biases adequately (eg, comprehensiveness of the search, amount of data missing from the synthesis due to unpublished studies and under-reported outcomes) are fragmented. A similar problem was occurring a decade ago with the assessment of risk of bias in randomised trials. Some authors assessed only problems with randomisation, while others focused on whether trials were not ‘double blinded’ or had any missing participant data.70 It was not until all the important bias domains were brought together into a structured, domain-based tool to assess the risk of bias in randomised trials,21 that systematic reviewers started to consider risk of bias in trials comprehensively. A similar initiative to link all the components needed to judge the risk of reporting biases into a comprehensive new tool may improve the credibility of evidence syntheses. In particular, there is an emergent need for a new tool to assess the risk that a synthesis is affected by reporting biases. This tool could guide users to consider risk of bias in a synthesis due to both selective publication and selective non-reporting, given that both practices lead to the same consequence: evidence missing from the synthesis.11 Such a tool would complement recently developed tools for assessing risk of bias within studies (RoB 2.041 and ROBINS-I4 which include a domain for assessing the risk of bias in selection of the reported result, but no mechanism to assess risk of bias due to selective non-reporting). Careful thought would need to be given as to how to weigh up various pieces of information underpinning the risk of bias judgement. For example, users will need guidance on how evidence of known, unpublished studies (as identified from trial registries, protocols or regulatory documents) should be considered alongside evidence that is more speculative (eg, funnel plots suggesting that studies may be missing). Further, guidance for the tool will need to emphasise the value of seeking documents other than published journal articles (eg, protocols) to inform risk of bias judgements. Preparation of a detailed guidance manual may enhance the usability of the tool, minimise misinterpretation and increase reliability in assessments. Once developed, evaluations of the measurement properties of the tool, such as inter-rater agreement and construct validity, should be conducted to explore whether modifications to the tool are necessary.

Conclusions

There are several limitations of existing tools for assessing risk of reporting biases in studies or syntheses of studies, in terms of their scope, guidance for reaching risk of bias judgements and measurement properties. Development and evaluation of a new, comprehensive tool could help overcome present limitations.
  55 in total

1.  GRADE: an emerging consensus on rating quality of evidence and strength of recommendations.

Authors:  Gordon H Guyatt; Andrew D Oxman; Gunn E Vist; Regina Kunz; Yngve Falck-Ytter; Pablo Alonso-Coello; Holger J Schünemann
Journal:  BMJ       Date:  2008-04-26

2.  The feasibility of creating a checklist for the assessment of the methodological quality both of randomised and non-randomised studies of health care interventions.

Authors:  S H Downs; N Black
Journal:  J Epidemiol Community Health       Date:  1998-06       Impact factor: 3.710

3.  GRADE guidelines: 5. Rating the quality of evidence--publication bias.

Authors:  Gordon H Guyatt; Andrew D Oxman; Victor Montori; Gunn Vist; Regina Kunz; Jan Brozek; Pablo Alonso-Coello; Ben Djulbegovic; David Atkins; Yngve Falck-Ytter; John W Williams; Joerg Meerpohl; Susan L Norris; Elie A Akl; Holger J Schünemann
Journal:  J Clin Epidemiol       Date:  2011-07-30       Impact factor: 6.437

4.  Can trial quality be reliably assessed from published reports of cancer trials: evaluation of risk of bias assessments in systematic reviews.

Authors:  Claire L Vale; Jayne F Tierney; Sarah Burdett
Journal:  BMJ       Date:  2013-04-22

5.  Recommendations for examining and interpreting funnel plot asymmetry in meta-analyses of randomised controlled trials.

Authors:  Jonathan A C Sterne; Alex J Sutton; John P A Ioannidis; Norma Terrin; David R Jones; Joseph Lau; James Carpenter; Gerta Rücker; Roger M Harbord; Christopher H Schmid; Jennifer Tetzlaff; Jonathan J Deeks; Jaime Peters; Petra Macaskill; Guido Schwarzer; Sue Duval; Douglas G Altman; David Moher; Julian P T Higgins
Journal:  BMJ       Date:  2011-07-22

6.  Assessing the potential for outcome reporting bias in a review: a tutorial.

Authors:  Kerry Dwan; Carrol Gamble; Ruwanthi Kolamunnage-Dona; Shabana Mohammed; Colin Powell; Paula R Williamson
Journal:  Trials       Date:  2010-05-12       Impact factor: 2.279

Review 7.  Failure of a numerical quality assessment scale to identify potential risk of bias in a systematic review: a comparison study.

Authors:  Seán R O'Connor; Mark A Tully; Brigid Ryan; Judy M Bradley; George D Baxter; Suzanne M McDonough
Journal:  BMC Res Notes       Date:  2015-06-06

8.  ROBINS-I: a tool for assessing risk of bias in non-randomised studies of interventions.

Authors:  Jonathan Ac Sterne; Miguel A Hernán; Barnaby C Reeves; Jelena Savović; Nancy D Berkman; Meera Viswanathan; David Henry; Douglas G Altman; Mohammed T Ansari; Isabelle Boutron; James R Carpenter; An-Wen Chan; Rachel Churchill; Jonathan J Deeks; Asbjørn Hróbjartsson; Jamie Kirkham; Peter Jüni; Yoon K Loke; Theresa D Pigott; Craig R Ramsay; Deborah Regidor; Hannah R Rothstein; Lakhbir Sandhu; Pasqualina L Santaguida; Holger J Schünemann; Beverly Shea; Ian Shrier; Peter Tugwell; Lucy Turner; Jeffrey C Valentine; Hugh Waddington; Elizabeth Waters; George A Wells; Penny F Whiting; Julian Pt Higgins
Journal:  BMJ       Date:  2016-10-12

9.  ROBIS: A new tool to assess risk of bias in systematic reviews was developed.

Authors:  Penny Whiting; Jelena Savović; Julian P T Higgins; Deborah M Caldwell; Barnaby C Reeves; Beverley Shea; Philippa Davies; Jos Kleijnen; Rachel Churchill
Journal:  J Clin Epidemiol       Date:  2015-06-16       Impact factor: 6.437

10.  SYRCLE's risk of bias tool for animal studies.

Authors:  Carlijn R Hooijmans; Maroeska M Rovers; Rob B M de Vries; Marlies Leenaars; Merel Ritskes-Hoitinga; Miranda W Langendam
Journal:  BMC Med Res Methodol       Date:  2014-03-26       Impact factor: 4.615

View more
  51 in total

Review 1.  Impacts of Gentrification on Health in the US: a Systematic Review of the Literature.

Authors:  Genee S Smith; Hannah Breakstone; Lorraine T Dean; Roland J Thorpe
Journal:  J Urban Health       Date:  2020-12       Impact factor: 3.671

Review 2.  Legal regulations, ethical guidelines and recent policies to increase transparency of clinical trials.

Authors:  Jan Borysowski; Agata Wnukiewicz-Kozłowska; Andrzej Górski
Journal:  Br J Clin Pharmacol       Date:  2020-02-19       Impact factor: 4.335

Review 3.  Adverse Drug Reactions in Pediatric Oncohematology: A Systematic Review.

Authors:  Kristopher Amaro-Hosey; Immaculada Danés; Antònia Agustí
Journal:  Front Pharmacol       Date:  2022-02-03       Impact factor: 5.810

4.  Occupational advice to help people return to work following lower limb arthroplasty: the OPAL intervention mapping study.

Authors:  Paul Baker; Carol Coole; Avril Drummond; Sayeed Khan; Catriona McDaid; Catherine Hewitt; Lucksy Kottam; Sarah Ronaldson; Elizabeth Coleman; David A McDonald; Fiona Nouri; Melanie Narayanasamy; Iain McNamara; Judith Fitch; Louise Thomson; Gerry Richardson; Amar Rangan
Journal:  Health Technol Assess       Date:  2020-09       Impact factor: 4.014

Review 5.  A systematic review and meta-analysis of microbial contamination of parenteral medication prepared in a clinical versus pharmacy environment.

Authors:  Karin H M Larmené-Beld; Henderik W Frijlink; Katja Taxis
Journal:  Eur J Clin Pharmacol       Date:  2019-01-25       Impact factor: 2.953

6.  Psychological interventions for depression and anxiety in pregnant Latina and Black women in the United States: A systematic review.

Authors:  Carolyn Ponting; Nicole E Mahrer; Hannah Zelcer; Christine Dunkel Schetter; Denise A Chavira
Journal:  Clin Psychol Psychother       Date:  2020-01-30

Review 7.  The relationship between minority stress and biological outcomes: A systematic review.

Authors:  Annesa Flentje; Nicholas C Heck; James Michael Brennan; Ilan H Meyer
Journal:  J Behav Med       Date:  2019-12-20

Review 8.  Kinesio taping for rotator cuff disease.

Authors:  Silvia Gianola; Valerio Iannicelli; Edoardo Fascio; Anita Andreano; Linda C Li; Maria Grazia Valsecchi; Lorenzo Moja; Greta Castellini
Journal:  Cochrane Database Syst Rev       Date:  2021-08-08

9.  Aortic valve function post-replacement of severe aortic stenosis by transcatheter procedure versus surgery: a systematic review and metanalysis.

Authors:  Barbara Ignatiuk; Guliz Erdem; Hani Jneid; Nikolaos Bonaros; Charbel Abi Khalil; Hiam Chemaitelly; Fabio Barilli; Mohamed El-Shazly; Jassim Al Suwaidi; Samar Aboulsoud; Markus Kofler; Lukas Stastny
Journal:  Sci Rep       Date:  2021-06-07       Impact factor: 4.379

Review 10.  Methodological review to develop a list of bias items used to assess reviews incorporating network meta-analysis: protocol and rationale.

Authors:  Carole Lunny; Andrea C Tricco; Areti-Angeliki Veroniki; Sofia Dias; Brian Hutton; Georgia Salanti; James M Wright; Ian White; Penny Whiting
Journal:  BMJ Open       Date:  2021-06-24       Impact factor: 3.006

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.