Literature DB >> 24618053

Reporting of results from network meta-analyses: methodological systematic review.

Aïda Bafeta1, Ludovic Trinquart, Raphaèle Seror, Philippe Ravaud.   

Abstract

OBJECTIVE: To examine how the results of network meta-analyses are reported.
DESIGN: Methodological systematic review of published reports of network meta-analyses. DATA SOURCES: Cochrane Database of Systematic Reviews, Database of Abstracts of Reviews of Effects, Medline, and Embase, searched from inception to 12 July 2012. STUDY SELECTION: All network meta-analyses comparing the clinical efficacy of three or more interventions in randomised controlled trials were included, excluding meta-analyses with an open loop network of three interventions. DATA EXTRACTION AND SYNTHESIS: The reporting of the network and results was assessed. A composite outcome included the description of the network (number of interventions, direct comparisons, and randomised controlled trials and patients for each comparison) and the reporting of effect sizes derived from direct evidence, indirect evidence, and the network meta-analysis.
RESULTS: 121 network meta-analyses (55 published in general journals; 48 funded by at least one private source) were included. The network and its geometry (network graph) were not reported in 100 (83%) articles. The effect sizes derived from direct evidence, indirect evidence, and the network meta-analysis were not reported in 48 (40%), 108 (89%), and 43 (36%) articles, respectively. In 52 reports that ranked interventions, 43 did not report the uncertainty in ranking. Overall, 119 (98%) reports of network meta-analyses did not give a description of the network or effect sizes from direct evidence, indirect evidence, and the network meta-analysis. This finding did not differ by journal type or funding source.
CONCLUSIONS: The results of network meta-analyses are heterogeneously reported. Development of reporting guidelines to assist authors in writing and readers in critically appraising reports of network meta-analyses is timely.

Entities:  

Mesh:

Year:  2014        PMID: 24618053      PMCID: PMC3949412          DOI: 10.1136/bmj.g1741

Source DB:  PubMed          Journal:  BMJ        ISSN: 0959-8138


Introduction

When several healthcare interventions are available for the same condition, the corresponding network of results from randomised controlled trials must be considered (that is, which of the considered interventions have been compared against each other or against a common comparator such as placebo).1 2 A network meta-analysis allows for a quantitative synthesis of the network by combining direct evidence from comparisons of interventions within randomised trials and indirect evidence across randomised trials on the basis of a common comparator.3 4 5 6 Networks of trials and network meta-analyses are increasingly used to evaluate healthcare interventions.7 8 9 10 11 Compared with pairwise meta-analyses, network meta-analyses allow for visualisation of a larger amount of evidence, estimation of the relative effectiveness among all interventions (even if some head to head comparisons are lacking), and rank ordering of the interventions. The conduct of network meta-analyses poses multiple challenges. Several reviews have explained how to approach these challenges in practice11 12 13; others have evaluated how indirect comparison meta-analyses and network meta-analyses have been conducted and reported.8 9 14 15 Most reviews have focused on checking the validity of key assumptions required in network meta-analyses (homogeneity, similarity/transitivity, consistency, or an overarching assumption of exchangeability); others have assessed the presence of essential methodological components of the systematic review process (for example, conducting a literature search and assessing the risk of bias of individual studies).10 These reviews did not evaluate the presentation of findings in reports of network meta-analyses. Although reporting guidelines have been developed for systematic reviews and pairwise meta-analyses, we lack reporting guidelines for network meta-analyses. An international group is preparing an extension of the PRISMA statement for network meta-analyses. Inadequate reporting of findings and inadequate evaluation of the required assumptions may impede confidence in the findings and conclusions of network meta-analyses.16 Reports of network meta-analyses should allow the reader to assess the amount of evidence in the network of randomised trials and the relative effect sizes between interventions, along with their uncertainty.9 17 On the basis of a previous methodological systematic review of published reports of network meta-analyses,10 we aimed to examine how network meta-analysis results are reported.

Methods

Selection of articles

The search strategy and the selection criteria for reports of network meta-analyses have been described elsewhere.10 In brief, we considered networks of randomised trials that assessed three or more treatments but excluded meta-analyses with an open loop network of three interventions. The articles were screened from a sample of 893 potentially relevant publications indexed in the Cochrane Database of Systematic Reviews, the Database of Abstracts of Reviews of Effects, Medline, and Embase. The date of the last search was 12 July 2012. We identified three additional network meta-analyses from methodological articles. Two independent reviewers excluded non-network meta-analysis reports on the basis of the title and abstract of retrieved citations, and the full text articles were then evaluated according to pre-specified eligibility criteria. Disagreements were discussed to reach consensus.

Data extraction

We used a standardised data collection form to collect all data, for general characteristics and those pertaining to reporting of results, from 121 original reports of network meta-analyses and all supplementary appendices referenced in the articles, when available. Two reviewers independently extracted all data for a random sample of 20% of reports, except items that involved subjective interpretation, which were extracted independently and in duplicate for all reports. These items were the information on the structure of the network, whether the authors applied a Bayesian approach, which types of prior distributions for the basic parameters and the between trial variance were used, and whether the authors applied a method to investigate the impact of effect modifiers. Disagreements were resolved by discussion.

General characteristics

We collected data on the following general characteristics: type of journal (general or specialty), funding source (we classified network meta-analyses as having public funding if they were explicitly not funded and were conducted by academic authors or were funded by a public source, private funding if they were funded by at least one private source, and unclear funding if the authors did not report any information on the funding), type of interventions included in the network (pharmacological, non-pharmacological (defined as any intervention that did not include an active substance), or both), the type of outcome (binary, continuous, or time to event) and number of outcomes assessed per network meta-analysis, the number of printed pages (excluding the supplement or appendices), existence of a supplement or appendix, and the number of tables and figures (including the supplement or appendices).

Statistical aspects

We evaluated whether the authors applied a Bayesian approach, because this affects the reporting of results such as treatment rankings. In that case, we also assessed whether the authors used non-informative or informative priors for the basic parameters and the between trial variance.18 We also assessed whether the authors reported a method to investigate non-transitivity, a key to the validity of a network meta-analysis, which occurs when studies are not sufficiently similar in important clinical and methodological characteristics (effect modifiers).11 In accordance with Donegan and colleagues,19 we considered two methods: comparing patients’ or trials’ characteristics across treatment comparisons (for example, using descriptive statistics describing the characteristics across the comparisons) and investigating potential treatment effect modifying covariates (for example, using subgroup analysis or network meta-regression analysis).

Presentation of results

In accordance with recommendations from Ioannidis,20 Mills and colleagues,21 Cipriani and colleagues,11 and Ades and colleagues,12 we chose items to assess the reporting of the amount of evidence in the network of randomised trials, and of effect size estimates for pairwise comparisons between interventions. Furthermore, we assessed the reporting of rank ordering of interventions.

Amount of evidence in network of randomised trials

We assessed whether the reports included data to allow for gauging of the description of the network and its geometry. In particular, we assessed whether each report allowed for explicit identification of the interventions included in the network, the direct comparisons between the interventions, and the number of randomised trials and patients for each comparison. We also assessed whether a network graph or diagram was reported. In cases of star networks (that is, all interventions were compared with a common comparator but not against each other) and fully connected networks (all interventions were compared with each other), we considered that the network structure was clear provided that the authors explicitly mentioned the interventions and direct comparisons between the interventions included in the network.

Effect size estimates for pairwise comparisons between interventions

We evaluated whether the following components were reported: results of all individual randomised trials used in the network meta-analysis (summary data for each intervention group or estimates of treatment effect) and effect estimates with measures of uncertainty (such as confidence intervals, standard error, variance) from pairwise meta-analyses (direct evidence), indirect comparison meta-analyses (indirect evidence), or network meta-analyses (mixed evidence). In the case of the last of these, we assessed the reporting of all pairwise comparisons between interventions or only some selected pairwise comparisons between interventions (for example, all comparisons with placebo or reference intervention). We also assessed how these effect sizes were reported. Possible formats included a table, a matrix, a forest plot, a network graph (each edge labelled with the corresponding effect size), or narrative reporting in the results section.

Rank ordering of interventions

We assessed whether the reports gave some form of rank ordering of interventions. We assessed whether the authors presented the probabilities of each intervention being the best, the probabilities of each intervention taking each possible rank (in an absolute or cumulative ranking curve), or the mean ranks (or the surface under the cumulative ranking curve) and whether authors stated which intervention was the best. We also assessed whether the authors considered the uncertainty in rank ordering (by reporting confidence/credible intervals for the ranks or by reporting all probabilities for each intervention with each possible rank, in an absolute or cumulative ranking curve22).

Composite outcome assessing inadequate reporting

We built a composite outcome describing inadequate reporting of results based on guidelines from Ioannidis,20 Mills and colleagues,21 Ades and colleagues,12 and Cipriani and colleagues.11 We considered reporting to be inadequate in the case of absence of one of the following items: no reporting of interventions included in the network, direct comparisons between interventions, and number of randomised trials and patients for each comparison; or no reporting of results from direct, indirect, or mixed evidence.

Statistical analysis

We summarised quantitative data as medians and interquartile ranges and categorical data as numbers and percentages. We used the Fisher exact test for categorical data to analyse the composite outcome. All tests were two sided, and we considered P<0.05 to be significant. We used SAS 9.3 for statistical analyses.

Results

General characteristics

We identified 121 eligible reports of network meta-analyses (table 1): 55 (45%) published in general journals and 66 (55%) in specialty journals; 55 (45%) network meta-analyses were funded by public sources and 48 (40%) by at least one private source. A total of 100 (83%) reports assessed pharmacological interventions. The median number of outcomes assessed per network meta-analysis was 2 (interquartile range 1-3). In all, 70 (58%) reports of network meta-analyses included a supplement or appendix.
Table 1

 General characteristics of 121 reports of network meta-analyses

Item and subcategoryNo (%) of reports
Journal type:
 General journal55 (45)
 Specialty journal66 (55)
Funding source:
 Private (at least one private source)48 (40)
 Public55 (45)
 Unclear18 (15)
Type of intervention assessed:
 Pharmacological intervention100 (83)
 Non-pharmacological intervention11 (9)
 Both (pharmacological and non-pharmacological intervention) 10 (8)
Bayesian statistical approach:91 (75)
 Basic parameter:
  Non-informative prior51/91 (56)
  Informative prior0 (0)
  Not reported40/91 (44)
 Between trial variance:
  Non-informative prior30/91 (33)
  Informative prior4/91 (4)
  Not reported57/91 (63)
Methods to assess non-transitivity:43 (36)
 Comparing patients’ or trials’ characteristics (across comparisons)40/43 (93)
 Investigating potential intervention effect modifying covariates3/43 (7)
Type of outcomes:
 Binary 78 (64)
 Continuous 28 (23)
 Time to event:15 (12)
No of outcomes assessed per network*2 (1-3)
No of printed pages*11 (9-15)
Supplement or appendix published70 (58)
No of tables per network*5 (3-6)
No of figures per network*3 (2-5)

*Data are medians (interquartile range).

General characteristics of 121 reports of network meta-analyses *Data are medians (interquartile range).

Statistical aspects

Overall, 91 (75%) network meta-analyses applied a Bayesian approach: 51 used non-informative priors for the basic parameters, and 30 used them for the between trial variance. In all, 43 (36%) used a method to assess non-transitivity.

Amount of evidence in the network of randomised trials

Of 121 network meta-analyses, 52 (43%) did not report the number of randomised trials for each comparison and 96 (79%) did not report the number of patients (table 2). Overall, 100 (83%) network meta-analyses did not give a description of the network (that is, the interventions included in the network, direct comparisons between interventions, number of randomised trials and patients for each comparison) and its geometry (a network graph or information needed to know the network structure) (table 2, figure, appendix table A).
Table 2

 Presentation of results in 121 reports of network meta-analysis. Values are numbers (percentages)

Items—yes if reportedOverall (n=121)General journals (n=55)Specialty journals (n=66)
Amount of evidence in network of randomised trials
Interventions included in network121 (100)55 (100) 66 (100)
Direct comparisons between interventions82 (68)39 (71)43 (65)
No of trials for each comparison69 (57)34 (62)35 (53)
No of patients for each comparison25 (21)14 (25)11 (17)
Network graph*89 (74)44 (80)45 (68)
Effect size estimates for pairwise comparisons between interventions
Results of all individual trials used in network meta-analysis68 (56)28 (51)40 (61)
Direct evidence 73 (60)33 (60)40 (61)
Indirect evidence13 (11)6 (11)7 (11)
Mixed evidence78 (64)34 (62)44 (67)
Effect estimates for some selected pairwise comparisons between interventions from network meta-analysis43 (36)21 (38)22 (33)
Format of results†:78 (64)29 (53)49 (74)
 Table41 (34)24 (44)17 (26)
 Forest plot7 (6)3 (5)4 (6)
 Matrix6 (5)3 (5)3 (5)
 Figure4 (3)2 (4)2 (3)
 Text
Rank order of interventions
Rank order of interventions‡52/91 (57)22/38 (58)30/53 (57)
Rank order of interventions with uncertainty9/52 (17)6/22 (27)3/30 (10)

*7 star networks and 4 fully connected networks reported no network graph, but authors explicitly mentioned interventions and direct comparisons in network.

†Multiple answers were possible, so total does not equal 100%.

‡Network meta-analyses that did not use Bayesian approach cannot derive intervention rankings; 91 network meta-analyses used Bayesian approach.

Reporting items for results of network meta-analyses. The idea is to show which items were reported in each of the 121 network meta-analyses. Each horizontal line of the gap chart corresponds to one network meta-analysis report. A specific colour was attributed to each of the 10 items studied. The colour bands show which of these items (labelled at the top) were reported for each network meta-analysis. Items were grouped into three categories: 4 items in blue pertain to the description of network, and 1 item in purple pertains to its geometry (network graph); 4 items in green pertain to effect size estimates for pairwise comparisons between interventions; 1 item pertains to intervention ranking. The 121 network meta-analysis reports were sorted according to the total number of reported items, in decreasing order. The diagram on the left shows the distribution of the total number of reported items across the 121 network meta-analyses. The diagram at the bottom shows the proportion of network meta-analysis reports that reported each item

Presentation of results in 121 reports of network meta-analysis. Values are numbers (percentages) *7 star networks and 4 fully connected networks reported no network graph, but authors explicitly mentioned interventions and direct comparisons in network. †Multiple answers were possible, so total does not equal 100%. ‡Network meta-analyses that did not use Bayesian approach cannot derive intervention rankings; 91 network meta-analyses used Bayesian approach. Reporting items for results of network meta-analyses. The idea is to show which items were reported in each of the 121 network meta-analyses. Each horizontal line of the gap chart corresponds to one network meta-analysis report. A specific colour was attributed to each of the 10 items studied. The colour bands show which of these items (labelled at the top) were reported for each network meta-analysis. Items were grouped into three categories: 4 items in blue pertain to the description of network, and 1 item in purple pertains to its geometry (network graph); 4 items in green pertain to effect size estimates for pairwise comparisons between interventions; 1 item pertains to intervention ranking. The 121 network meta-analysis reports were sorted according to the total number of reported items, in decreasing order. The diagram on the left shows the distribution of the total number of reported items across the 121 network meta-analyses. The diagram at the bottom shows the proportion of network meta-analysis reports that reported each item

Effect size estimates for pair-wise comparisons between interventions

Of 121 reports of network meta-analyses, 53 (44%) did not give the results of all individual randomised trials used in the network meta-analysis. A total of 48 (40%) did not report the effect sizes for all pairwise comparisons between interventions with a measure of uncertainty derived from pairwise meta-analyses (that is, direct evidence), and 108 (89%) did not give indirect evidence (the effect sizes for all pairwise comparisons between interventions with a measure of uncertainty from indirect evidence). In all, 109 (90%) reports did not give both direct and indirect evidence and 71 (59%) did not give both direct and mixed evidence (that is, the effect sizes for all pairwise comparisons between interventions with a measure of uncertainty from network meta-analyses); 112 (93%) did not describe direct as well as indirect and mixed evidence (table 2, figure , appendix table A).

Rank order of interventions

Among 91 network meta-analyses with a Bayesian approach, 52 reports gave the rank order of interventions; 40 of these gave the probabilities of each intervention being the best, 7 used absolute or cumulative rankograms, and 5 stated which intervention was the best. However, 43 of the 52 reports did not report a measure of uncertainty in ranking (table 2). When reported, the uncertainty in ranking was assessed by absolute or cumulative rankograms (7/9) or confidence/credible intervals (2/9). Overall, 119 (98%) reports of network meta-analyses showed inadequate reporting of results (that is, the authors did not report the interventions included in the network or direct comparisons between the interventions; number of randomised trials and patients for each comparison; or results from direct, indirect, or mixed evidence). This finding did not differ by journal type (general journals: 100%, 95% confidence interval 94% to 100%; specialty journals: 97%, 93% to 100%; P=0.5) or funding source (public: 98%, 94% to 100%; private: 100%, 93% to 100%; P=0.53).

Discussion

Our sample of 121 network meta-analyses included a large number of recently published network meta-analyses covering a wide range of medical areas. Essential components of reporting results of network meta-analyses were missing in most of the articles. An important element in understanding a network meta-analysis is assessing the network geometry.20 21 Firstly, knowing precisely the shape of the network (that is, which interventions are included in the network and which interventions have been compared against each other in randomised trials) is a preliminary step for researchers and readers to assess the available evidence base. The geometry of the network may show that some (head to head) comparisons have been ignored or that some comparators were preferred.23 24 25 26 Identifying trials that have not been conducted may help in designing the research agenda. Secondly, the reader should consider the number of randomised trials and patients for each comparison. In fact, differences in the numbers and sizes of trials across interventions and comparisons may affect the reliability of some network meta-analyses’ estimates.27 The amount of available evidence may be imbalanced because of differential reporting bias across comparisons, which may substantially affect the results.28 Thirdly, readers should be able to assess whether similar but not identical interventions were grouped, which may weaken the interpretability of a network meta-analysis. For transparency, authors of network meta-analyses should clearly report the interventions included in the network, the direct comparisons between the interventions, and the number of randomised trials and patients for each comparison and should give a graphical representation of the network.11 However, only 17% of our reports of network meta-analyses adequately described the network and its geometry. The most important element for readers may be the assessment of the relative effectiveness of the considered interventions. When direct and indirect evidence is available, a potential difficulty is choosing among direct, indirect, and mixed evidence, especially with discrepancies between direct and indirect evidence. Checking consistency is a preliminary step before proceeding to the network meta-analysis.29 Empirical studies have detected statistically significant inconsistency between direct and indirect comparisons in 14% of 112 single loops of evidence and 2-9% of 303 loops from complete networks, depending on the measure of effect and the method for estimating heterogeneity.30 31 However, excluding inconsistency on the basis of the results of statistical tests is difficult because the results may suggest no major inconsistency although it may be present. Moreover, statistically non-significant inconsistency does not necessarily imply clinical consistency.30 Consequently, readers should be able to compare effect estimates from direct, indirect, and mixed evidence, even in the absence of statistical inconsistency. Indirect evidence may be considered of lower quality than evidence from head to head trials because it is prone to similarity/transitivity problems.32 However, in some cases, indirect evidence may be more reliable than direct and mixed evidence,33 34 35 and when trials are similarly biased an indirect comparison is not biased.36 Transparency is important to allow readers to make their own judgment according to explicitly presented discrepancies. However, only 10% of our network meta-analysis reports gave both direct and indirect evidence and only 7% gave direct, indirect, and mixed evidence. One appealing feature of a network meta-analysis is the rank ordering of interventions. This information is attractive for clinical researchers and physicians because it seems to respond to their main concern: determining the best available intervention. However, an intervention may be ranked highly even though it was assessed in only a few trials and with a few patients, which resulted in a misleadingly strong endorsement for the intervention despite large uncertainty. Discussion is ongoing as to whether rankings must be reported and are appropriate to derive inference. Clinical researchers and physicians should always be able to examine the uncertainty in the ranking of interventions, because the difference between interventions might be small and not clinically relevant. However, only 17% of our network meta-analysis reports that ranked interventions reported the uncertainty.

Strengths and limitations of study

To the best of our knowledge, no previous study has tackled the problem of presentation of results in a large sample, such as 121 network meta-analyses. Recently, the Agency for Healthcare Research and Quality assessed the statistical methods used in a sample of 34 closed loop Bayesian mixed treatment comparisons,15 37 and Tan et al explored the presentational approaches used in the United Kingdom for reporting evidence synthesis of 19 reports of studies involving indirect and mixed treatment comparisons.38 These authors focused mainly on the statistical methods or possible formats of results but did not assess reporting of the amount of evidence in the network of trials (that is, the interventions included in the network, direct comparisons between the interventions, number of trials and patients for each comparison, network graph/or diagram) or the reporting of effect estimates with uncertainty from direct, indirect, or mixed evidence. Our study had some limitations. Several journals have space constraints, so the study authors may have omitted important details from their reports or key information may have been deleted during the publication process. Nevertheless, most journals now allow an online appendix for extended descriptions of methods and results, and we assessed all data from both the original reports of network meta-analyses and the supplementary appendices when available. Another limitation is that we searched only for reports of network meta-analyses published in journals and did not search for health technology assessment (HTA) reports. Network meta-analyses are increasingly being used to support HTAs. However, identifying network meta-analyses in HTA reports by usual search strategies is difficult. The reporting of results may differ in HTA reports and journal articles. Finally, we built a composite outcome based on the available recommendations from Ioannidis,20 Mills and colleagues,21 Ades and colleagues,12 and Cipriani and colleagues,11 as a possible core set of information that we would like to see in any report of a network meta-analysis. However, we acknowledge that consensus is lacking on which items should be required in network meta-analysis reports. This is an area of ongoing debate.

Conclusions

Our study shows that the results of network meta-analyses are heterogeneously reported. These results might reflect the fact that we lack a general consensus on what should be reported in a network meta-analysis. This review reinforces the need to develop reporting guidelines for network meta-analyses to facilitate their reporting. It may provide a basis for the future development of a quality assessment tool for reporting network meta-analyses and aiding appropriate interpretation. The findings from network meta-analyses may be difficult for clinicians and decision makers to interpret Inadequate reporting of results may affect the interpretation of network meta-analyses and mislead clinical researchers Reports of network meta-analysis should allow the reader to assess the amount of evidence in the network of trials and the relative effectiveness of the considered interventions The findings from network meta-analyses are heterogeneously reported These results might reflect the lack of a general consensus on what should be reported in a network meta-analysis A clear need exists to extend reporting guidelines to network meta-analyses to improve the reporting of network meta-analyses
  37 in total

1.  Network meta-analysis for indirect treatment comparisons.

Authors:  Thomas Lumley
Journal:  Stat Med       Date:  2002-08-30       Impact factor: 2.373

2.  Lack of head-to-head trials and fair control arms: randomized controlled trials of biologic treatment for rheumatoid arthritis.

Authors:  Candice Estellat; Philippe Ravaud
Journal:  Arch Intern Med       Date:  2012-02-13

3.  Graphical methods and numerical summaries for presenting results from multiple-treatment meta-analysis: an overview and tutorial.

Authors:  Georgia Salanti; A E Ades; John P A Ioannidis
Journal:  J Clin Epidemiol       Date:  2010-08-05       Impact factor: 6.437

4.  Exploring the geometry of treatment networks.

Authors:  Georgia Salanti; Fotini K Kavvoura; John P A Ioannidis
Journal:  Ann Intern Med       Date:  2008-04-01       Impact factor: 25.391

5.  Impact of evergreening on patients and health insurance: a meta analysis and reimbursement cost analysis of citalopram/escitalopram antidepressants.

Authors:  Ali A Alkhafaji; Ludovic Trinquart; Gabriel Baron; Moïse Desvarieux; Philippe Ravaud
Journal:  BMC Med       Date:  2012-11-20       Impact factor: 8.775

Review 6.  Indirect comparisons: a review of reporting and methodological quality.

Authors:  Sarah Donegan; Paula Williamson; Carrol Gamble; Catrin Tudur-Smith
Journal:  PLoS One       Date:  2010-11-10       Impact factor: 3.240

7.  Impact of reporting bias in network meta-analysis of antidepressant placebo-controlled trials.

Authors:  Ludovic Trinquart; Adeline Abbé; Philippe Ravaud
Journal:  PLoS One       Date:  2012-04-20       Impact factor: 3.240

8.  Evidence synthesis for decision making 7: a reviewer's checklist.

Authors:  A E Ades; Deborah M Caldwell; Stefanie Reken; Nicky J Welton; Alex J Sutton; Sofia Dias
Journal:  Med Decis Making       Date:  2013-07       Impact factor: 2.583

9.  Evaluation of inconsistency in networks of interventions.

Authors:  Areti Angeliki Veroniki; Haris S Vasiliadis; Julian P T Higgins; Georgia Salanti
Journal:  Int J Epidemiol       Date:  2013-02       Impact factor: 7.196

Review 10.  Analysis of the systematic reviews process in reports of network meta-analyses: methodological systematic review.

Authors:  Aïda Bafeta; Ludovic Trinquart; Raphaèle Seror; Philippe Ravaud
Journal:  BMJ       Date:  2013-07-01
View more
  54 in total

1.  Critical Appraisal of Published Indirect Comparisons and Network Meta-Analyses of Competing Interventions for Multiple Myeloma.

Authors:  Shannon Cope; Kabirraaj Toor; Evan Popoff; Rafael Fonseca; Ola Landgren; María-Victoria Mateos; Katja Weisel; Jeroen Paul Jansen
Journal:  Value Health       Date:  2020-04-06       Impact factor: 5.725

2.  CORR Insights(®): There Are No Differences in Short- to Mid-term Survivorship Among Total Hip-bearing Surface Options: A Network Meta-analysis.

Authors:  Raphaël Porcher
Journal:  Clin Orthop Relat Res       Date:  2015-01-23       Impact factor: 4.176

3.  Borrowing of strength from indirect evidence in 40 network meta-analyses.

Authors:  Lifeng Lin; Aiwen Xing; Michael J Kofler; Mohammad Hassan Murad
Journal:  J Clin Epidemiol       Date:  2018-10-17       Impact factor: 6.437

4.  Quantifying and presenting overall evidence in network meta-analysis.

Authors:  Lifeng Lin
Journal:  Stat Med       Date:  2018-07-18       Impact factor: 2.373

5.  Evidence inconsistency degrees of freedom in Bayesian network meta-analysis.

Authors:  Lifeng Lin
Journal:  J Biopharm Stat       Date:  2020-12-09       Impact factor: 1.051

6.  The effectiveness of different interventions to promote poison prevention behaviours in households with children: a network meta-analysis.

Authors:  Felix A Achana; Alex J Sutton; Denise Kendrick; Persephone Wynn; Ben Young; David R Jones; Stephanie J Hubbard; Nicola J Cooper
Journal:  PLoS One       Date:  2015-04-20       Impact factor: 3.240

Review 7.  Network meta-analysis using R: a review of currently available automated packages.

Authors:  Binod Neupane; Danielle Richer; Ashley Joel Bonner; Taddele Kibret; Joseph Beyene
Journal:  PLoS One       Date:  2014-12-26       Impact factor: 3.240

8.  Monotherapy for Alopecia Areata: A Systematic Review and Network Meta-Analysis.

Authors:  Aditya K Gupta; Jessie L Carviel; Kelly A Foley; Neil H Shear; Bianca Maria Piraccini; Vincent Piguet; Antonella Tosti
Journal:  Skin Appendage Disord       Date:  2019-08-29

9.  Mesh fixation methods in open inguinal hernia repair: a protocol for network meta-analysis and trial sequential analysis of randomised controlled trials.

Authors:  Long Ge; Jin-hui Tian; Lun Li; Quan Wang; Ke-hu Yang
Journal:  BMJ Open       Date:  2015-11-19       Impact factor: 2.692

10.  The Global Research Collaboration of Network Meta-Analysis: A Social Network Analysis.

Authors:  Lun Li; Ferrán Catalá-López; Adolfo Alonso-Arroyo; Jinhui Tian; Rafael Aleixandre-Benavent; Dawid Pieper; Long Ge; Liang Yao; Quan Wang; Kehu Yang
Journal:  PLoS One       Date:  2016-09-29       Impact factor: 3.240

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.