Literature DB >> 24363797

Publication bias in recent meta-analyses.

Michal Kicinski1.   

Abstract

INTRODUCTION: Positive results have a greater chance of being published and outcomes that are statistically significant have a greater chance of being fully reported. One consequence of research underreporting is that it may influence the sample of studies that is available for a meta-analysis. Smaller studies are often characterized by larger effects in published meta-analyses, which can be possibly explained by publication bias. We investigated the association between the statistical significance of the results and the probability of being included in recent meta-analyses.
METHODS: For meta-analyses of clinical trials, we defined the relative risk as the ratio of the probability of including statistically significant results favoring the treatment to the probability of including other results. For meta-analyses of other studies, we defined the relative risk as the ratio of the probability of including biologically plausible statistically significant results to the probability of including other results. We applied a Bayesian selection model for meta-analyses that included at least 30 studies and were published in four major general medical journals (BMJ, JAMA, Lancet, and PLOS Medicine) between 2008 and 2012.
RESULTS: We identified 49 meta-analyses. The estimate of the relative risk was greater than one in 42 meta-analyses, greater than two in 16 meta-analyses, greater than three in eight meta-analyses, and greater than five in four meta-analyses. In 10 out of 28 meta-analyses of clinical trials, there was strong evidence that statistically significant results favoring the treatment were more likely to be included. In 4 out of 19 meta-analyses of observational studies, there was strong evidence that plausible statistically significant outcomes had a higher probability of being included.
CONCLUSIONS: Publication bias was present in a substantial proportion of large meta-analyses that were recently published in four major medical journals.

Entities:  

Mesh:

Year:  2013        PMID: 24363797      PMCID: PMC3868709          DOI: 10.1371/journal.pone.0081823

Source DB:  PubMed          Journal:  PLoS One        ISSN: 1932-6203            Impact factor:   3.240


Introduction

When some study outcomes are more likely to be published than other, the literature that is available to doctors, scientists, and policy makers provides misleading information. The tendency to decide to publish a study based on its results has been long acknowledged as a major threat to the validity of conclusions from medical research[1,2]. During the past 25 years, the phenomenon of research underreporting has been extensively investigated. It is clear that statistically significant results supporting the hypothesis of the researcher often have a greater chance of being published and fully reported[3-7]. Meta-analysis, a statistical approach to estimate a parameter of interest based on multiple studies, plays an essential role in medical research. One consequence of research underreporting is that it influences the sample of studies that is available for a meta-analysis[8,9]. This causes a bias, unless the process of study selection is modeled correctly[10]. Such modeling requires strong assumptions about the nature of the publication bias, especially when the size of a meta-analysis is not very large and when robust techniques cannot be used[11-13]. As a result, when publication bias occurs, the validity of the meta-analysis is uncertain. It is well-known that smaller studies are often characterized by larger effects in published meta-analyses[14-16]. Publication bias is one of the possible explanations of this phenomenon[17]. Although a meta-analysis is typically preceded by an investigation of the presence of publication bias, the standard detection methods are characterized by a low power[11,18-22]. Therefore, the sample of included studies may be unrepresentative of the population of all conducted studies even when publication bias has not been detected. In this study, we investigated whether statistically significant outcomes that showed a positive effect of the treatment (in the case of clinical trials) and plausible statistically significant outcomes (in the case of observational studies and interventional studies) had a greater probability of being included in recent meta-analyses than other outcomes. We considered all meta-analyses of aggregate data that included at least 30 effect sizes from individual studies and were published between 2008 and 2012 in four major general medical journals: BMJ, JAMA, Lancet, and PLOS Medicine. We applied a Bayesian approach, which allows estimation of a parametric function that describes the selection process[23].

Methods

For meta-analyses of clinical trials, we a priori decided to estimate the ratio of the probability of including statistically significant results favoring the treatment to the probability of including other results. For other meta-analyses, we a priori decided to estimate the ratio of the probability of including plausible statistically significant results to the probability of including other results. The definition of plausibility was straightforward and a priori chosen. For meta-analyses of an association between a risk factor and an undesired outcome (disease, mortality, etc.), results showing a positive association were regarded as plausible. For meta-analyses of an association between the absence of a risk factor and an undesired outcome, results showing a negative association were regarded as plausible. In meta-analyses of associations between alcohol consumption and cardiovascular parameters, we estimated the ratio of the probability of including statistically significant results to the probability of including not statistically significant results because both a positive and negative association is biologically plausible[24,25]. A two-sided significance level of 0.05 was assumed. Further in the article, we refer to the ratio of probabilities as the relative risk (RR).

Identification of meta-analyses

We used PubMed to identify meta-analyses published between 2008 and 2012 in four general medical journals (BMJ, JAMA, Lancet, and PLOS Medicine). The term ‘meta-analysis’ was required to appear in the title or the abstract. Meta-analyses that combined at least 30 estimates from individual studies were considered. Only large meta-analyses were included because a substantial sample size is required to distinguish between the effects of heterogeneity and publication bias[11,26]. We considered only meta-analyses of aggregate data because a selection model based on p-values seemed to be inappropriate to study the complicated process of study selection in individual participant data meta-analyses.

Statistical model

For each identified meta-analysis, we applied a hierarchical selection model[23,27]. In this approach, a weight function is incorporated in a meta-analysis to model the probability that an individual study is included. Similarly as in the case of a standard random effects meta-analysis, the conditional distribution of the observed study effects Yi (i=1, …, N), given the true study effect αi and the within-study variance σi 2, is assumed to be normal: f(yi|αi,σi 2)~N(αi,σi 2). The true study effect comes from a normal distribution: N(µ,τ2), were µ is the mean effect size and τ2 is the between-study variance. When a selection process is present some studies may fail to enter the meta-analysis sample. To model the process of study selection, the probability that any specific study enters the sample is assumed to be multiplied by a nonnegative weight function. In this case, the observed study effects Xj that enter the meta-analysis sample (j=1, …, n, n≤N) have a weighted density: fw(xj|αj,σj 2)=w(xj)f(xj|αj,σj 2)/∫w(x)f(x|αj,σj 2)dx, where w(x) is a weight function. We applied a step weight function that took two values: ϒ1 for statistically significant results favoring the treatment (in the case of clinical trials) or plausible statistically significant results (in the case of other studies) and ϒ2 for other results, so that the RR equaled ϒ1/ϒ2 Maximum likelihood estimation is one possible approach to fit the model described above[27,28]. We used Bayesian inference because it produces valid results when the sample size is small[29], allows a straightforward interval estimation, and examination of the sensitivity of the findings to the distribution of the random effects. Similarly to Silliman[23], we used diffuse uniform priors U(0,1) for the parameters of the weight function. We declared a diffuse prior N(0,1000) for the mean effect size and, following a recommendation of Gelman[30], a uniform prior for the between-study standard deviation. We used Gibbs sampling [31] to obtain samples from the posterior distribution of ϒ1/ϒ2. We applied the algorithm described by Silliman, who considered a general class of hierarchical selection models[23]. In our specific case, the full conditional distributions needed by the Gibbs sampler were: , and where, x= (x1, ..., xn), σ= (σ1, ..., σn), α= (α1, ..., αn), ϒ=(ϒ1,ϒ2), and s is the number of statistically significant results favoring the treatment (in the case of meta-analyses of clinical trials) or plausible statistically significant results (in the case of meta-analysis of other studies). A burn-in of 10000 iterations was sufficient to achieve convergence for all meta-analyses. The estimates were based on the subsequent 50000 iterations. For point estimation, we used the median of the posterior distribution of RR. For interval estimation, we used the 95% equal-tail credible interval (CI). When the posterior probability that the RR exceeded 1 was greater than 0.95, we concluded that there was strong evidence that the RR was greater than 1. An R program that was used to fit the models can be found in Appendix S1.

Quality of the statistical model

In order to evaluate whether the statistical model was suitable for the objectives of the study, we performed simulations. The settings were based on the characteristics of the meta-analyses of clinical trials included in the study (see Appendix S2). The posterior distribution provided reliable information about the size of the RR (Table 1). Although the estimate of the RR based on the median of the posterior distribution was characterized by a substantial variability and was biased in some scenarios, it gave a correct idea about the existence of a publication bias. The model tended to underestimate the RR when the mean effect size was small and overestimate the RR when the mean effect size was large. However, it was able to distinguish the RR from the mean effect size well for all simulation settings, as indicated by the quality of the interval estimates of the RR. For almost all scenarios, the lower bound of the 95% equal-tail credible interval was smaller than the assumed true value of the RR in at least 95% of the simulations. Although the upper bound was sometimes too small when the mean effect size was small, it was greater than the true RR in more than 95% of the simulations for most of the scenarios (Table 1). The performance of the model did not depend on the size of the between-study heterogeneity. The model was robust to the presence of small study effects (i.e., larger true effects in smaller studies, Table 1).
Table 1

Model performance: quality of estimation.

RR SSE I2 µ Bias ME MSE LB UB Total
1No0.360.0-0.070.360.240.9991.0000.999
1No0.360.2-0.030.330.210.9950.9950.990
1No0.360.80.090.320.220.9810.9890.970
1No0.361.40.210.380.380.9650.9940.959
1No0.670.0-0.020.350.270.9960.9960.992
1No0.670.20.010.330.220.9940.9940.988
1No0.670.80.100.320.240.9840.9930.977
1No0.671.40.250.420.450.9660.9940.960
4No0.360.0-1.451.985.210.9970.8530.850
4No0.360.2-1.071.713.960.9930.8550.848
4No0.360.80.261.868.320.9820.9520.934
4No0.361.40.932.7428.90.9780.9670.945
4No0.670.0-1.061.935.560.9960.8760.872
4No0.670.2-0.891.714.130.9970.9030.900
4No0.670.80.652.1513.40.9710.9680.939
4No0.671.41.412.9737.90.9780.9770.955
10No0.360.0-3.384.5126.50.9960.8310.827
10No0.360.2-2.524.1424.50.9950.8710.866
10No0.360.82.636.992090.9790.9540.933
10No0.361.40.526.3375.01.0000.9620.962
10No0.670.0-2.904.5228.60.9970.8660.863
10No0.670.2-1.794.4230.50.9940.9020.896
10No0.670.84.528.352690.9730.9700.943
10No0.671.42.727.641530.9950.9680.963
1Yes0.360.0-0.030.420.380.9980.9980.996
1Yes0.360.20.070.400.410.9870.9920.979
1Yes0.360.80.300.420.410.9500.9950.945
1Yes0.361.40.460.530.640.9310.9980.929
1Yes0.670.00.050.410.370.9960.9970.993
1Yes0.670.20.060.380.320.9920.9920.984
1Yes0.670.80.250.400.390.9650.9990.964
1Yes0.671.40.460.540.740.9420.9970.939
4Yes0.360.0-1.172.106.150.9970.8780.875
4Yes0.360.2-0.411.835.790.9890.9220.911
4Yes0.360.81.592.5019.30.9470.9820.929
4Yes0.361.42.393.3441.10.9680.9910.959
4Yes0.670.0-0.792.006.280.9950.9090.904
4Yes0.670.2-0.511.745.060.9920.9350.927
4Yes0.670.81.652.6420.40.9570.9810.938
4Yes0.671.42.573.5745.90.9560.9870.943
10Yes0.360.0-2.004.7237.30.9920.8880.880
10Yes0.360.2-0.694.2230.90.9860.9050.891
10Yes0.360.86.569.003500.9590.9910.950
10Yes0.361.43.467.231191.0000.9870.987
10Yes0.670.0-1.174.3530.00.9960.9180.914
10Yes0.670.2-0.414.3935.30.9880.9370.925
10Yes0.670.86.449.353790.9560.9830.939
10Yes0.671.45.669.382140.9950.9890.984

RR: relative risk, the ratio of the probability of including statistically significant outcomes favoring the treatment to the probability of including other outcomes (for RR=1, all results had the same probability of being included), SSE: small study effect, I2: proportion of the total variability due to heterogeneity, µ: mean effect size, Bias: average difference between the median of the posterior distribution of the RR and the true RR, ME: mean error, MSE: mean squared error, LB: proportion of the lower bounds of the 95% equal-tail credible intervals lower than the true RR, UB: proportion of the upper bounds of the 95% equal-tail credible intervals greater than the true RR, Total: proportion of the 95% equal-tail credible intervals including the true RR.

RR: relative risk, the ratio of the probability of including statistically significant outcomes favoring the treatment to the probability of including other outcomes (for RR=1, all results had the same probability of being included), SSE: small study effect, I2: proportion of the total variability due to heterogeneity, µ: mean effect size, Bias: average difference between the median of the posterior distribution of the RR and the true RR, ME: mean error, MSE: mean squared error, LB: proportion of the lower bounds of the 95% equal-tail credible intervals lower than the true RR, UB: proportion of the upper bounds of the 95% equal-tail credible intervals greater than the true RR, Total: proportion of the 95% equal-tail credible intervals including the true RR. Additionally, we compared the ability of our model to detect a selection process based on the statistical significance with publication bias methods widely used in medical research: the Egger’s test[32], correlation test of Begg and Mazumdar[33], and the trim and fill method[34]. When statistically significant outcomes had a higher probability of being included, the Bayesian selection model showed much higher detection rates compared to the standard methods (Table 2). This difference was especially apparent when small study effects were absent. Furthermore, in contrast to the standard methods, the Bayesian selection model was characterized by low false positive rates, even in the presence of small-study effects (Table 2).
Table 2

Model performance: type 1 error and power.

RR SSE I2 µ Current model Egger Rank correlation Trim and Fill
1No0.360.00.0080.0480.0190.027
1No0.360.20.0120.0590.0270.032
1No0.360.80.0390.0780.0320.066
1No0.361.40.0580.0840.0440.041
1No0.670.00.0090.0470.0190.012
1No0.670.20.0090.0540.0230.031
1No0.670.80.0300.0850.0260.045
1No0.671.40.0560.0930.0380.046
4No0.360.00.3080.0490.0170.043
4No0.360.20.5600.0680.0240.055
4No0.360.80.8290.5730.4460.231
4No0.361.40.7110.3960.4230.277
4No0.670.00.4150.0300.0040.016
4No0.670.20.5670.0510.0060.028
4No0.670.80.7940.4710.2730.170
4No0.671.40.7300.3700.2880.225
10No0.360.00.9220.0840.0290.073
10No0.360.20.9790.3910.2740.111
10No0.360.80.9890.8160.7830.470
10No0.361.40.9530.5880.6080.527
10No0.670.00.9390.0610.0200.029
10No0.670.20.9710.3090.1770.061
10No0.670.80.9860.7330.6370.354
10No0.671.40.9580.4920.5120.401
1Yes0.360.00.0070.1900.0890.094
1Yes0.360.20.0200.2330.0930.098
1Yes0.360.80.0980.2410.1310.154
1Yes0.361.40.1310.2400.1260.140
1Yes0.670.00.0090.1420.0450.070
1Yes0.670.20.0200.1700.0530.069
1Yes0.670.80.0600.1930.0670.112
1Yes0.671.40.1050.2160.0780.108
4Yes0.360.00.3040.2410.1150.165
4Yes0.360.20.5920.2710.0850.152
4Yes0.360.80.9260.8400.6360.432
4Yes0.361.40.8940.6800.6150.467
4Yes0.670.00.4240.0920.0280.071
4Yes0.670.20.6030.1410.0180.087
4Yes0.670.80.8900.6800.4390.261
4Yes0.671.40.8460.5500.4240.345
10Yes0.360.00.8850.3820.1630.252
10Yes0.360.20.9860.5800.3210.247
10Yes0.360.80.9990.9500.9230.641
10Yes0.361.40.9880.8070.7840.689
10Yes0.670.00.9560.1590.0450.088
10Yes0.670.20.9880.4000.2270.134
10Yes0.670.80.9940.8680.7910.492
10Yes0.671.40.9830.7150.6360.538

RR: relative risk, the ratio of the probability of including statistically significant outcomes favoring the treatment to the probability of including other outcomes (for RR=1, all results had the same probability of being included), SSE: small study effect, I2: proportion of the total variability due to heterogeneity, µ: mean effect size. Proportion of meta-analyses, in which publication bias was identified, is presented. For the Bayesian selection model, publication bias was indicated when the posterior probability that the RR was larger than 1 exceeded 95%. For the Egger’s test and the rank correlation test, one-sided procedures were used with a 0.05 significance level. For the trim and fill method, publication bias was indicated when the number of missing studies estimated by the R estimator in the first step of the algorithm was greater than 3[34].

RR: relative risk, the ratio of the probability of including statistically significant outcomes favoring the treatment to the probability of including other outcomes (for RR=1, all results had the same probability of being included), SSE: small study effect, I2: proportion of the total variability due to heterogeneity, µ: mean effect size. Proportion of meta-analyses, in which publication bias was identified, is presented. For the Bayesian selection model, publication bias was indicated when the posterior probability that the RR was larger than 1 exceeded 95%. For the Egger’s test and the rank correlation test, one-sided procedures were used with a 0.05 significance level. For the trim and fill method, publication bias was indicated when the number of missing studies estimated by the R estimator in the first step of the algorithm was greater than 3[34].

Sensitivity analysis

In order to investigate the robustness of the findings, two alternative models were considered. Because the conclusions drawn from a hierarchical model may be sensitive to the choice of the prior for the variance of the random effects[35], in the first model, we replaced the uniform prior for τ with a 1/τ2 prior for τ2 (for an R program: see Appendix S1). In the second model, the assumption of a normal distribution of αi was relaxed by allowing it to follow a t-distribution. A prior U(2,100) was declared for the number of degrees of freedom. A Winbugs program that was used to fit this model can be found in Appendix S3.

Results

Out of 406 articles that were identified in the initial search, 88 articles did not report meta-analyses of an association. We excluded 280 articles because they did not describe a meta-analysis including at least 30 effect sizes. Further, 14 articles did not report any results from a meta-analysis of aggregate data. Finally, four articles were excluded because they did not report the effect sizes from the individual studies and the corresponding author did not respond to a request to provide them. Twenty reports including 49 meta-analyses were used in this study (Figure 1 and Figure 2, references: Appendix S4, raw data available at www.plosone.org).
Figure 1

Publication bias in meta-analyses of clinical trials.

The RR is the ratio of the probability of including statistically significant results favoring the treatment to the probability of including other results. The median of the posterior distribution was used for point estimation. The interval estimate is the 95% equal-tail credible interval. P(RR>1) is the posterior probability that statistically significant results favoring the treatment had a higher chance of being included in the meta-analysis than other results.

Publication bias in meta-analyses of clinical trials.

The RR is the ratio of the probability of including statistically significant results favoring the treatment to the probability of including other results. The median of the posterior distribution was used for point estimation. The interval estimate is the 95% equal-tail credible interval. P(RR>1) is the posterior probability that statistically significant results favoring the treatment had a higher chance of being included in the meta-analysis than other results.

Publication bias in meta-analyses of studies other than clinical trials.

The RR is the ratio of the probability of including plausible statistically significant results to the probability of including other results. The median of the posterior distribution was used for point estimation. The interval estimate is the 95% equal-tail credible interval. P(RR>1) is the posterior probability that plausible statistically significant results had a higher chance of being included in the meta-analysis than other results.

RR in meta-analyses of clinical trials

We estimated the ratio of the probability of including statistically significant results favoring the treatment to the probability of including other results in 28 large meta-analyses of clinical trials that were described in nine articles published in BMJ, JAMA, Lancet, or PLOS Medicine from 2008 to 2012 (Figure 1). In 25 out of the 28 meta-analyses, the estimate of the RR was greater than 1. In 10 meta-analyses, there was strong evidence that statistically significant results favoring the treatment had a higher probability of being included than other outcomes (Figure 1). Trials that demonstrated the efficacy of adding lidocaine on prevention of pain on injection of propofol were estimated to be between 2.45 and 141 times more likely to be included in the meta-analysis than other trials (Figure 3A). Studies favoring pretreatment with lidocaine were estimated to have a between 2.21 and 61.5-fold higher probability of being included in the meta-analysis than other studies (Figure 3B). Changing the prior distribution for the between-study variance and the distribution of the true study effects had little effect on the estimates (Appendix S5).
Figure 3

The posterior distribution of the RR in the meta-analyses of the associations between A) adding lidocaine and the risk of pain on injection of propofol, and B) pretreatment with lidocaine and the risk of pain on injection of propofol.

The posterior distributions describe the knowledge about the RR. The higher the value of the density function, the more likely a given value of RR is in light of the prior knowledge (no prior knowledge was assumed) and the data from the meta-analysis. For both meta-analyses, there was much certainty that the RR was greater than 1, indicating that statistically significant results favoring the treatment had a greater probability of being included in the meta-analysis than other results.

The posterior distribution of the RR in the meta-analyses of the associations between A) adding lidocaine and the risk of pain on injection of propofol, and B) pretreatment with lidocaine and the risk of pain on injection of propofol.

The posterior distributions describe the knowledge about the RR. The higher the value of the density function, the more likely a given value of RR is in light of the prior knowledge (no prior knowledge was assumed) and the data from the meta-analysis. For both meta-analyses, there was much certainty that the RR was greater than 1, indicating that statistically significant results favoring the treatment had a greater probability of being included in the meta-analysis than other results.

RR in other meta-analyses

We identified 10 articles describing 19 meta-analyses of observational studies and one article describing two meta-analyses of interventional studies that were published in BMJ, JAMA, Lancet, or PLOS Medicine between 2008 and 2012. In four meta-analyses, there was strong evidence that plausible statistically significant results had a higher probability of being included than other outcomes (Figure 2). Studies that showed a statistically significant positive association between the C−reactive protein level and cardiovascular events were estimated to have a between 6.97 and 30.7-fold higher probability of being included in the meta-analysis than studies showing other outcomes (Figure 4A). Statistically significant results showing a positive association were estimated to be between 1.92 and 18.1 times more likely to be included in the meta-analysis on the association between child physical abuse and depressive disorders (Figure 4B). The results were robust to the choice of the prior distribution for the between-study variance and the assumption about the distribution of the true study effects (Appendix S6).
Figure 2

Publication bias in meta-analyses of studies other than clinical trials.

The RR is the ratio of the probability of including plausible statistically significant results to the probability of including other results. The median of the posterior distribution was used for point estimation. The interval estimate is the 95% equal-tail credible interval. P(RR>1) is the posterior probability that plausible statistically significant results had a higher chance of being included in the meta-analysis than other results.

Figure 4

The posterior distribution of the RR in the meta-analyses of the associations between A) c−reactive protein level and cardiovascular events and B) child physical abuse and depressive disorders.

The posterior distributions describe the knowledge about the RR. The higher the value of the density function, the more likely a given value of RR is in light of the prior knowledge (no prior knowledge was assumed) and the data from the meta-analysis. For both meta-analyses, there was much certainty that the RR was greater than 1, indicating that plausible statistically significant results had a greater probability of being included in the meta-analysis than other results.

The posterior distribution of the RR in the meta-analyses of the associations between A) c−reactive protein level and cardiovascular events and B) child physical abuse and depressive disorders.

The posterior distributions describe the knowledge about the RR. The higher the value of the density function, the more likely a given value of RR is in light of the prior knowledge (no prior knowledge was assumed) and the data from the meta-analysis. For both meta-analyses, there was much certainty that the RR was greater than 1, indicating that plausible statistically significant results had a greater probability of being included in the meta-analysis than other results.

Discussion

Clinical trials showing statistically significant results favoring the treatment and observational studies showing plausible statistically significant outcomes often had a higher probability of being included in the recent meta-analyses than studies showing other results. The magnitude of the publication bias differed greatly between the meta-analyses and was very large in some cases. For example, for a meta-analysis of the association between the C−reactive protein level and cardiovascular events, statistically significant outcomes showing a positive association were estimated to be between 6.97 and 30.7 times more likely to be included in the analyzed sample than other results. The effect of the higher probability of including for statistically significant outcomes on the combined estimates is unknown due to a lack of information about the exact nature of the bias. However, it is clear that the fundamental assumption of a lack of systematic bias in the process of study selection was strongly violated. Consequently, the validity of a substantial proportion of the recent meta-analyses published in major general medical journals is uncertain due to the presence of a publication bias. Only in 3 [36-38] out of the 14 meta-analyses, in which we found evidence that statistically significant outcomes had a higher probability to be included, the presence of a publication bias was acknowledged in the article. The study demonstrates an application of an attractive alternative to the standard publication bias detection methods for studying a selection process based on the statistical significance in large meta-analyses. Widely used publication bias methods such as the trim and fill method[34,39], Egger’s test[32], rank correlation test[33], and their modifications [18,19,40-42] are based on funnel plot asymmetry. These approaches have two major disadvantages. First, funnel plot asymmetry may be caused by processes other than publication bias, such as between-study variability or small study effects (i.e., larger true effects in smaller studies)[17,32]. As a result, these methods may incorrectly suggest that a publication bias is present[13,22,43,44]. Second, some selection processes introduce little asymmetry to the funnel plot. As a result, widely used publication bias tests often have a low power[18,20,22]. In contrast to these methods, selection models do not rely on the funnel plot but incorporate a model for publication bias in the random effects meta-analysis. While standard approaches investigate the association between effect sizes and some measure of precision to draw conclusions about publication bias, selection models allow to directly estimate parameters that describe the selection process. Several alternatives to the methods based on the funnel plot have been suggested. Iyengar and Greenhouse introduced selection models in meta-analysis[45]. Hedges proposed a class of selection models that incorporated between-study variance[27]. The advantage of these two frequentist methods compared to the class of Bayesian hierarchical selection models described by Silliman [23] is their computational simplicity. We chose the Bayesian approach because it produces valid conclusions when the sample size is small[29], allows a straightforward interval estimation and examination of the sensitivity of the findings to the distribution of the random effects. When a selection process based on the p-values is a point of interest but the weight function is difficult to specify a priori, non-parametric selection models can be used[46,47]. In this study, a parametric model was applied because the aim was to estimate a specific parametric function. Ioannidis and Trikalinos introduced a method based on a comparison of the number of expected and observed statistically significant results in a meta-analysis[48]. A major advantage of this approach is that it does not require a large sample size. An advantage of selection models compared to the method of Ioannidis and Trikalinos is that they allow to take between-study variance into account. Different selection mechanisms have been considered. Dear and Begg developed a selection model based on the assumption that the probability of publishing can be described with a step function with discontinuities at alternate observed p-values[46]. Rufibach described a method that imposed a monotonicity constrained on this function[47]. The trim and fill method handles publication bias defined as the absence of studies with most extreme negative estimates[34,39]. The model of Copas and Shi assumes that the selection probability, given the size of the study, is an increasing function of the observed study effect[49]. Ioannidis and Trikalinos developed a test to investigate an excess of statistically significant findings[48]. We investigated whether statistically significant results favoring the treatment had a higher probability of being included in the meta-analyses of clinical trials. We focused on this selection process because its existence in the medical literature is well-documented by empirical studies following research from inception or submission to a regulatory authority[4,6]. In the case of other meta-analyses, we estimated the ratio of the probability of including biologically plausible statistically significant results to the probability of including other results. As demonstrated by the simulation study, our model performed well in detecting a selection process based on the statistical significance and direction of the effect. However, the power of the model to detect publication bias may be lower when a selection mechanism of a different nature occurs. The main limitation of the study is that we focused on the largest meta-analyses. Possibly, the size of the association between the statistical significance of the results and the probability of including is different for small and medium meta-analyses than for the largest meta-analyses that we considered. When publication bias is detected, an analyst can attempt to account for it[23,28,39,47,49-51]. Although the methods to conduct meta-analysis in the presence of a publication bias provide a powerful sensitivity analysis tool, their validity depends on the correctness of strong and unverifiable assumptions[12]. In light of the studies on publication bias in medicine, including the one presented here, it is clear that the quality of evidence from medical research greatly benefits from policies that aim to reduce underreporting. Several measures that regulate clinical trials have been recently taken. Since 2005, the International Committee of Medical Journal Editors requires a prospective public registration of clinical trials as a condition for publication. Since 2007, the U.S. Food and Drug Administration has also required the registration of trial results. Similar initiatives are needed for observational studies in order to make a clear distinction between a predefined hypothesis testing and exploratory analysis[52]. A prospective registration of all study protocols including a detailed description of the data analysis, a requirement of consistency between the protocol and the study report, and an obligatory disclosure of the results are recommended to further improve the quality of medical literature. R code to fit the main model and the model with a modified prior for the between-study variance. (TXT) Click here for additional data file. Simulation details. (PDF) Click here for additional data file. Winbugs code for the model with t-distributed study-specific means. (PDF) Click here for additional data file. References of meta-analyses included in the study. (PDF) Click here for additional data file. Sensitivity analysis for meta-analyses of clinical trials. (PDF) Click here for additional data file. Sensitivity analysis for meta-analyses of observational and interventional studies. (PDF) Click here for additional data file. Raw data. (ZIP) Click here for additional data file.
  39 in total

1.  Reported methodologic quality and discrepancies between large and small randomized trials in meta-analyses.

Authors:  L L Kjaergard; J Villumsen; C Gluud
Journal:  Ann Intern Med       Date:  2001-12-04       Impact factor: 25.391

2.  A sensitivity analysis for publication bias in systematic reviews.

Authors:  J B Copas; J Q Shi
Journal:  Stat Methods Med Res       Date:  2001-08       Impact factor: 3.021

3.  How vague is vague? A simulation study of the impact of the use of vague prior distributions in MCMC using WinBUGS.

Authors:  Paul C Lambert; Alex J Sutton; Paul R Burton; Keith R Abrams; David R Jones
Journal:  Stat Med       Date:  2005-08-15       Impact factor: 2.373

4.  The performance of tests of publication bias and other sample size effects in systematic reviews of diagnostic test accuracy was assessed.

Authors:  Jonathan J Deeks; Petra Macaskill; Les Irwig
Journal:  J Clin Epidemiol       Date:  2005-09       Impact factor: 6.437

5.  Fate of biomedical research protocols and publication bias in France: retrospective cohort study.

Authors:  Evelyne Decullier; Véronique Lhéritier; François Chapuis
Journal:  BMJ       Date:  2005-06-20

6.  Comparison of two methods to detect publication bias in meta-analysis.

Authors:  Jaime L Peters; Alex J Sutton; David R Jones; Keith R Abrams; Lesley Rushton
Journal:  JAMA       Date:  2006-02-08       Impact factor: 56.272

7.  The existence of publication bias and risk factors for its occurrence.

Authors:  K Dickersin
Journal:  JAMA       Date:  1990-03-09       Impact factor: 56.272

8.  Bias in meta-analysis detected by a simple, graphical test.

Authors:  M Egger; G Davey Smith; M Schneider; C Minder
Journal:  BMJ       Date:  1997-09-13

9.  Operating characteristics of a rank correlation test for publication bias.

Authors:  C B Begg; M Mazumdar
Journal:  Biometrics       Date:  1994-12       Impact factor: 2.571

10.  Underreporting research is scientific misconduct.

Authors:  I Chalmers
Journal:  JAMA       Date:  1990-03-09       Impact factor: 56.272

View more
  24 in total

1.  Consequences of exotic host use: impacts on Lepidoptera and a test of the ecological trap hypothesis.

Authors:  Su'ad Yoon; Quentin Read
Journal:  Oecologia       Date:  2016-01-28       Impact factor: 3.225

2.  Do School-Based Programs Prevent HIV and Other Sexually Transmitted Infections in Adolescents? A Systematic Review and Meta-analysis.

Authors:  Ali Mirzazadeh; M Antonia Biggs; Amanda Viitanen; Hacsi Horvath; Li Yan Wang; Richard Dunville; Lisa C Barrios; James G Kahn; Elliot Marseille
Journal:  Prev Sci       Date:  2018-05

Review 3.  Effectiveness and safety of procalcitonin evaluation for reducing mortality in adults with sepsis, severe sepsis or septic shock.

Authors:  Brenda Ng Andriolo; Regis B Andriolo; Reinaldo Salomão; Álvaro N Atallah
Journal:  Cochrane Database Syst Rev       Date:  2017-01-18

4.  Evidence of citation bias in the pesticide ecotoxicology literature.

Authors:  M L Hanson; L E Deeth; R S Prosser
Journal:  Ecotoxicology       Date:  2018-03-02       Impact factor: 2.823

5.  Empirical Comparison of Publication Bias Tests in Meta-Analysis.

Authors:  Lifeng Lin; Haitao Chu; Mohammad Hassan Murad; Chuan Hong; Zhiyong Qu; Stephen R Cole; Yong Chen
Journal:  J Gen Intern Med       Date:  2018-04-16       Impact factor: 5.128

6.  Vaccines for preventing herpes zoster in older adults.

Authors:  Anna Mz Gagliardi; Brenda Ng Andriolo; Maria Regina Torloni; Bernardo Go Soares; Juliana de Oliveira Gomes; Regis B Andriolo; Eduardo Canteiro Cruz
Journal:  Cochrane Database Syst Rev       Date:  2019-11-07

7.  Successful decannulation of patients with traumatic spinal cord injury: A scoping review.

Authors:  Gordon H Sun; Stephanie W Chen; Mark P MacEachern; Jing Wang
Journal:  J Spinal Cord Med       Date:  2020-11-09       Impact factor: 2.040

Review 8.  Predictive medicine: outcomes, challenges and opportunities in the Synergy-COPD project.

Authors:  Felip Miralles; David Gomez-Cabrero; Magí Lluch-Ariet; Jesper Tegnér; Marta Cascante; Josep Roca
Journal:  J Transl Med       Date:  2014-11-28       Impact factor: 5.531

9.  How does under-reporting of negative and inconclusive results affect the false-positive rate in meta-analysis? A simulation study.

Authors:  Michal Kicinski
Journal:  BMJ Open       Date:  2014-08-28       Impact factor: 2.692

10.  Hyaluronan for knee osteoarthritis: an updated meta-analysis of trials with low risk of bias.

Authors:  Pascal Richette; Xavier Chevalier; Hang Korng Ea; Florent Eymard; Yves Henrotin; Paul Ornetti; Jérémie Sellam; Michel Cucherat; Marc Marty
Journal:  RMD Open       Date:  2015-05-14
View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.