Literature DB >> 28003684

Weighing Evidence "Steampunk" Style via the Meta-Analyser.

Jack Bowden1, Chris Jackson2.   

Abstract

The funnel plot is a graphical visualization of summary data estimates from a meta-analysis, and is a useful tool for detecting departures from the standard modeling assumptions. Although perhaps not widely appreciated, a simple extension of the funnel plot can help to facilitate an intuitive interpretation of the mathematics underlying a meta-analysis at a more fundamental level, by equating it to determining the center of mass of a physical system. We used this analogy to explain the concepts of weighing evidence and of biased evidence to a young audience at the Cambridge Science Festival, without recourse to precise definitions or statistical formulas and with a little help from Sherlock Holmes! Following on from the science fair, we have developed an interactive web-application (named the Meta-Analyser) to bring these ideas to a wider audience. We envisage that our application will be a useful tool for researchers when interpreting their data. First, to facilitate a simple understanding of fixed and random effects modeling approaches; second, to assess the importance of outliers; and third, to show the impact of adjusting for small study bias. This final aim is realized by introducing a novel graphical interpretation of the well-known method of Egger regression.

Entities:  

Keywords:  Bias adjustment; Egger regression; Funnel plot; Meta-analysis

Year:  2016        PMID: 28003684      PMCID: PMC5125286          DOI: 10.1080/00031305.2016.1165735

Source DB:  PubMed          Journal:  Am Stat        ISSN: 0003-1305            Impact factor:   8.710


Introduction

Meta-analysis is the science of combining the evidence available from different sources to arrive at the most sensible and informed conclusion to a given question. The technique has a long history in medical research, where it is routinely used to aggregate results from independent clinical trials testing competing therapies (e.g., which is better: treatment A or B? And by how much?). While each study may be too small or imprecise to definitively answer the clinical question on its own, the overall estimate delivered by a meta-analysis can often do so, thus negating the need to perform further costly research. Its application is not limited to medicine, and spans the entire scientific spectrum. For example, it has been used to synthesize studies measuring: aspects of human psychology such as emotion or motivation (Elfenbein and Ambady 2002); the performance of standard equipment across laboratories (Paule and Mandel 1982); and the rest mass of elementary particles in physics (Baker and Jackson 2013). Scientists tasked with performing a meta-analysis are often hindered by the suspicion that the set of studies they have collected form an incomplete and biased selection of the total evidence base. One reason why this may occur is that a study's outcome (or result) can affect its chances of being published and put into the public domain (Rosenthal 1979). This can, in turn, cause authors to selectively report study results to increase their chances of being published (Ioannidis 2005; Bowden, Jackson, and Thompson 2010). Therefore, bias can be due as much to studies that are missing as to those that are present. In 2014, we attempted to explain the notion of meta-analysis and the concept of biased evidence to a young audience at the Cambridge Science Festival (), an event for Cambridge academics working across the scientific spectrum to promote and explain their current research to the general public. To achieve our specific aim without recourse to mathematical formulas, or any precise definitions of bias, we came up with a mechanical device, named the “Meta-Analyser.” We introduced our invention within the context of a Sherlock Holmes inspired mystery—see Box 1 for an explanation and Figure 1 for pictures of the Meta-Analyser in action at the festival.
Figure 1.

The Meta-Analyser in situ at the Cambridge Science Festival.

The Meta-Analyser in situ at the Cambridge Science Festival. The Meta-Analyser shown in Figures 1 and 2 is a machine that can be used to find the balance or pivot point of a set of weights by hanging the weights from a pole in such a way that the pole is parallel to the ground. It was explained to the children that the weights reflected the importance of each source of information in the calculation.
Figure 2.

Left: Unbalanced Meta-Analyser. Middle: Rebalanced Meta-Analyser (Inspector Lestrade approach). Right: Rebalanced Meta-Analyser (Sherlock Holmes approach).

Left: Unbalanced Meta-Analyser. Middle: Rebalanced Meta-Analyser (Inspector Lestrade approach). Right: Rebalanced Meta-Analyser (Sherlock Holmes approach). Although it does not materially affect the balance point, the length of cord connecting each weight to the pole was chosen to be proportional to its weight. With this convention in place, the Meta-Analyser can be interpreted as a glorified Funnel plot (Sterne and Egger 2001). Funnel plots are used in meta-analysis to visually assess whether the data are potentially affected by bias. If the study results are not symmetrical about the pivot point (or funnel shaped) then missing studies could be responsible. By augmenting the Funnel plot with an additional pole, cord, and pivot, the Meta-Analyser gives this abstract object, and the overall estimate that it implies, a clear physical interpretation. It also facilitates a simple, graphical explanation of the general notion of biased evidence. Box 1 shows the story as explained to the (mainly young) science fair attendees: A small number of study results were removed, leading to an imbalance in the Meta-Analyser. Inspector Lestrade's typically lacklustre approach is to ignore the missing information altogether, and recalculate the pivot point. Sherlock Holmes’ solution is to add in “missing” studies to rebalance and restore symmetry to the Meta-Analyser. Aficionados will notice this as a simplistic representation of the “Trim and Fill” method Duval and Tweedie (2000). The Meta-Analyser was intended to be a simple educational tool to demonstrate the basics of evidence synthesis to a lay audience, capitalizing on the growing recognition of Ada Lovelace in the birth of computation through initiatives such as Ada Lovelace Day () and jumping on the bandwagon of “steampunk” fiction, which melds Victorian fantasy and futurism (Onion 2008; Tanenbaum, Tanenbaum, and Wakkary 2012). However, we continued to find the physical system it describes useful, more broadly, to explain some common secondary issues in meta-analysis. We have therefore developed an online web application (the Meta-Analyser) to implement the approach in practice and to bring this idea to a wider audience of students and researchers. Moreover, our work revealed that the physical analogy the Meta-Analyser promotes is perhaps a deeper and more accurate description of the mathematics underlying a meta-analysis than many statisticians realize. In Section 2, we give a more formal introduction to meta-analysis and introduce the Meta-Analyser web app, showing that it helps to transparently demonstrate the implications of moving from a fixed to a random effects model and to assess the influence of outlying studies. In Section 3, we discuss the issue of biased evidence and how small study bias is commonly addressed by medical statisticians (the authors’ field) using Egger regression (Egger et al. 1997). We then show that by viewing Egger regression from an estimating equation perspective, a novel causal interpretation of this method is found that can be intuitively visualized via the Meta-Analyser. We conclude with a discussion of the issues raised in Section 4 and point to future research. Sections 3.1 and 4.3 formulate previously explained ideas using the estimating equation framework. They may be of most interest to the statistically inclined reader, but can be safely ignored by those wishing to focus on the graphics content of this article alone.

The Meta-Analyser As a Research Tool

A First Statistical Model for Meta-Analysis

Let y, i = 1, …, k, represent summary estimates of the same apparent quantity from k independent information sources in a meta-analysis. The ith estimate is associated with a variance, s 2 that is assumed to be known. The standard fixed effect model for combining the ys assumes and the focus for inference is the population mean effect parameter μ. If s is uncorrelated with ε in Equation (1), so that E[sε] = 0, then each y is an unbiased estimate for μ. The fixed effect estimate and its variance are given by the well-known formulas where the weight given to study i is inversely proportional to its variance. This weighting is attractive, since it minimizes the variance of the overall estimate, . The Meta-Analyser () is an online free-to-access tool to visualize the results from a meta-analysis using the physical analogy previously described. Figure 3 shows the Meta-Analyser populated with a fictional collection of k = 13 studies. This is the “Symmetric” example dataset included in our tool. It shows a standard funnel plot (Sterne and Egger 2001) that has been augmented so that the area representing point i is proportional to w = 1/s 2 , to promote its interpretation as a physical mass. Points are joined by vertical cord to a pole that is joined itself to a vertical stand at a pivot point, p. Study weights are imagined to exert a downward force due to gravity. Since the pole is perfectly horizontal, it is intuitively understood that p satisfies the physical law: and is therefore equal to the center of mass (Beatty 2005). Formula (3) can be viewed as a rudimentary estimating equation, a construction we continue to use throughout this article. It is simple to verify that solving for p yields the fixed effect estimate in (2). The length of the stand base along the x-axis shows the 95% confidence interval for the overall effect μ, which is approximately (−0.75,0.75).
Figure 3.

The Meta-Analyser supporting the symmetric data, a fictional body of 13 study results. The center of mass (overall estimate) is located at zero.

The Meta-Analyser supporting the symmetric data, a fictional body of 13 study results. The center of mass (overall estimate) is located at zero. If the modeling assumption E[sε] = 0 is true, this would imply that study effect sizes and precisions are also uncorrelated. The Meta-Analyser should then appear symmetrical, in that less precise estimates should funnel in from either side toward the most precise estimates. The Meta-Analyser therefore provides a simple means to visually assess the reliability of the overall estimate . Note that this is exactly satisfied for the toy example in Figure 3. When populated with such idealized data, the Meta-Analyser may remind some of Galton's bean board or quincunx, a tool that also illustrates how statistical laws can emerge from a physical process in the presence of gravity. Humans tend to be good at perceiving lengths in graphs, but usually under-estimate areas, and do even worse with volumes (Cleveland and McGill 1984). However, as a reviewer commented, the Meta-Analyser is not relying solely on areas, but rather the entire system in balance and the ability to interact. The Meta-Analyser gives users the flexibility to show the weights in absolute (w = 1/s 2 ) or percentage (w/∑w) terms and to change the scale of the y-axis to improve its aesthetic quality. This is particularly useful for those wishing to include an image of the Meta-Analyser graphical output in a separate document, as done here.

Random Effects Models

In a fixed effect meta-analysis, all studies are assumed to provide an estimate of the same quantity, and therefore any differences between study estimates should be due to chance variation alone. If the fixed effect model is true, then we would expect Cochran's Q-statistic, Q = to be approximately equal to k − 1. If, as is often the case, Q is significantly larger than k − 1, then the fixed effect model is thought implausible. A common secondary issue in meta-analysis, therefore, is accounting for between study heterogeneity via extended modeling approaches. Two distinct methods for incorporating heterogeneity have emerged. The first is via an additional additive random effect, as in model (4) (e.g., DerSimonian and Laird 1986; Higgins and Thompson 2002). The second is via the addition of a multiplicative scale factor, as in model (5) (e.g., Thompson and Sharp 1999; Baker and Jackson 2013): The point estimate for μ obtained by fitting model (5) is always identical to that obtained from fitting model (1), since the common term φ simply cancels from the numerator and denominator in (2). Only the variance of the estimate is altered, due to the weight given to each study being rescaled by a factor of φ (when φ is not equal to one). Consequently, when the study weights are presented in percentage terms, moving from a fixed effect model to a multiplicative random effects model does not induce any physical change in the Meta-Analyser at all. In contrast, both the point estimate and variance for change under the additive random effect model (4) when τ2 is estimated to be greater than zero. Random effects model (4) is popular in medical research; it is promoted by the Cochrane collaboration (Higgins and Green 2011) who are responsible for a large proportion of all meta-analyses in bio-medicine. Moreover, since its implementation via the Meta-Analyser demands a significant graphical extension (as discussed in Section 3.2), we focus on model (4) for the remainder of this section. The assumption E[sδε] = 0 implicit in model (4) is required to ensure that each study still provides an unbiased estimate for μ. It is common practice to estimate τ2 under model (4) using the simple formula defined by DerSimonian and Laird (1986), to give . This estimate conveniently provides a link between Q and a popular measure of heterogeneity, I 2 (Higgins and Thompson 2002) as follows: where s 2 approximates the typical within study variance. When Q is less than k − 1, is simply set to zero. Estimation of μ (and its variance) then follows by simple application of formula (2) except now the weight given to study i, w, changes to .

An Estimating Equation Formulation of Model (4)

To continue to provide a physical interpretation of the calculations underpinning a meta-analysis, and as a natural progression of the single estimating equation for fixed effect model (1), we can formulate random effects model (4) via a system of estimating equations as follows: Formula (8) is referred to as the generalized Q statistic (Bowden et al. 2011) and, when solved in conjunction with (6) and (7), it returns the Paule–Mandel (PM) estimate for τ2 (Paule and Mandel 1982), which we denote by . As with the PM estimate is constrained to be positive, and it is known to provide a more reliable estimate for the between study heterogeneity than (Veroniki et al. 2016).

Random Effects Models via the Meta-Analyser

Application of the random effects model, with additional variance component τ2, leads to study results being both down-weighted and more similarly weighted. Furthermore, the original weight given to large studies is reduced to a greater extent than those of smaller studies. This issue is fairly subtle and hard to comprehend, but the Meta-Analyser provides a very simple visualization, by directly adjusting the mass of each weight. Specifically, we represent the weight “loss” induced by moving from a fixed to an additive random effects model by “drilling out” a square of length x to satisfy the Pythagorean identity = , as illustrated in Figure 4.
Figure 4.

(a) Weight given to study i in a fixed effect meta-analysis. (b) Weight given to study i in an additive random effects meta-analysis. A square of length x is removed from weight i to satisfy the Pythagorean formula.

(a) Weight given to study i in a fixed effect meta-analysis. (b) Weight given to study i in an additive random effects meta-analysis. A square of length x is removed from weight i to satisfy the Pythagorean formula. Figure 5 (left) shows the Meta-Analyser populated with a dataset of eight randomized trial results, each assessing the use of magnesium to treat myocardial infarction. The effect measure, y, is the log-odds ratio of death between treatment and control groups for study i. These data were previously analyzed by Higgins and Spiegelhalter (2002). Between trial heterogeneity is present for these data ( = 0.095, I 2 = 27.6%), so our analysis is under random effects model (4) using to estimate τ2. Full results obtained via the estimating equation approach (and so using the PM estimate for τ2) are shown in Table 1. Under the random effects model, w is reduced from to 1/(s 2 + ). The center of mass or balance point defined by the holed-out weights in Figure 5 (left) is automatically consistent with the random effects estimate for μ. It is immediately apparent that large studies lose a higher proportion of their fixed effect weight than small studies under this model.
Figure 5.

The Meta-Analyser supporting the magnesium data under a random effects model for all trials (left); and with the Shechter trial removed (right).

Table 1.

Meta-analysis of the magnesium data under random effects model (3), with and without the Shechter trial. S.E. = standard error.

Model    
parameterEstimateS.E.t-Valuep-Value
All studies
μ−0.5160.214−2.4080.047
τ2PM0.084
τ2DL0.095 (I2 = 27.6%)
Shechter study removed
μ−0.3620.219−1.6530.149
τ2PM0.008
τ2DL0.012 (I2 = 5.1%)
Figure 6.

Left: The Meta-Analyser supporting the Aspirin data under a random effects model. Right: The Meta-Analyser incorporating Egger regression enacted on the Aspirin data and shown under the potential outcome transform.

The Meta-Analyser supporting the magnesium data under a random effects model for all trials (left); and with the Shechter trial removed (right). Meta-analysis of the magnesium data under random effects model (3), with and without the Shechter trial. S.E. = standard error.

Outliers and Sensitivity Analysis

The amount of heterogeneity estimated in a meta-analysis can depend heavily on extreme, and often small, study results (Bowden et al. 2011). It is therefore useful in some circumstances to perform a sensitivity analysis, in which an outlying study result is excluded. Figure 5 (right) shows the Meta-Analyser supporting the magnesium data with the Shechter study (shown in gray) excluded. This is achieved by simply clicking on its weight. The solid black support stand shows the overall estimate and corresponding 95% confidence interval in this case. For comparison, the original point estimate and confidence interval using all the data is kept but shown in gray. In this example, exclusion of the outlying study removes a large proportion of the between trial heterogeneity (updated = 0.012, I 2 = 5.2%). From looking at the scale of the y-axis, one can see that the weight given to each study has increased, with the largest study benefitting the most. The Meta-Analyser facilitates easy transitions between various models like this as part of a sensitivity analysis. Users also see the Meta-Analyser dynamically tip and rebalance in response to the latest analysis choice.

Small Study Bias

The Aspirin Data

Figure 6 (left) shows the Meta-Analyser enacted on 63 randomized controlled trials reported by Edwards et al. (2000) that each investigated the benefit of oral Aspirin for pain relief. Study estimates y represent the log-odds ratios for the proportion of patients in each arm who had at least a 50% reduction in pain. Between trial heterogeneity was present for these data ( = 0.04, I 2 = 10%) and Figure 6 (left) reflects the weight given to each study by the Meta-Analyser under the random effects model (4) using as the heterogeneity parameter estimate. Despite the apparent between trial heterogeneity, the conclusion of the random effects meta-analysis is that oral Aspirin is an effective treatment, the combined log-odds ratio estimate is 1.26 in favor of Aspirin with a 95% confidence interval (1.1,1.41). Full results are shown in Table 2.
Table 2.

Results for Meta-analyses of the Aspirin data. S.E. = standard error.

Model    
parameterEstimateS.E.t-Valuep-Value
Random effects model (3)
μ1.260.08215.4<10e-16
τ20.04
Random effects model (4)
μ1.230.08015.4<10e-16
φ1.11
Egger regression model (10)
β02.110.316.775.8e-09
μ0.0250.190.130.89
φ0.64
Left: The Meta-Analyser supporting the Aspirin data under a random effects model. Right: The Meta-Analyser incorporating Egger regression enacted on the Aspirin data and shown under the potential outcome transform. Results for Meta-analyses of the Aspirin data. S.E. = standard error. The hypothetical data shown in Figure 1 is perfectly symmetrical about its center of mass, indicating that there is no correlation between effect size and precision across studies. However, there is a clear asymmetry present in the Aspirin data, smaller studies tend to show larger effect size estimates, whereas larger studies tend to report more modest results. For these data, Cor(y, 1/s) = −0.7, which suggests that E[sδε] = 0 does not hold for model (4). The phenomenon of observing a negative correlation between study precision and effect size is often given the umbrella term “small study bias” (Egger et al. 1997; Rücker et al. 2011; Sterne et al. 2011). Small study bias could actually be caused by real differences between small and large studies. Small trials may employ a more intensive intervention and therefore generate a greater effect on disease outcomes than larger trials (Egger et al. 1997; Bowater and Escarela 2013). Asymmetry could also be a simple artefact of the data. For example, point estimates are not strictly independent of their estimated variances when calculated from binary or count outcomes, (Harbord, Egger, and Sterne 2006, Peters et al. 2010). However, its cause could also be more sinister. Publication bias, or the file-drawer problem (Rosenthal 1979) occurs when journals selectively publish study results that achieve a high level of statistical significance, and also induces asymmetry.

Egger Regression

When attempting to explain the concept of biased evidence and of bias adjustment to the science festival audience, we opted for a simplified version of Trim and Fill (Duval and Tweedie 2000). Trim and Fill can be essentially understood as follows. First, smaller studies causing the asymmetry are removed (or “trimmed”) from the data, and the remaining studies are used to estimate the overall effect. Then the trimmed studies are replaced (or “filled”), along with their mirror images, by reflecting them about vertical line defined by the overall effect estimate. Trim and Fill therefore provides an estimate for the number of missing studies and a bias-adjusted estimate for the overall effect. Because of its intuitive nature and visual appeal, it was natural to use this idea within our Sherlock Holmes mystery (Box 1). However, the mathematics behind Trim and Fill are quite complicated and some dislike the way its results depend on imputed data. Perhaps due to its relative simplicity, the most popular approach to testing and adjusting for small study bias in medical research is Egger regression (Egger et al. 1997). This assumes the following linear fixed effect model to explain the correlation between y and 1/s: The addition of an intercept parameter β0 in model (9) allows small study bias to be accounted for. Testing for small study bias is then equivalent to testing H 0: β0 = 0. If E[sε] = 0 for model (9), then the overall effect estimate, , adjusted for possible small study bias (via ) is a consistent estimate for μ. Several authors have considered the addition of random effects into model (9), to account for possible residual heterogeneity after adjustment for small study bias, see, for example, Moreno et al. (2009), Peters et al. (2010), and Rücker et al. (2011). Their approaches have been straightforward generalizations of the additive and multiplicative random effects models (4) and (5), respectively, as below: At first sight, model (10) seems the most natural extension to model (9). However, when a nonzero value for τ2 is estimated under model (10), the resulting overall estimate for μ differs from, and can often exhibit more substantial bias than the fixed effect estimate (Rücker et al. 2011). By contrast, multiplicative model (11) is far more well behaved in this respect since its point estimate is identical to that obtained from fitting model (9). Nullifying the influence of variance components on the overall mean, a property enjoyed by model (11) is so attractive in the presence of small study bias, that approaches have been developed to artificially incorporate this feature into additive random effects models as well (Henmi and Copas 2010). For these reasons, we analyze the Aspirin data using the multiplicative Egger regression model only. This is straightforward, because model (11) is automatically fitted by the process of standard linear regression, in which the variance of the residual error is estimated, rather than assumed to be 1. For these data, = 2.11, with a p-value of approximately 1 × 10− 8, is equal to 0.025, with a p-value of 0.89, and = 0.64. In summary, Egger regression detects a highly significant presence of small study bias and, after this has been removed, no evidence of a treatment effect whatsoever.

An Estimating Equation Formulation of Model (11)

We now show that, like Trim and Fill, Egger regression can be understood as a means to de-bias a meta-analysis by restoring symmetry to the funnel plot, in a way that complements its application via the Meta-Analyser and makes connections with the world of causal inference. We start by assuming model (11). Multiplying each side by s and subtracting β0 s yields The term y(β0) is a transformed version of the effect size estimates, and when E[sε] = 0, y(β0) is mean-independent of s. The intercept estimate obtained from fitting model (11), , can be viewed as defining a particular transform of the data that forces to be independent of s across all studies. We note the similarity between this formulation of Egger regression and the structural mean model framework in causal inference, in which observed outcomes are related to potential outcomes (Rubin 2005) using a parametric transform to satisfy statistical independence rules. Parameter estimation via this strategy is termed G-estimation, and has been used extensively to adjust for biases due to noncompliance or dropout in clinical trials and confounding bias in observational studies. For a general overview, see Vansteelandt and Joffe (2014) or, in the context of epidemiology, Bowden and Vansteelandt (2011). For this reason, we refer to y(β0) as the potential outcome transform. Employing the estimating equation framework, model (11) is equivalent to solving the following system: where is the arithmetic mean of the s terms. We now clarify the connection between the above system of estimating equations and estimation of β0, μ, and φ using standard linear regression theory. Fitting model (9) to obtain estimates for β0 and μ is equivalent to solving Equations (12) and (13) leaving φ unspecified (in the Appendix we provide some simple R code to verify this). We then formally define φ as a parameter and solve equation (14) to give We note the equivalence of the numerator of Equation (15) to the Q statistic defined in Rücker et al. (2011). The variances of and are given by where is the sample mean of the 1/s 2 terms. Concerns over the use of Egger regression have been raised when analyzing binary data because a study's outcome estimate will not truly be independent of its standard error, even when small study bias is not present. For this reason, Peters et al. (2010) proposed to replace s in the Egger regression equation with a measure of precision based on study size, 1/n say. This could easily be represented in our estimating equation with suitable modification. For example, G-estimation Equation (13) above would then be used to find the β0 that forces independence between y(β0) = y − β0/n and 1/n.

Reanalysis of the Aspirin Data via the Meta-Analyser

Returning to the Aspirin data, Figure 6 (right) shows the final resting point of the Meta-Analyser upon enacting Egger regression using the estimating equation interpretation described above. Once the Egger regression option is ticked, users observe a dynamical change in the Meta-Analyser from its initial starting point of the standard random effects analysis in Figure 6 (left). The x-axis position is now the transformed or potential outcome scale = y(2.11) for study i. The meta-analysis now exhibits a high amount of symmetry that can be immediately visualized by the user. This transformation is highlighted in Figure 7, which plots potential outcomes on the horizontal versus on the vertical axis with the original data (y vs. 1/s) also shown for comparison. When small study bias is present, estimates from small studies are shifted by a relatively large amount in the horizontal dimension (in this case to the left), whereas those from large studies are shifted horizontally by a relatively small amount.
Figure 7.

Funnel plot of the original Aspirin data (y vs. 1/s, hollow red dots) versus its transformed counterpart ( vs. , solid black dots).

Funnel plot of the original Aspirin data (y vs. 1/s, hollow red dots) versus its transformed counterpart ( vs. , solid black dots). Because φ is estimated to be 0.64 in this instance, indicating under-dispersion after adjustment for small study bias, study precisions under the potential outcome transformation are increased in proportion to their size, so that the precisions of small studies are shifted vertically upward by a small amount and those of large studies are shifted by a large amount. If over-dispersion had been observed after adjustment, we would have seen a shift vertically downward instead.

Discussion

From its original conception and success as a science festival exhibit aimed at a lay audience, we believe that the Meta-Analyser will prove a useful online tool for explaining the rationale, and interpreting the effect of, extended modeling choices in meta-analysis to a more advanced audience. It has also proved a useful vehicle for forging new connections between methods of bias adjustment from the statistical literature. To keep the Meta-Analyser as faithful as possible to the original funnel plot, the length of the cord connecting each weight to the horizontal pole was also set to be proportional to the weight. Clearly, only the weight matters for the physical interpretation. This redundancy could be exploited to come up with different versions of the Meta-Analyser. For example, the cord lengths could all be uniform, but this would lose the ability to detect asymmetry. One could also modify the weights. At present they are holed out when between study heterogeneity is present, to show how the weight of each study decreases. Other ways of achieving the same end are surely possible. The program code for the Meta-Analyser is available on the GitHub repository . An R package containing the Meta-Analyser is also available, which allows users to run the application off-line. This can be installed from the repository as follows: install.packages(''devtools'') # if necessary devtools::install_github(''rstudio/DT'') devtools::install_github(''chjackson/MetaAnalyser'') We hope to continue to develop and improve the Meta-Analyser to incorporate new features and analysis choices. New ideas and offers of collaboration are very welcome.
  19 in total

1.  Trim and fill: A simple funnel-plot-based method of testing and adjusting for publication bias in meta-analysis.

Authors:  S Duval; R Tweedie
Journal:  Biometrics       Date:  2000-06       Impact factor: 2.571

2.  On the universality and cultural specificity of emotion recognition: a meta-analysis.

Authors:  Hillary Anger Elfenbein; Nalini Ambady
Journal:  Psychol Bull       Date:  2002-03       Impact factor: 17.737

3.  Quantifying heterogeneity in a meta-analysis.

Authors:  Julian P T Higgins; Simon G Thompson
Journal:  Stat Med       Date:  2002-06-15       Impact factor: 2.373

4.  Treatment-effect estimates adjusted for small-study effects via a limit meta-analysis.

Authors:  Gerta Rücker; Guido Schwarzer; James R Carpenter; Harald Binder; Martin Schumacher
Journal:  Biostatistics       Date:  2010-07-22       Impact factor: 5.899

5.  Confidence intervals for random effects meta-analysis and robustness to publication bias.

Authors:  Masayuki Henmi; John B Copas
Journal:  Stat Med       Date:  2010-10-20       Impact factor: 2.373

6.  A modified test for small-study effects in meta-analyses of controlled trials with binary endpoints.

Authors:  Roger M Harbord; Matthias Egger; Jonathan A C Sterne
Journal:  Stat Med       Date:  2006-10-30       Impact factor: 2.373

7.  Modelling multiple sources of dissemination bias in meta-analysis.

Authors:  Jack Bowden; Dan Jackson; Simon G Thompson
Journal:  Stat Med       Date:  2010-03-30       Impact factor: 2.373

8.  Meta-analysis in clinical trials.

Authors:  R DerSimonian; N Laird
Journal:  Control Clin Trials       Date:  1986-09

Review 9.  Single dose oral aspirin for acute pain.

Authors:  J E Edwards; A Oldman; L Smith; S L Collins; D Carroll; P J Wiffen; H J McQuay; R A Moore
Journal:  Cochrane Database Syst Rev       Date:  2000

10.  Recommendations for examining and interpreting funnel plot asymmetry in meta-analyses of randomised controlled trials.

Authors:  Jonathan A C Sterne; Alex J Sutton; John P A Ioannidis; Norma Terrin; David R Jones; Joseph Lau; James Carpenter; Gerta Rücker; Roger M Harbord; Christopher H Schmid; Jennifer Tetzlaff; Jonathan J Deeks; Jaime Peters; Petra Macaskill; Guido Schwarzer; Sue Duval; Douglas G Altman; David Moher; Julian P T Higgins
Journal:  BMJ       Date:  2011-07-22
View more
  5 in total

1.  A new justification of the Hartung-Knapp method for random-effects meta-analysis based on weighted least squares regression.

Authors:  Robbie C M van Aert; Dan Jackson
Journal:  Res Synth Methods       Date:  2019-08-14       Impact factor: 5.273

2.  Charting the landscape of graphical displays for meta-analysis and systematic reviews: a comprehensive review, taxonomy, and feature analysis.

Authors:  Michael Kossmeier; Ulrich S Tran; Martin Voracek
Journal:  BMC Med Res Methodol       Date:  2020-02-07       Impact factor: 4.615

3.  Verification of autoclaving-cooling treatment to increase the resistant starch contents in food starches based on meta-analysis result.

Authors:  Didah Nur Faridah; Rhoito Frista Silitonga; Dias Indrasti; Frendy Ahmad Afandi; Anuraga Jayanegara; Maria Putri Anugerah
Journal:  Front Nutr       Date:  2022-07-19

4.  Improving the visualization, interpretation and analysis of two-sample summary data Mendelian randomization via the Radial plot and Radial regression.

Authors:  Jack Bowden; Wesley Spiller; Fabiola Del Greco M; Nuala Sheehan; John Thompson; Cosetta Minelli; George Davey Smith
Journal:  Int J Epidemiol       Date:  2018-08-01       Impact factor: 7.196

5.  The median and the mode as robust meta-analysis estimators in the presence of small-study effects and outliers.

Authors:  Fernando P Hartwig; George Davey Smith; Amand F Schmidt; Jonathan A C Sterne; Julian P T Higgins; Jack Bowden
Journal:  Res Synth Methods       Date:  2020-03-10       Impact factor: 5.273

  5 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.