Literature DB >> 23804508

Evidence synthesis for decision making 4: inconsistency in networks of evidence based on randomized controlled trials.

Sofia Dias1, Nicky J Welton1, Alex J Sutton2, Deborah M Caldwell1, Guobing Lu1, A E Ades1.   

Abstract

Inconsistency can be thought of as a conflict between "direct" evidence on a comparison between treatments B and C and "indirect" evidence gained from AC and AB trials. Like heterogeneity, inconsistency is caused by effect modifiers and specifically by an imbalance in the distribution of effect modifiers in the direct and indirect evidence. Defining inconsistency as a property of loops of evidence, the relation between inconsistency and heterogeneity and the difficulties created by multiarm trials are described. We set out an approach to assessing consistency in 3-treatment triangular networks and in larger circuit structures, its extension to certain special structures in which independent tests for inconsistencies can be created, and describe methods suitable for more complex networks. Sample WinBUGS code is given in an appendix. Steps that can be taken to minimize the risk of drawing incorrect conclusions from indirect comparisons and network meta-analysis are the same steps that will minimize heterogeneity in pairwise meta-analysis. Empirical indicators that can provide reassurance and the question of how to respond to inconsistency are also discussed.

Entities:  

Keywords:  Bayesian; Network meta-analysis; inconsistency; indirect evidence

Mesh:

Year:  2013        PMID: 23804508      PMCID: PMC3704208          DOI: 10.1177/0272989X12455847

Source DB:  PubMed          Journal:  Med Decis Making        ISSN: 0272-989X            Impact factor:   2.583


Introduction

Network meta-analysis (NMA), also referred to as mixed treatment comparisons or multiple treatment meta-analysis, combines information from multiple randomized comparisons of treatments A versus B, A versus C, B versus C, A versus D, and so on,[1-6] while preserving randomization.[7] Given a connected network of comparisons, NMA produces an internally coherent set of estimates of the efficacy of any treatment in the network relative to any other, under the key assumption of evidence consistency. This requires that in every trial i in the network, regardless of the actual treatments that were compared, the true effect δ of treatment Y relative to treatment X is the same in a fixed effects (FE) model, i.e., δ = d XY, or exchangeable between trials in a random effects (RE) model, i.e., δ ~ Normal(d XY, σ2). From this assumption, the consistency equations can be deduced,[6,8-10] asserting that for any 3 treatments X, Y, Z, the FE, or mean effects in an RE model, are related as follows: d YZ = d XZ − d XY. Where doubts have been expressed about NMA, these have focused on the consistency equations.[11,12] This is because, unlike the exchangeability assumptions from which they are derived, which are notoriously difficult to verify, the consistency equations offer a prediction about relationships in the data that can be statistically tested. Note that consistency concerns the relation between the treatment contrasts whereas heterogeneity concerns the variation between trials within each contrast (we use contrast to refer to a pairwise comparison between 2 treatments). This tutorial suggests methods for detection of inconsistency in evidence networks, clarifies the measures that can be taken to minimize the risk of drawing incorrect conclusions from indirect comparisons and NMA, and suggests some empirical indicators that might help assess what that risk might be. Sample code using the WinBUGS 1.4.3 package[13] is set out in the appendix. This tutorial should be seen as an adjunct to Dias and others,[10] which sets out a generalized linear modeling framework for NMA, indirect comparisons, and pairwise meta-analysis and explains how the same core model can be applied with different likelihoods and linking functions. It should be understood that this carries over entirely to the Bayesian models for inconsistency.

Network Structure

Evidence Loops

The first step in checking for inconsistency is to examine network diagrams carefully, as the structure can reveal particular features that may assist in the choice of analysis method. We begin by considering networks that consist only of 2-arm trials, starting with a triangular network ABC (Figure 1a), in which each edge represents direct evidence comparing the treatments it connects. Taking treatment A as our reference treatment, a consistency model[8,10] has 2 basic parameters, say d AB and d AC, but we have data on 3 contrasts, d AB, d AC, and d BC. The latter, however, is not an independent parameter but is wholly determined by the 2 other parameters through the consistency equations. Setting aside the question of the number of trials informing each pairwise contrast, we can see that there are 2 independent parameters to estimate and 3 sources of data. This generates 1 degree of freedom with which to detect inconsistency. Thus, if all trials are 2-arm trials, the inconsistency degrees of freedom (ICDF) can be calculated from the number of treatments (nt) and the number of contrasts (N) on which there is evidence as[6]
Figure 1

Possible treatment networks: treatments are represented by letters; lines connecting 2 treatments indicate that a comparison between these treatments has been made (in 1 or more randomized controlled trials).

Possible treatment networks: treatments are represented by letters; lines connecting 2 treatments indicate that a comparison between these treatments has been made (in 1 or more randomized controlled trials). This accords with the commonsense notion of inconsistency, which views it as a property of loops of evidence.[14,15] Every additional independent loop in a network of 2-arm trials represents 1 additional ICDF and one further way in which potential inconsistency can be realized. In the square network in Figure 1b, there are N = 4 independent pieces of evidence, nt = 4 treatments, and nt − 1 = 3 parameters in a consistency model, giving ICDF = 4 − (4 − 1) = 1. In Figure 1c, there are N = 9 contrasts on which there is evidence, nt = 7 treatments, and 6 parameters, giving ICDF = 3. Note that the ICDF is equal to the number of independent loops. In Figure 1c, there are 2 separate structures in which inconsistency could be detected: the triangle EFG and the square ABCD. In the square, one could count a total of 3 loops: ABC, BCD, and ABCD. However, there are only 2 independent loops in this part of the structure: If we know all the edges of any 2 loops, we immediately know the edges of the third. Therefore, there can be only 2 inconsistencies in the ABCD square. Similarly, in Figure 1d, one can count a total of 7 loops: 4 three-treatment loops (ACD, BCD, ABD, ABC) and 3 four-treatment loops (ABCD, ACDB, CABD). But there are only 3 independent loops: N = 6, nt = 4, and ICDF = 3. It is not possible to specify which loops are independent, only how many there are because knowing the edges of any 3 loops will mean we know the edges of the others.

Multiarm Trials

When multiarm trials (i.e., trials with more than 2 arms) are included in the network, the definition of inconsistency becomes more complex. A 3-arm trial provides evidence on all 3 edges of an ABC triangle, and yet it cannot be inconsistent. In other words, although trial i estimates 3 parameters, δ, δ, δ, only 2 are independent because δ = δ − δ. There can therefore be no inconsistency within a 3-arm trial. Similarly, if all the evidence was from 3-arm trials on the same 3 treatments, there could be no inconsistency, only between-trials heterogeneity. The difficulty in defining inconsistency comes when we have 2- and 3-arm trial evidence, for example, AB, AC, BC, and ABC trials. Because ICDF corresponds to the number of independent loops, if a loop is formed from a multiarm trial alone, it is not counted as an independent loop and must therefore be discounted from the total ICDF.[6] Thus, where there are mixtures of 2-arm and multiarm trials, our definition of inconsistency as arising in loops creates inherent technical difficulties that cannot, as far as is known, be avoided.

Testing for Inconsistency

A key consideration in consistency assessment is whether independent tests for inconsistency can be constructed. These should be used wherever possible as they provide the simplest, most complete, and easiest to interpret analyses of inconsistency. We show how to construct independent tests, explain the circumstances where this is not possible, and set out methods for the more general case, which can be applied to any network.

Bucher Method for Single Loops of Evidence

The simplest method for testing consistency of evidence is essentially a 2-stage method.[16] The first stage is to separately synthesize the evidence in each pairwise contrast; the second stage is to test whether direct and indirect evidence are in conflict. A direct estimate of the C versus B effect, , is compared with an indirect estimate, , formed from the AB and AC direct evidence The direct estimates can be either from individual trials or from pairwise meta-analyses, whether fixed or random effects. Attached to each direct estimate is a variance, for example, (. As the direct estimates are statistically independent, we have . Estimates of the inconsistency, ω, and its variance can be formed by subtracting the direct and indirect estimates: An approximate test of the null hypothesis that there is no inconsistency is obtained by referring to the standard normal distribution. It makes no difference whether we compare the direct BC evidence to the indirect evidence formed through AB and AC, or compare the direct AB evidence to the indirect AC and BC, or compare the AC with the AB and BC. The absolute values of the inconsistency estimates will be identical, as will the variances. This agrees with the intuition that, in a single loop, there can be only 1 inconsistency. However, this method can only be applied to 3 independent sources of data. Three-arm trials cannot be included because they are internally consistent and will reduce the chances of detecting inconsistency. This method generalizes naturally to the square network in Figure 1b, which, like the triangle and any other simple circuit structure, has ICDF = 1. An indirect estimate of any edge can be formed from the remaining edges, and the variance of the inconsistency term is the sum of the variances of all the comparisons. For example, an indirect estimate can be formed as , by successive application of the consistency equations. Clearly, as the number of edges in the loop increases, it becomes less and less likely that a real inconsistency will be detected because of the higher variance of the inconsistency estimate.

Extension to networks with multiple loops

Figure 1c represents a further pattern in which the inconsistency analysis can be broken down into separate independent elements: There are a total of 3 independent loops, and ICDF = 9 − (7 − 1) = 3. In this case, one inconsistency relates to the loop EFG, where there are 2 sources of evidence on any edge, whereas the other concerns the edge BC, on which there are 3 independent sources of evidence, 1 direct and 2 indirect. To analyze inconsistency in this structure, the problem is broken down into 2 separate and unrelated components. First, inconsistency in the EFG triangle is examined using the simple Bucher approach. Second, consistency between the 3 sources of evidence on the BC edge is examined by calculating a statistic to refer to a distribution.[17] These 2 independent tests provide a complete analysis of the inconsistency in this network. These methods are based on 2-arm trials. Inclusion of multiarm trials will lower their power to detect inconsistency. Our suggestion is that when a test on a loop ABC is being constructed, evidence from 3-arm ABC trials is excluded. However, ABC evidence on AB should be included when testing, for example, the ABD loop.

Methods for General Networks

Figure 1d shows a 4-treatment network in which there are data on every contrast and 3 possible inconsistencies. The difference between the networks in Figure 1d and Figure 1c is that in the former, there are 4 three-treatment loops (ACD, BCD, ABD, ABC) and 3 four-treatment loops (ABCD, ACDB, CABD), but these loops are not statistically independent. It is therefore not possible to construct a set of independent tests to examine the 3 inconsistencies. Applying the Bucher method to each of the 7 loops in the network in turn would be a simple way to check inconsistency in this network. However, the number of loops, and hence the number of tests carried out, will far exceed the maximum number of possible inconsistencies in the network. For example, in a network where N = 42, nt = 12, and ICDF = 31,[18] repeated use of the Bucher method on each of the 3-way loops in this network gave 70 estimates of inconsistency for the response outcome and 63 estimates for the acceptability outcome. In total, 6 loops showed statistically significant inconsistency, and the authors concluded that this was compatible with chance as 133 separate tests were performed. However, this could be questioned on the grounds that the 133 tests were not independent; there could not be more than 62 independent tests, and even this assumes that the 2 outcomes are unrelated. Difficulties in the interpretation of statistical tests arise if any of the loops show significant inconsistency, at say a P < 0.05 level. One cannot immediately reject the null hypothesis at this level because multiple testing has taken place, and adjustment of significance levels would need to be considered. However, because the tests are not independent, calculating the correct level of adjustment becomes a complex task. Furthermore, in networks with multiple treatments, the total number of triangular, quadrilateral, and higher-order loops may be extremely large.

Unrelated mean effects model

In complex networks, where independent tests cannot be constructed, we propose that the standard consistency model[8,10] is compared with a model not assuming consistency. In the consistency model, a network with nt treatments, A, B, C, … defines nt − 1 basic parameters[19] d, …, which estimate the effects of all treatments relative to treatment A, chosen as the reference treatment. Prior distributions are placed on these parameters. All other contrasts can be defined as functions of the basic parameters by making the consistency assumption. We propose an unrelated mean effects (UME) model in which each of the N contrasts for which evidence is available represents a separate, unrelated, basic parameter to be estimated: no consistency is assumed. This model has also been termed inconsistency model.[9,20,21] Formally, suppose we have a set of M trials comparing nt = 4 treatments, A, B, C, and D, in any connected network. In an RE model, the study-specific treatment effects for a study comparing a treatment X to another treatment Y, δ, are assumed to follow a normal distribution In a consistency model, nt − 1 = 3 basic parameters are given vague priors: d ~N(0,1002), and the consistency equations define all other possible contrasts as In an RE UME model, each of the mean treatment effects in equation 2 is treated as a separate (independent) parameter to be estimated, sharing a common variance σ2. So, for the network in Figure 1d, the 6 treatment effects are all given vague priors: d ~ N(0,1002). Note that the extra number of parameters in this model is equal to the ICDF. In an FE UME model, no shared variance parameter needs to be considered. The model is then equivalent to performing completely separate pairwise meta-analyses of the contrasts. However, fitting a UME model to all the data has the advantage of easily accommodating multiarm trials as well as providing a single global measure of model fit. When multiarm trials are included in the evidence, the UME model can have different parameterizations depending on which of the multiple contrasts defined by a multiarm trial are chosen. For example, a 3-arm trial ABC can inform the AB and AC independent effects, or it can be chosen to inform the AB and BC effects (if B was the reference treatment), or the AC and BC effects (with C as reference). The code presented in the appendix arbitrarily chooses the contrasts relative to the first treatment in the trial. Thus, ABC trials inform the AB and AC contrasts, BCD trials inform BC and BD, and so forth. For FE models, the choice of parameterization makes no difference to the results, but in the RE model, the choice of parameterization will affect both the heterogeneity estimate and the tests of inconsistency.

Illustrative Examples

Smoking cessation

Twenty-four studies, including 2 three-arm trials, compared 4 smoking cessation counseling programs and recorded the number of individuals with successful smoking cessation at 6 to 12 mo.[3,6] All possible contrasts were compared, forming the network in Figure 1d, where A = no intervention, B = self-help, C = individual counseling, and D = group counseling. We contrast a consistency model[8,10] with an RE model estimating 6 independent mean treatment effects. Results for both models are presented in Table 1, along with the posterior mean of the residual deviance and deviance information criterion (DIC), measures to assess model fit.[22] Comparison between the deviance and DIC statistics of the consistency and UME models provides an omnibus test of consistency. In this case, the heterogeneity estimates, the posterior means of the residual deviance, and the DICs are very similar for both models.
Table 1

Smoking Example: Posterior Summaries from Random Effects Consistency and Unrelated Mean Effects Models

Network Meta-analysis[a] (Consistency Model)
Unrelated Mean Effects Model
Mean/MedianSDCrIMean/MedianSDCrI
dAB 0.490.40(−0.29, 1.31)0.340.58(−0.81, 1.50)
dAC 0.840.24(0.39, 1.34)0.860.27(0.34, 1.43)
dAD 1.100.44(0.26, 2.00)1.430.88(−0.21, 3.29)
dBC 0.350.41(−0.46, 1.18)−0.050.74(−1.53, 1.42)
dBD 0.610.49(−0.34, 1.59)0.650.73(−0.80, 2.12)
dCD 0.260.41(−0.55, 1.09)0.200.78(−1.37, 1.73)
σ0.820.19(0.55, 1.27)0.890.22(0.58, 1.45)
resdev[b] 54.053.4
pD45.046.1
DIC99.099.5

Note: Mean, standard deviation (SD), 95% credible interval (CrI) of relative treatment effects, and median of between-trial standard deviation (σ) on the log-odds scale and posterior mean of the residual deviance (resdev), effective number of parameters (pD), and deviance information criterion. Results are based on 100 000 iterations on 3 chains after a burn-in period of 20 000 for the consistency model and after a burn-in of 30 000 for the inconsistency model. Treatments: A= no intervention, B = self-help, C = individual counseling, D = group counseling.

d calculated using the consistency equations.

Compare to 50 data points.

Smoking Example: Posterior Summaries from Random Effects Consistency and Unrelated Mean Effects Models Note: Mean, standard deviation (SD), 95% credible interval (CrI) of relative treatment effects, and median of between-trial standard deviation (σ) on the log-odds scale and posterior mean of the residual deviance (resdev), effective number of parameters (pD), and deviance information criterion. Results are based on 100 000 iterations on 3 chains after a burn-in period of 20 000 for the consistency model and after a burn-in of 30 000 for the inconsistency model. Treatments: A= no intervention, B = self-help, C = individual counseling, D = group counseling. d calculated using the consistency equations. Compare to 50 data points. Plotting the posterior mean deviance of the individual data points in the UME model against their posterior mean deviance in the consistency model (Figure 2) provides information that can help identify the loops in which inconsistency is present. We expect each data point to have a posterior mean deviance contribution of about 1, with higher contributions suggesting a poorly fitting model.[22] In this example, the contributions to the deviance are very similar and close to 1 for both models. Two points have a higher than expected posterior mean deviance—these are the arms of 2 trials that have a zero cell—but the higher deviance is seen in both models. In general, trial arms with zero cells will have a high deviance as the model will never predict a zero cell exactly. The parameter estimates are similar for both models, and there is considerable overlap in the 95% credible intervals. This suggests no evidence of inconsistency in the network.
Figure 2

Plot of the individual data points’ posterior mean deviance contributions for the consistency model (horizontal axis) and the unrelated mean effects model (vertical axis) along with the line of equality.

Plot of the individual data points’ posterior mean deviance contributions for the consistency model (horizontal axis) and the unrelated mean effects model (vertical axis) along with the line of equality.

Thrombolytic treatments

Figure 3 represents the treatment network for a data set consisting of 50 trials comparing 8 thrombolytic drugs and percutaneous transluminal coronary angioplasty, following acute myocardial infarction.[23,24] Data consist of the number of deaths in 30 or 35 days and the number of patients in each treatment arm. Note that in this network, not all treatment contrasts have been compared in a trial. There are 9 treatments in total and information on 16 pairwise comparisons, which would suggest an ICDF of 8. However, there is 1 loop, SK, Acc t-PA, SK+t-PA (highlighted in bold), which is only informed by a 3-arm trial and therefore cannot contribute to the number of possible inconsistencies. Discounting this loop gives ICDF = 7.[25]
Figure 3

Thrombolytics example network. Lines connecting 2 treatments indicate that a comparison between these treatments (in 1 or more randomized controlled trials) has been made. The triangle highlighted in bold represents comparisons that have been made in only a 3-arm trial. Treatments: streptokinase (SK), alteplase (t-PA), accelerated alteplase (Acc t-PA), reteplase (r-PA), tenecteplase (TNK), urokinase (UK), anistreptilase (ASPAC), percutaneous transluminal coronary angioplasty (PTCA).

Thrombolytics example network. Lines connecting 2 treatments indicate that a comparison between these treatments (in 1 or more randomized controlled trials) has been made. The triangle highlighted in bold represents comparisons that have been made in only a 3-arm trial. Treatments: streptokinase (SK), alteplase (t-PA), accelerated alteplase (Acc t-PA), reteplase (r-PA), tenecteplase (TNK), urokinase (UK), anistreptilase (ASPAC), percutaneous transluminal coronary angioplasty (PTCA). An FE NMA (consistency model) with a binomial likelihood and logit link[8,10] was fitted to the data, taking SK as the reference treatment, that is, the 8 treatment effects relative to SK are the basic parameters and have been estimated, whereas the remaining relative effects were obtained from the consistency assumptions. An FE model without the consistency assumptions was also fitted, which estimated 15 independent mean treatment effects (Table 2).
Table 2

Thrombolitics Example: Posterior Summaries, Mean, Standard Deviation (SD), and 95% Credible Interval (CrI) on the Log-Odds Ratio Scale for Treatments Y versus X for Contrasts That Are Informed by Direct Evidence and Posterior Mean of the Residual Deviance (resdev), Number of Parameters (pD), and DIC for the Fixed Effects Network Meta-analysis and Inconsistency Models

TreatmentNetwork Meta-analysis[a] (Consistency Model)Unrelated Mean Effects Model
XYMeanSDCrIMeanSDCrI
SKt-PA0.0020.030(−0.06, 0.06)−0.0040.030(−0.06, 0.06)
SKAcc t-PA−0.1770.043(−0.26, −0.09)−0.1580.049(−0.25, −0.06)
SKSK + t-PA−0.0490.046(−0.14, 0.04)−0.0440.047(−0.14, 0.05)
SKr-PA−0.1240.060(−0.24, −0.01)−0.0600.089(−0.23, 0.11)
SKPTCA−0.1730.077(−0.32, −0.02)−0.6650.185(−1.03, −0.31)
SKUK−0.4760.101(−0.67, −0.28)−0.3690.518(−1.41, 0.63)
SKASPAC−0.2030.221(−0.64, 0.23)0.0050.037(−0.07, 0.08)
t-PAPTCA0.0160.037(−0.06, 0.09)−0.5440.417(−1.38, 0.25)
t-PAUK−0.1800.052(−0.28, −0.08)−0.2940.347(−0.99, 0.37)
t-PAASPAC−0.0520.055(−0.16, 0.06)−0.2900.361(−1.01, 0.41)
Acc t-PAr-PA−0.1260.067(−0.26, 0.01)0.0190.066(−0.11, 0.15)
Acc t-PATNK−0.1750.082(−0.34, −0.01)0.0060.064(−0.12, 0.13)
Acc t-PAPTCA−0.4780.104(−0.68, −0.27)−0.2160.119(−0.45, 0.02)
Acc t-PAUK−0.2060.221(−0.64, 0.23)0.1460.358(−0.54, 0.86)
Acc t-PAASPAC0.0130.037(−0.06, 0.09)1.4050.417(0.63, 2.27)
resdev[b] 105.999.7
pD5865
DIC163.9164.7

Note: Results are based on 50 000 iterations on 2 chains after a burn-in period of 50 000 for the consistency model and after a burn-in of 20 000 for the inconsistency model. Treatments: streptokinase (SK), alteplase (t-PA), accelerated alteplase (Acc t-PA), reteplase (r-PA), tenecteplase (TNK), urokinase (UK), anistreptilase (ASPAC), percutaneous transluminal coronary angioplasty (PTCA).

All relative treatment effects not involving SK were calculated using the consistency equations.

Compare to 102 data points.

Thrombolitics Example: Posterior Summaries, Mean, Standard Deviation (SD), and 95% Credible Interval (CrI) on the Log-Odds Ratio Scale for Treatments Y versus X for Contrasts That Are Informed by Direct Evidence and Posterior Mean of the Residual Deviance (resdev), Number of Parameters (pD), and DIC for the Fixed Effects Network Meta-analysis and Inconsistency Models Note: Results are based on 50 000 iterations on 2 chains after a burn-in period of 50 000 for the consistency model and after a burn-in of 20 000 for the inconsistency model. Treatments: streptokinase (SK), alteplase (t-PA), accelerated alteplase (Acc t-PA), reteplase (r-PA), tenecteplase (TNK), urokinase (UK), anistreptilase (ASPAC), percutaneous transluminal coronary angioplasty (PTCA). All relative treatment effects not involving SK were calculated using the consistency equations. Compare to 102 data points. Although the UME model is a better fit (lower posterior mean of the residual deviance), the DICs are very similar for both models because the UME model has 7 more parameters than the NMA model does. A plot of the individual data points’ posterior mean deviance contribution in each of the 2 models highlights 4 data points that fit poorly to the consistency model (Figure 4). These points correspond to the 2 arms of trials 44 and 45, which were the only 2 trials comparing Acc t-PA to ASPAC. Furthermore, the posterior estimates of the treatment effects of ASPAC versus Acc t-PA (Table 2) in the consistency and UME models differ markedly. The fact that the 2 trials on this contrast give similar results to each other but are in conflict with the remaining evidence supports the notion that there is a systematic inconsistency.
Figure 4

Plot of the individual data points’ posterior mean deviance contributions for the consistency model (horizontal axis) and the unrelated mean effects model (vertical axis) along with the line of equality. Points that have a better fit in the unrelated mean effects model have been marked with the trial number.

Plot of the individual data points’ posterior mean deviance contributions for the consistency model (horizontal axis) and the unrelated mean effects model (vertical axis) along with the line of equality. Points that have a better fit in the unrelated mean effects model have been marked with the trial number.

Other Methods for Detecting Inconsistency

Variance measures of inconsistency

In the UME models described above, a different basic parameter represents each contrast. One can reparameterize the 6-parameter UME model so that instead of 6 treatment effect parameters (d AB, d AC, d AD, d BC, d BD, d CD) we have (d AB, d AC, d AD, ωBC, ωBD, ωCD), where The ω, ω, ω parameters are the inconsistencies between the direct and indirect evidence on these 3 edges. However, rather than considering the 3 inconsistency parameters as unrelated, we might assume that they all come from a random distribution, for example, ω, ~ N(0,), where this additional between-contrast variance serves as a measure of inconsistency.[6,15] We do not recommend this, however, because measures of variance will have very wide credible intervals unless the ICDF is extremely high. Even then, large numbers of large trials on each contrast would be required to obtain a meaningful estimate. Furthermore, where there is a single loop (ICDF = 1), it should be impossible to obtain any estimate of . See Salanti and others[26] for further comments on this issue.

Node splitting

A more sophisticated approach, which needs to be implemented in a Bayesian MCMC framework, is node splitting.[25] This is a powerful and robust method that can be recommended as a further option for inconsistency analysis in complex networks. It allows the user to split the information contributing to estimates of a parameter (node), say, d XY, into 2 distinct components: the direct based on all the XY data (which may come from XY, XYZ, WXY trials) and the indirect based on all the remaining evidence. The process can be applied to any contrast (node) in the network and in networks of any complexity. Like the UME model above, a shared variance term solves the difficulties created in an RE model when some contrasts are supported by only 1 or 2 trials. Node splitting can also generate intuitive graphics showing the difference between the estimates based on direct, indirect, and combined evidence.

Discussion

Although it is essential to carry out tests for inconsistency, this should not be considered in an overly mechanical way. Detection of inconsistency, like the detection of any statistical interaction, requires far more data than is needed to establish the presence of a treatment effect. The null hypothesis of consistency will therefore nearly always fail to be rejected, although this does not mean that there is no inconsistency. The mechanisms that potentially could create bias in indirect comparisons appear be to identical to those that cause heterogeneity in pairwise meta-analysis. Thus, to ensure that conclusions based on indirect evidence are sound, we must attend to the direct evidence on which they are based, as is clear from equation 1. This states that if the direct estimates of the AB and AC effects are unbiased estimates of the treatment effects in the target population, the indirect estimate of the BC effect must be unbiased as well. Conversely, any bias in the direct estimates, for example, due to effect-modifying covariates arising from the patients not being drawn from the target population, will be passed on to the indirect estimates in equal measure. The term bias in this context must be seen broadly, comprising both internal and external threats to validity.[27] So if the direct evidence on AB is based on trials conducted on a patient population different from the target, and a treatment effect modifier is present, using the AB trials to draw inferences about the target population can be considered as (external) bias, which will be inherited by any indirect estimates based on these data. Thus, the question, “Are conclusions based on indirect evidence reliable?” should be considered alongside the question, “Are conclusions based on pairwise meta-analysis reliable?” Any steps that can be taken to avoid between-trial heterogeneity will be effective in reducing the risk of drawing incorrect conclusions from pairwise meta-analysis, indirect comparisons, and NMA alike. In the decision-making context, the most obvious sources of potential heterogeneity of effect, such as differences in dose or differences in cotherapies, will already have been eliminated when defining the scope, which is likely to restrict the set of trials to specific doses and cotherapies. Clear cases in which direct and indirect evidence are in conflict are rare in the literature.[14,28] Where inconsistency has been evident, it illustrates the danger introduced by heterogeneity and in particular by the practice of trying to combine evidence on disparate treatment doses or treatment combinations within meta-analyses, often termed lumping.[1,29] The place for an enquiry into consistency is alongside a consideration of heterogeneity and its causes and, where appropriate, the reduction of heterogeneity through covariate adjustment (meta-regression) and bias adjustment.[27,30,31] This suggests that the risk of inconsistency is greatly reduced if between-trial heterogeneity is low. Empirical assessment of heterogeneity can therefore provide some reassurance or alert investigators to the risk of inconsistency. Tests of homogeneity in the pairwise comparisons can be used, or the posterior summaries of the distribution of the between-trials standard deviation can be compared to the size of the mean treatment effects. A second useful indicator is the between-trials variation in the trial baselines. If the treatment arms representing placebo or a standard treatment have similar proportions of events, this suggests that the trial populations are relatively homogeneous and that there will be little heterogeneity in the treatment effects. If, on the other hand, the baselines are highly heterogeneous, there is a potential risk of heterogeneity in the relative effects. Heterogeneity in baselines can be examined via a Bayesian synthesis.[32,33] One possible cause of inconsistency is a poor choice of scale of measurement, which can also lead to increased heterogeneity.[20,34] It is not always obvious whether to model treatment effects on a risk difference, logit, or other scale. The choice of the most appropriate scale is essentially an empirical one, although there is seldom enough evidence to decide on the basis of goodness of fit.[8,10] The choice of method used to test for inconsistency should be guided by the evidence structure. If it is possible to construct independent tests, then the Bucher method or its extensions represent the most simple and complete approach. In more complex networks, a repeated application of the Bucher method to all the possible loops produces interpretable results as long as no “significant” inconsistencies are found. Each application of the Bucher method is a valid test at its stated significance level. However, if inconsistencies are found when applying the test to all loops in the network, correction for multiple testing is needed, but it is difficult to specify how this should be done. In networks where multiarm trials are included, assessment of inconsistency becomes more problematic, as the presence of such internally consistent trials tends to hide potential inconsistencies. Our suggestion of removing multiarm trials involved in the loop being checked can become quite cumbersome when there are multiple multiarm trials and multiple loops. Furthermore, removal of some trials may affect the estimated between-trials heterogeneity, which in turn may affect the detection of inconsistency. A careful examination of the network, paying special attention to which contrasts are informed by multiarm trials, how large these trials are, and how they are likely to affect estimates, is recommended. This can inform both the simple Bucher approach and the parameterization of the UME model. Within a Bayesian framework, a consistency model can be compared with a model without the consistency assumptions. Analyses of residual deviance can provide an omnibus test of global inconsistency and can also help locate it. Node splitting[25] is another effective method for comparing direct evidence to indirect evidence in complex networks, but measures of inconsistency variance[6] or incoherence variance[15] are not recommended as indicators of inconsistency. Although the Bucher approach is conceptually simpler and easy to apply, it requires 2 stages, whereas Bayesian approaches have the advantage of being one stage: There is no need to summarize the findings on each contrast first. The 2-stage approach introduces a particular difficulty in networks in which the evidence on some contrasts may be limited to a small number of trials. This is that the decision as to whether to fit an RE model must be taken for each contrast separately, and if there is only one study, only an FE analysis is available, even when there is clear evidence of heterogeneity on other contrasts. This causes a further problem: Under the null hypothesis of consistency that the method sets out to test, the true variances have to conform to special relationships known as “triangle inequalities,”[9] but separate estimation makes it hard to ensure these inequalities are met. The likelihood of detecting an inconsistency, therefore, will be highly sensitive to the pattern of evidence. The choice of FE or RE summaries in the first stage can determine whether inconsistency is detected in the second stage.[17] Interestingly, the UME model with its shared variance parameter offers a way of smoothing the estimates of between-trial heterogeneity. Sparse data also show drawbacks in the Bayesian methods, especially when an RE analysis is used. The difficulty is that the greater the degree of between-trials heterogeneity, the less likely it is for inconsistency to be detectable, but there is seldom enough data to estimate the between-trials variation. The practice of using vague prior distributions for the between-trials variation, combined with a lack of data, will generate posteriors that allow an unrealistically high variance. This, in turn, is likely to mask all but the most obvious signs of inconsistency. Although possible inconsistency should be investigated thoroughly, the preferred approach should be to consider potential sources of heterogeneity in advance. Finally, there has been little work on how to respond to inconsistency when it is detected in a network. It is a reasonable principle that decisions should be based on models that are internally coherent, that is, models in which d YZ = d XZ − d XY, and that these models should fit the data. If the data cannot be fitted by a coherent model, then some kind of adjustment must be made. Any adjustment in response to inconsistency is post hoc, and usually there will be a large number of different adjustments to the body of data that could eliminate the inconsistency. There are clear examples of this in the literature on multiparameter evidence synthesis in epidemiology applications,[35,36] emphasizing the importance of identifying potential causes of heterogeneity of effect at the scoping stage and potential internal biases in advance of synthesis. Similarly, although inconsistency in one part of the network does not necessarily imply that the entire body of evidence is to be considered suspect, a reexamination of all included studies is desirable. Inconsistency is not a property of individual studies but of loops of evidence, and it may not always be possible to isolate which loop is responsible for the detected inconsistency, let alone which edge.[6] Where several alternative adjustments are available, a sensitivity analysis is essential.
  26 in total

1.  Issues in the selection of a summary statistic for meta-analysis of clinical trials with binary outcomes.

Authors:  Jonathan J Deeks
Journal:  Stat Med       Date:  2002-06-15       Impact factor: 2.373

2.  The results of direct and indirect treatment comparisons in meta-analysis of randomized controlled trials.

Authors:  H C Bucher; G H Guyatt; L E Griffith; S D Walter
Journal:  J Clin Epidemiol       Date:  1997-06       Impact factor: 6.437

3.  Modeling between-trial variance structure in mixed treatment comparisons.

Authors:  Guobing Lu; Ae Ades
Journal:  Biostatistics       Date:  2009-08-17       Impact factor: 5.899

4.  Borrowing strength from external trials in a meta-analysis.

Authors:  J P Higgins; A Whitehead
Journal:  Stat Med       Date:  1996-12-30       Impact factor: 2.373

Review 5.  Primary angioplasty versus intravenous thrombolytic therapy for acute myocardial infarction: a quantitative review of 23 randomised trials.

Authors:  Ellen C Keeley; Judith A Boura; Cindy L Grines
Journal:  Lancet       Date:  2003-01-04       Impact factor: 79.321

Review 6.  Methodological problems in the use of indirect comparisons for evaluating healthcare interventions: survey of published systematic reviews.

Authors:  Fujian Song; Yoon K Loke; Tanya Walsh; Anne-Marie Glenny; Alison J Eastwood; Douglas G Altman
Journal:  BMJ       Date:  2009-04-03

7.  Evidence synthesis for decision making 3: heterogeneity--subgroups, meta-regression, bias, and bias-adjustment.

Authors:  Sofia Dias; Alex J Sutton; Nicky J Welton; A E Ades
Journal:  Med Decis Making       Date:  2013-07       Impact factor: 2.583

8.  Combination of direct and indirect evidence in mixed treatment comparisons.

Authors:  G Lu; A E Ades
Journal:  Stat Med       Date:  2004-10-30       Impact factor: 2.373

9.  Bias modelling in evidence synthesis.

Authors:  Rebecca M Turner; David J Spiegelhalter; Gordon C S Smith; Simon G Thompson
Journal:  J R Stat Soc Ser A Stat Soc       Date:  2009-01       Impact factor: 2.483

10.  Evidence synthesis for decision making 2: a generalized linear modeling framework for pairwise and network meta-analysis of randomized controlled trials.

Authors:  Sofia Dias; Alex J Sutton; A E Ades; Nicky J Welton
Journal:  Med Decis Making       Date:  2012-10-26       Impact factor: 2.583

View more
  222 in total

1.  Role of metformin in overweight and obese people without diabetes: a systematic review and network meta-analysis.

Authors:  Fuhai Hui; Yingshi Zhang; Tianshu Ren; Xiang Li; Mingyi Zhao; Qingchun Zhao
Journal:  Eur J Clin Pharmacol       Date:  2018-12-03       Impact factor: 2.953

Review 2.  Indirect Comparisons and Network Meta-Analyses.

Authors:  Corinna Kiefer; Sibylle Sturtz; Ralf Bender
Journal:  Dtsch Arztebl Int       Date:  2015-11-20       Impact factor: 5.594

Review 3.  Dressings and topical agents for treating pressure ulcers.

Authors:  Maggie J Westby; Jo C Dumville; Marta O Soares; Nikki Stubbs; Gill Norman
Journal:  Cochrane Database Syst Rev       Date:  2017-06-22

4.  Rapid network meta-analysis using data from Food and Drug Administration approval packages is feasible but with limitations.

Authors:  Lin Wang; Benjamin Rouse; Arielle Marks-Anglin; Rui Duan; Qiyuan Shi; Kevin Quach; Yong Chen; Christopher Cameron; Christopher H Schmid; Tianjing Li
Journal:  J Clin Epidemiol       Date:  2019-06-18       Impact factor: 6.437

5.  Comparison of the Efficacy and Safety of Tofacitinib and Apremilast in Patients with Active Psoriatic Arthritis: A Bayesian Network Meta-Analysis of Randomized Controlled Trials.

Authors:  Gwan Gyu Song; Young Ho Lee
Journal:  Clin Drug Investig       Date:  2019-05       Impact factor: 2.859

6.  Comparison of the efficacy and tolerability of tocilizumab, sarilumab, and sirukumab in patients with active rheumatoid arthritis: a Bayesian network meta-analysis of randomized controlled trials.

Authors:  Sang-Cheol Bae; Young Ho Lee
Journal:  Clin Rheumatol       Date:  2018-02-05       Impact factor: 2.980

7.  Comparative efficacy and safety of tacrolimus, mycophenolate mofetil, azathioprine, and cyclophosphamide as maintenance therapy for lupus nephritis : A Bayesian network meta-analysis of randomized controlled trials.

Authors:  Y H Lee; G G Song
Journal:  Z Rheumatol       Date:  2017-12       Impact factor: 1.372

8.  Comparative efficacy and safety of biosimilar adalimumab and originator adalimumab in combination with methotrexate in patients with active rheumatoid arthritis: a Bayesian network meta-analysis of randomized controlled trials.

Authors:  Sang-Cheol Bae; Young Ho Lee
Journal:  Clin Rheumatol       Date:  2018-02-01       Impact factor: 2.980

9.  Is It Necessary to Perform the Pharmacological Interventions for Intrahepatic Cholestasis of Pregnancy? A Bayesian Network Meta-Analysis.

Authors:  Yi Shen; Jie Zhou; Sheng Zhang; Xu-Lin Wang; Yu-Long Jia; Shu He; Yuan-Yuan Wang; Wen-Chao Li; Jian-Guo Shao; Xun Zhuang; Yuan-Lin Liu; Gang Qin
Journal:  Clin Drug Investig       Date:  2019-01       Impact factor: 2.859

Review 10.  Efficacy and Safety of Antibiotic Therapy in Early Cutaneous Lyme Borreliosis: A Network Meta-analysis.

Authors:  Gabriel Torbahn; Heidelore Hofmann; Gerta Rücker; Karin Bischoff; Michael H Freitag; Rick Dersch; Volker Fingerle; Edith Motschall; Joerg J Meerpohl; Christine Schmucker
Journal:  JAMA Dermatol       Date:  2018-11-01       Impact factor: 10.282

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.