Literature DB >> 34961377

A Revised Framework to Evaluate the Consistency Assumption Globally in a Network of Interventions.

Loukia M Spineli1.   

Abstract

BACKGROUND: The unrelated mean effects (UME) model has been proposed for evaluating the consistency assumption globally in the network of interventions. However, the UME model does not accommodate multiarm trials properly and omits comparisons between nonbaseline interventions in the multiarm trials not investigated in 2-arm trials.
METHODS: We proposed a refinement of the UME model that tackles the limitations mentioned above. We also accompanied the scatterplots on the posterior mean deviance contributions of the trial arms under the network meta-analysis (NMA) and UME models with Bland-Altman plots to detect outlying trials contributing to poor model fit. We applied the refined and original UME models to 2 networks with multiarm trials.
RESULTS: The original UME model omitted more than 20% of the observed comparisons in both networks. The thorough inspection of the individual data points' deviance contribution using complementary plots in conjunction with the measures of model fit and the estimated between-trial variance indicated that the refined and original UME models revealed possible inconsistency in both examples.
CONCLUSIONS: The refined UME model allows proper accommodation of the multiarm trials and visualization of all observed evidence in complex networks of interventions. Furthermore, considering several complementary plots to investigate deviance helps draw informed conclusions on the possibility of global inconsistency in the network. HIGHLIGHTS: We have refined the unrelated mean effects (UME) model to incorporate multiarm trials properly and to estimate all observed comparisons in complex networks of interventions.Forest plots with posterior summaries of all observed comparisons under the network meta-analysis and refined UME model can uncover the consequences of potential inconsistency in the network.Using complementary plots to investigate the individual data points' deviance contribution in conjunction with model fit measures and estimated heterogeneity aid in detecting possible inconsistency.

Entities:  

Keywords:  Bland-Altman plot; consistency assumption; deviance information criterion; network meta-analysis

Mesh:

Year:  2021        PMID: 34961377      PMCID: PMC9189723          DOI: 10.1177/0272989X211068005

Source DB:  PubMed          Journal:  Med Decis Making        ISSN: 0272-989X            Impact factor:   2.749


Systematic reviews with network meta-analysis (NMA) have been at the forefront of evidence-based medicine over the past 2 decades. The explosive rate of published systematic reviews with NMA from several health care fields and a recent comprehensive review on the methodological advances of NMA attest to the reception of this evidence synthesis design from the wide research community.[1,2] NMA has the advantage of providing a hierarchy of interventions for a specific research question to assist the end users of systematic reviews in selecting the best intervention for a condition. The intervention hierarchy results from modeling direct evidence from the relevant clinical trials and indirect evidence for interventions never compared in any clinical trial simultaneously. Consequently, NMA provides coherent evidence for all possible comparisons of interventions under the investigated outcome. The credibility of the results from NMA strongly depends on the validity of the consistency assumption that underlies this evidence synthesis tool. The consistency assumption dictates the agreement of direct and indirect evidence for any pairwise comparison in a closed loop of interventions (i.e., a path that starts and ends with the same intervention). The evaluation of the consistency assumption includes methods for local and global detection of possible inconsistency.[2,5] The local evaluation is the most prevalent in the published systematic reviews.[6,7] Among the methods for global evaluation, the unrelated mean effects (UME) model, introduced by Dias et al., is the most frequently applied. The UME model is particularly useful in complex networks, in which the implementation of several statistical tests of inconsistency, such as the loop-specific approach, may become cumbersome, challenging when loops are also informed by multiarm trials and prone to multiplicity issues. A global evaluation using the UME model is achieved by comparing the Bayesian NMA model with the Bayesian UME model using measures of model fit, such as the deviance information criterion (DIC). The model with the smaller DIC value by 3 or 5 units may be preferred regarding model fit and complexity. If the UME model fits the data better, this is evidence of possible inconsistency in the network. A scatterplot of the posterior mean deviance of the individual data points under the UME model against the NMA model can reveal the trials with a higher than expected posterior mean deviance. These trials may help identify the loops with possible evidence of inconsistency. Furthermore, the inspection of the between-trial variance for substantial reductions also offers valuable information on the suitability of the contrasted models. Suppose the between-trial variance estimated from a model such as the UME that does not incorporate the consistency assumption is substantially lower than that estimated from the NMA model. This implies that the estimated between-trial variance had to increase for the NMA model to fit well given the lack of consistent effects, thus suggesting potential inconsistency. The presence of multiarm trials in the network may challenge the application of the UME model. Suppose a comparison is informed by a multiarm trial alone. In that case, selecting a different baseline intervention for that trial may omit this comparison from the estimation process. This was the case with the network of thrombolytic treatments that Dias et al. considered to illustrate the UME model (figure 3 in Ref 5). A closed-loop of 3 interventions (SK, SK plus t-PA, and Acc t-PA) was informed by a multiarm trial and a 2-arm trial for 1 of the comparisons (SK versus SK plus t-PA). Of the 16 observed comparisons, the UME model estimated 15 treatment effects, inevitably omitting 1 of the comparisons in the multiarm trial (SK plus t-PA versus Acc t-PA). Considering a different baseline intervention for this trial (e.g., SK plus t-PA) would have resulted in the omission of a different comparison (in that case, SK versus Acc t-PA). The omitted comparisons in the multiarm trials may carry evidence of possible design inconsistency in the network. The design inconsistency indicates disagreement in the treatment effects across different designs (i.e., 2-arm and multiarm trials) for the same comparison. Design inconsistency can be formally investigated using the design-by-treatment interaction model.[12,13] Contrasting the NMA model with the UME model regarding the treatment effect of all observed comparisons in the network offers an informal, exploratory investigation of the design inconsistency. We aimed to propose a straightforward refinement of the UME model that accommodates the multiarm trials properly and yields treatment effects for all observed comparisons in networks with multiarm trials. Therefore, the proposed refinement allows the scrutiny of all observed evidence in the network to draw informed conclusions about the possibility of consistency. The article has the following structure. We first introduce 2 published systematic reviews with NMA as motivating examples. Then we present the Bayesian random-effects NMA and UME models—the latter as proposed by Dias et al. (called the UME-Dias model). We follow by proposing a straightforward refinement of the Bayesian random-effects UME model (called the refined UME model). We demonstrate the refined UME model using the motivating examples. Finally, we conclude with a discussion of the results, strengths, and limitations of the proposed refinement in complex networks with multiarm trials.

Motivating Examples

We considered 2 motivating examples: 1) the network of antimanic drugs for the mean change on mania rating scales, and 2) the network of pharmacologic interventions for the exacerbation of chronic obstructive pulmonary disease (COPD). Both networks included multiarm trials. The corresponding articles analyzed these networks in the standardized mean difference (SMD) and log odds ratio (OR) scales. Both outcomes were harmful; therefore, a negative SMD or log OR favored the first intervention in the comparison, and a positive SMD or log OR favored the second intervention. There was information on the number of missing (participant) outcome data in each arm of every trial for both networks. We excluded the missing outcome data from the analysis for illustrative purposes. Namely, we subtracted the number of missing outcome data from the number randomized in each arm of every trial. Methods to properly account for aggregate missing outcome data in NMA have been described elsewhere.[16,17]

Methods

Random-Effects NMA Model with Multiarm Trials

For a network of trials comparing different sets of interventions, we have the following information for the investigated outcome in each trial arm. For a binary outcome, we collect the number of events, , as reported in arm of trial out of the total randomized participants, . We assume that is sampled from a binomial distribution with an underlying probability of an event, . Then the underlying log odds in arm of trial is a function of the underlying log odds of the baseline arm, , and the underlying log OR (the treatment effect), , where For a continuous outcome, we extract the mean outcome, , as reported in arm of trial alongside the standard deviation, , as measured in the total randomized participants, . We assume a normal distribution for with an underlying mean outcome, , and a variance, that is commonly assumed known, even though it has been estimated. Then, the underlying mean in arm of trial is a function of the underlying mean of the baseline arm, , and the underlying SMD (the treatment effect), , where and is the pooled standard deviation in that trial, with being the number of arms in trial . Under the random-effects model, is assumed to follow a normal distribution with mean (the consistency equation) and between-trial variance, , assumed common within the network. With , we indicate the intervention studied in arm of trial , and with , we indicate the reference intervention of the whole network. The NMA model estimates the treatment effect of comparisons with the selected reference intervention (i.e., with and ) and uses the consistency equation to obtain the remaining possible comparisons. Likewise, in a multiarm trial , in which the consistency assumption is inherent, we estimate the treatment effects of comparisons with the baseline intervention of the trial. The vector of random effects, , is assumed to follow a multivariate normal distribution with mean vector and variance-covariance matrix (equation (10) in Dias et al. ), which is equivalent to conditional univariate normal distributions of arm given the arms from 2 to (equation (11) in Dias et al. ), By assuming a common in the whole network, the correlation between any 2 random effects in the multiarm trial equals 0.5. These random effects refer to contrasts with the baseline arm of the multiarm trial. Choosing a different baseline arm will yield a different vector .

Random-Effects UME Model of Dias and Colleagues

Contrary to the NMA model, the UME-Dias model does not pose consistency equations. Therefore, the UME-Dias model comprises separate random-effects pairwise meta-analyses for the observed comparisons. The random-effects UME-Dias model also considers a shared between-trial variance across the observed comparisons to borrow strength from comparisons with many trials.[5,11] Hence, the estimation of between-trial variance is greatly improved. Suppose the network includes a total of observed comparisons with . Then, under the random-effects model, the random effects follow separate normal distributions with mean ( , and ) for the observed comparisons and shared between-trial variance, . In the absence of multiarm trials, the UME-Dias model estimates the treatment effects of all observed comparisons. When the network includes multiarm trials, the UME-Dias model estimates the same vector with the NMA model in multiarm trial . However, contrary to the NMA model, the UME model treats the random effects in vector as separate univariate normal distributions (se pp 651–2, appendix in Ref 5), By making it similar to fitting separate pairwise meta-analyses to the data, the random-effects UME-Dias model retains its simplicity. However, of the possible comparisons in that multiarm trial, the UME-Dias model will not estimate a total of comparisons. These comparisons do not include the baseline arm. If these comparisons are not informed by any 2-arm trial and are not found in the vector of other multiarm trials in the network, they will be omitted completely by the UME-Dias model. In the Bayesian framework, the posterior distribution of for the omitted comparisons will coincide with the prior distribution. Hence, if a noninformative normal prior distribution with zero mean and variance equal to 10,000 is assigned on , the posterior standard deviation of will be approximately 100 for the omitted comparisons.

Refined Random-Effects UME Model

The UME-Dias model does not properly accommodate the multiarm trials. The random effects are inherently correlated in multiarm trials. Hence, as a matter of principle, the conditional univariate normal distributions of the random effects for the multiarm trials (equation (11) in Dias et al. ) should be maintained in the UME model. Dias et al. also suggested accounting for the correlated effects in the multiarm trials; however, the authors did not formally implement this model in their book. In our proposed refinement of the UME-Dias model, we have maintained the conditional univariate normal distributions of the random effects for the multiarm trials. We have also developed an algorithm to automatically detect pairwise comparisons in the multiarm trials that the UME-Dias model would omit. When there is at least 1 omitted comparison, we perform another random-effects NMA in the subset of multiarm trials, and we use the consistency equation to obtain the summary treatment effect of the omitted comparisons. When the subset of multiarm trials forms subnetworks rather than a fully connected network, we perform random-effects NMA with consistency equations in each subnetwork separately. To prevent the multiarm trials from contributing twice to the estimation of , we consider a different for the subset of multiarm trials, indicated by . This “companion” model does not share any common parameter with the rest of the model. Therefore, using the multiarm trials twice does not affect the estimation of the parameters outside the companion model. In the absence of multiarm trials, the refined UME model boils down to the UME-Dias model. In the presence of multiarm trials, comparisons can also be omitted when the fixed-effect model is considered. In that case, the weighting approach proposed by Rücker and Schwarzer can be used to prevent comparison omission. Dias et al. discussed this weighting approach for the fixed-effect UME-Dias model in a Bayesian context.

Model Implementation

For each network, we applied Bayesian random-effects NMA, the refined UME model, and the UME-Dias model. We considered a normal prior distribution with zero mean and variance 10,000 for the location parameters of the models, and we assigned a half-normal prior distribution with scale parameter 1 on and . We considered 3 chains of different initial values, 100,000 iterations with 10,000 burn-in and thinning equal to 5 and 20 for the network of antimanic drugs under the NMA and both UME models, respectively, and thinning equal to 10 and 20 for the network for COPD under the NMA and both UME models, respectively. We inspected the autocorrelation plots to define the values for the thinning. Convergence was assessed using the Gelman-Rubin convergence diagnostic, where : indicates a lack of convergence for the corresponding model node. We applied JAGS via the R-package R2jags (statistical software R, version 4.1.1) to run the models.[22-24] We tabulated the posterior median and posterior standard deviation of , the posterior mean of the residual deviance ( ), the DIC, and the number of effective parameters under all 3 models. A larger DIC by 5 units for the NMA model or exceeding the number of data points under the NMA model indicates a possible inconsistency in the network. We created scatterplots and Bland-Altman plots on the posterior mean deviance contribution of the individual points to detect possible data points where the compared models have a poor fit. We considered the refined UME model to calculate the bias and limits of agreement in the Bland-Altman plot. We also obtained the leverage plots separately for the NMA and UME models to supplement our observations from the scatterplot and Bland-Altman plots. We used forest plots to illustrate the posterior mean and 95% credible interval of all observed comparisons under the NMA, the refined UME, and UME-Dias models. All figures were created using the R-package ggplot2. For the network plots, we used the R-package pcnetmeta. All functions related to the present article can be found as supplementary material (Supplementary Material 1).

Results

Network of Antimanic Drugs

Figure 1A illustrates the network of 13 antimanic drugs and placebo. There were 17 (27%) 3-arm trials in the network. Of the 33 (36%) observed comparisons, 7 (21%) were informed solely by at least 1 multiarm trial and contained no baseline intervention. The UME-Dias model omitted these comparisons. The maximum value of across the parameters of all 3 models was 1.03, thus indicating convergence. All models yielded a posterior mean of residual deviance that exceeded the total number of 141 trial arms. The NMA model provided the largest , followed by the UME-Dias and refined UME models ( 157.82, 149.73, and 146.67, respectively; Table 1). The DIC was similar for the NMA and UME-Dias model but lower by 2.54 units for the refined UME model (Table 1). The posterior median and 95% credible interval (CrI) of were almost identical across the 3 models (Table 2). According to the DIC and the estimated , there was little to choose between a model with and without consistency.
Figure 1

The network of antimanic drugs for the mean change on mania rating scales (plot A). The network of pharmacologic interventions for exacerbating chronic obstructive pulmonary disease (plot B). Each node refers to an intervention, and each link refers to a pairwise comparison. The size of the nodes is proportional to the number of observed comparisons that include that node. The thickness of the edge is proportional to the number of trials that investigated that comparison. The colored intervention loops indicate multiarm trials.

Table 1

Measures of Model Assessment for the NMA and UME Models

Model Assessment MeasuresNetwork of Antimanic Drugs
Network for COPD
NMAUMEUME-DiasNMAUMEUME-Dias
DIC264.28261.54264.0892.1590.4790.56
pD 106.46114.87114.3635.5835.6235.85
D¯ 157.82146.67149.7356.5854.8554.70
Total trial arms14150

COPD, chronic obstructive pulmonary disease; , posterior mean of the residual deviance; DIC, deviance information criterion; NMA, network meta-analysis; , number of effective parameters; UME, unrelated mean effects.

Table 2

Estimated between-Trial Standard Deviation ( ) under the Compared Models

τ Network of Antimanic Drugs
Network for COPD
NMAUMEUME-DiasNMAUMEUME-Dias
Posterior median0.140.140.140.180.150.14
95% CrI(0.09, 0.21)(0.09, 0.21)(0.08, 0.20)(0.02, 0.40)(0.01, 0.35)(0.01, 0.35)

COPD, chronic obstructive pulmonary disease; CrI, credible interval; NMA, network meta-analysis; UME, unrelated mean effects.

The network of antimanic drugs for the mean change on mania rating scales (plot A). The network of pharmacologic interventions for exacerbating chronic obstructive pulmonary disease (plot B). Each node refers to an intervention, and each link refers to a pairwise comparison. The size of the nodes is proportional to the number of observed comparisons that include that node. The thickness of the edge is proportional to the number of trials that investigated that comparison. The colored intervention loops indicate multiarm trials. Measures of Model Assessment for the NMA and UME Models COPD, chronic obstructive pulmonary disease; , posterior mean of the residual deviance; DIC, deviance information criterion; NMA, network meta-analysis; , number of effective parameters; UME, unrelated mean effects. Estimated between-Trial Standard Deviation ( ) under the Compared Models COPD, chronic obstructive pulmonary disease; CrI, credible interval; NMA, network meta-analysis; UME, unrelated mean effects. Overall, including the consistency equations yielded similar posterior mean deviance contributions of the trial arms with discounting these equations and almost identical deviance contributions for both UME models (Figure 2A). An exception was 7 trial arms associated with larger deviance under the NMA model compared with the UME models that greatly exceeded the expected deviance contribution equal to 1. Of those trial arms, the arms of the 40th trial yielded a remarkably large posterior mean deviance contribution under all models.
Figure 2

Scatterplot on the posterior mean deviance contributions of the trial arms under the unrelated mean effects (UME) models (refined and Dias and colleagues; y axis) and the network meta-analysis (NMA) model (x axis) for the network of antimanic drugs (plot A). The gray dotted lines refer to 1 posterior mean deviance contribution. The Bland-Altman plot on the difference in the posterior mean deviance contributions between the NMA and the UME models (refined and Dias and colleagues) against the average posterior mean deviance contributions of the compared models (plot B). Each data point corresponds to a trial arm, indicated by a pair of numbers. The first number refers to the trial ID, and the second number refers to the trial’s arm, as placed in the analyzed data set (Supplementary Table S1).

Scatterplot on the posterior mean deviance contributions of the trial arms under the unrelated mean effects (UME) models (refined and Dias and colleagues; y axis) and the network meta-analysis (NMA) model (x axis) for the network of antimanic drugs (plot A). The gray dotted lines refer to 1 posterior mean deviance contribution. The Bland-Altman plot on the difference in the posterior mean deviance contributions between the NMA and the UME models (refined and Dias and colleagues) against the average posterior mean deviance contributions of the compared models (plot B). Each data point corresponds to a trial arm, indicated by a pair of numbers. The first number refers to the trial ID, and the second number refers to the trial’s arm, as placed in the analyzed data set (Supplementary Table S1). In the Bland-Altman plot, the bias was slightly positive, indicating a tendency for the NMA model to yield slightly larger posterior mean deviance contributions on average than the refined UME model (Figure 2B). The trial arms associated with larger deviance under the NMA model were found outside the upper limit of agreements. These trial arms contributed 1 posterior mean deviance under the UME models (Figure 2A). Two of these points referred to 2-arm trials that were the sole contributors in divalproex versus carbamazepine (the 28th trial) and lithium versus divalproex (the 26th trial; Supplementary Table S1 in Supplementary Material 2). The remaining outlying points referred to haloperidol and lithium compared only in the multiarm trial (the 53rd trial), which included the omitted comparison of paliperidone versus lithium (Supplementary Table S1 in Supplementary Material 2). Supplementary Figure S1 (in Supplementary Material 2) located trials 28, 26, and 53 outside the red parabola (i.e., ) of the leverage plot for the NMA model only (plot C) and trial 40 (olanzapine versus lithium) outside the red parabola of the leverage plot for all models. These points contributed to a DIC an amount larger than 3. Overall, there were similar posterior estimates and sufficient overlapping in the 95% CrIs of SMD for the NMA and either UME model except for the comparisons of divalproex versus carbamazepine and of lithium versus divalproex and haloperidol (Figure 3). These comparisons corresponded to trials that contributed to the poor fit of the NMA model (Figure 2; Supplementary Figure S1), thus signaling possible inconsistency in the network
Figure 3

A panel of forest plots on the standardized mean difference (SMD) for all observed comparisons under the network meta-analysis (NMA), the refined unrelated mean effects (UME) model, and the UME-Dias model. Results refer to the posterior mean and 95% credible interval of the SMD. Gray panels refer to the omitted comparisons. Red and black indicate weak and strong evidence, respectively. Namely, the corresponding 95% credible interval includes and excludes the null value, respectively.

A panel of forest plots on the standardized mean difference (SMD) for all observed comparisons under the network meta-analysis (NMA), the refined unrelated mean effects (UME) model, and the UME-Dias model. Results refer to the posterior mean and 95% credible interval of the SMD. Gray panels refer to the omitted comparisons. Red and black indicate weak and strong evidence, respectively. Namely, the corresponding 95% credible interval includes and excludes the null value, respectively.

Network of Pharmacologic Interventions for COPD

The network of 5 interventions (and their combinations) and placebo for COPD included 5 (24%) multiarm trials: 2 three-arm and 3 four-arm trials (Figure 1B). Of the 15 (54%) observed comparisons, 6 (40%) were omitted from the UME-Dias model. All models converged according to the diagnostic: the maximum value across the parameters of all models was 1.01. Both UME models yielded a similarly smaller posterior mean of residual deviance ( ) than the NMA model that exceeded the total number of 50 trial arms. The DIC indicated little to choose between these models (Table 1). The estimated τ was substantial in all models but slightly larger and less precise under the NMA model (Table 2). Overall, both UME models yielded similar posterior mean deviance contributions (Figure 4A). The scatterplot indicated a poor fit of the NMA model for 2 trial arms that exhibited substantial deviance under the NMA model and deviance close to 1 for both UME models (Figure 4A). These trial arms were flagged as outliers in the Bland-Altman plot as they exceeded the upper limit of agreement (Figure 4B). The 2 points referred to the unique trial that compared formoterol with tiotropium (Supplementary Table S2 in Supplementary Material 2). This trial was found outside the red parabola of the leverage plot for NMA (Supplementary Figure S2, plot C, in Supplementary Material 2) alongside the 2-arm trial 1 (0 events in fluticasone; Supplementary Table S2). Trial 1 also contributed to the poor fit of both UME models (Supplementary Figure S2, plots A and B) as both arms were associated with a substantial posterior mean deviance (Figure 4A).
Figure 4

Scatterplot of the posterior mean deviance contributions of the trial arms under the unrelated mean effects (UME) models (refined and Dias and colleagues; y axis) and the network meta-analysis (NMA) model (x axis) for the network of pharmacologic interventions for chronic obstructive pulmonary disease (plot A). The gray dotted lines refer to 1 posterior mean deviance contribution. The Bland-Altman plot on the difference in the posterior mean deviance contributions between the NMA and the UME models (refined and Dias and colleagues) against the average posterior mean deviance contributions of the compared models (plot B). Each data point corresponds to a trial arm, indicated by a pair of numbers. The first number refers to the trial ID, and the second number refers to the trial’s arm, as placed in the analyzed data set (Supplementary Table S2).

Scatterplot of the posterior mean deviance contributions of the trial arms under the unrelated mean effects (UME) models (refined and Dias and colleagues; y axis) and the network meta-analysis (NMA) model (x axis) for the network of pharmacologic interventions for chronic obstructive pulmonary disease (plot A). The gray dotted lines refer to 1 posterior mean deviance contribution. The Bland-Altman plot on the difference in the posterior mean deviance contributions between the NMA and the UME models (refined and Dias and colleagues) against the average posterior mean deviance contributions of the compared models (plot B). Each data point corresponds to a trial arm, indicated by a pair of numbers. The first number refers to the trial ID, and the second number refers to the trial’s arm, as placed in the analyzed data set (Supplementary Table S2). The panel of forest plots illustrated almost identical results for the UME models and, overall, similar results for the NMA and either UME model (Figure 5). For the comparison of formoterol with tiotropium, the posterior estimate of OR differed remarkedly between the NMA and UME models, which, in conjunction with the substantial deviance contribution of the corresponding trial to the poor fit of the NMA model (Figure 4), may suggest possible inconsistency in the network.
Figure 5

A panel of forest plots on the odds ratio (OR) for all observed comparisons under the network meta-analysis (NMA), refined unrelated mean effects (UME) model, and the UME-Dias model. The results refer to the posterior mean and 95% credible interval of the OR. Gray panels refer to the omitted comparisons. Red and black indicate weak and strong evidence, respectively. Namely, the corresponding 95% credible interval includes and excludes the null value, respectively.

A panel of forest plots on the odds ratio (OR) for all observed comparisons under the network meta-analysis (NMA), refined unrelated mean effects (UME) model, and the UME-Dias model. The results refer to the posterior mean and 95% credible interval of the OR. Gray panels refer to the omitted comparisons. Red and black indicate weak and strong evidence, respectively. Namely, the corresponding 95% credible interval includes and excludes the null value, respectively.

Discussion

The refined UME model and the UME-Dias model gave the same conclusions regarding possible inconsistency in both examples. Both models elucidated the same data points that contributed to possible inconsistency in the investigated networks. They also yielded similar measures of model fit and almost identical estimated . However, proper accommodation of the multiarm trials and visualization of all observed comparisons were possible only with the refined UME model, which are the main strengths of this model. The present study considered a series of complementary plots to thoroughly investigate the individual data points’ deviance contribution, which laid the foundation for signaling possible inconsistency in both examples. Specifically, the scatterplot flagged the data points with higher posterior mean deviance than expected for the NMA model. The Bland-Altman plot complemented the scatterplot by detecting the outlying trial arms among those with substantial deviance contribution as they laid outside the 95% limits of agreement. Furthermore, this plot offers further exploratory insights that are not obvious by looking at the scatterplot. For instance, the lack of randomness in the scattered points and a nonzero bias may indicate a possible mismatch between the direct and NMA evidence that questions the whole evidence base. The leverage plot for NMA revealed that the outlying data points were found outside the red parabola, thus contributing substantially to the DIC (an amount larger than 3) and the model’s poor fit. The panel of forest plots pinpointed poor overlapping in the 95% CrI of the treatment effects of comparisons informed by the outlying trials, showing the consequences of possible inconsistency. The panel of forest plots should be used to aid model critique and highlight the issues caused when potential inconsistency is detected rather than scrutinize the relative effects between the NMA and UME models. A comparison of the posterior mean of the treatment effects obtained via the NMA model with those obtained via the UME model has been criticized as an inappropriate method to evaluate inconsistency. This is because an NMA estimate is an amalgamation of direct and indirect evidence. In the present study, we elaborated on the random-effects UME model for 2 reasons. First, the random-effects model is more appropriate than the fixed-effect model in systematic reviews, in which clinical and methodological heterogeneity should be expected and may manifest as statistical heterogeneity. Second, the proper accommodation of multiarm trials and related parameterization issues are relevant in the random-effects model. We have developed an algorithm to automatically detect the omitted comparisons (if any) in the network and incorporate them in the refined UME model. In Supplementary Material 1, we provide user-defined functions in R to run the refined UME model in 1 step and obtain the necessary plots of the present work. The user can employ the following effect measures: OR for binary outcomes, mean difference, SMD, and ratio of means for continuous outcomes. The refined UME model is not immune to different parameterization, which comprises our work’s major limitation. The choice of parameterization will affect the derived omitted comparisons, because selecting a different baseline intervention for the multiarm trials will derive different omitted comparisons.[5,11] Suppose a connected network comprises an ABC, an AC, and an AB trial. Selecting intervention A to be the baseline arm in the ABC trial will yield the BC comparison as omitted. Selecting intervention B or C as the baseline arm in the ABC trial will not yield omitted comparisons. Suppose the network did not include an AB trial. Selecting intervention C as the baseline arm in the ABC trial will yield the AB as an omitted comparison. However, regardless of the selected parameterization, the refined UME model will estimate all observed comparisons of the network, contrary to the UME-Dias model. Another limitation, common to the UME models, is that different parameterizations of the multiarm trials may affect the estimates and possibly the measures of model fit. Suppose there is also a BC trial in the example above, and we are interested in the BC comparison. Since the multiarm and 2-arm trials inform all comparisons, there are no omitted comparisons. However, selecting a different baseline arm will lead to one comparison being informed solely by the corresponding 2-arm trial. When A is the baseline arm in the ABC trial, the BC trial supports only the BC comparison. When B or C is the baseline arm, both ABC and BC trials contribute to the estimation of BC. The extent to which different parameterizations lead to different conclusions may also depend on the extent of between-trial variance. Different parameterizations may not affect the conclusions if the between-trial variance is low. Useful clinical decisions can be made based only on models that assume the underlying evidence is consistent to ensure coherent estimates for a proper incremental assessment of benefits and costs. Models like the UME that synthesize evidence without the consistency assumption are useful for assessing the feasibility of this assumption. When the consistency assumption is not deemed feasible, the whole evidence base should be called into question. Decisions should not be based on results from any syntheses that do not properly account for the reasons for this inconsistency.

Conclusion

The proposed refinement of the UME model handles multiarm trials properly, and it yields treatment effects for all observed comparisons. A thorough inspection of the deviance contribution of the individual data points in conjunction with visualizing the posterior summaries of all observed comparisons under the NMA and refined UME models can aid our conclusions about possible global inconsistency in the network. In the presence of inconsistency, we should not be making inferences based on any of the models because they do not adequately describe the totality of the evidence available. Click here for additional data file. Supplemental material, sj-7z-1-mdm-10.1177_0272989X211068005 for A Revised Framework to Evaluate the Consistency Assumption Globally in a Network of Interventions by Loukia M. Spineli in Medical Decision Making Click here for additional data file. Supplemental material, sj-docx-2-mdm-10.1177_0272989X211068005 for A Revised Framework to Evaluate the Consistency Assumption Globally in a Network of Interventions by Loukia M. Spineli in Medical Decision Making
  19 in total

1.  The results of direct and indirect treatment comparisons in meta-analysis of randomized controlled trials.

Authors:  H C Bucher; G H Guyatt; L E Griffith; S D Walter
Journal:  J Clin Epidemiol       Date:  1997-06       Impact factor: 6.437

2.  Borrowing strength from external trials in a meta-analysis.

Authors:  J P Higgins; A Whitehead
Journal:  Stat Med       Date:  1996-12-30       Impact factor: 2.373

3.  Indirect and mixed-treatment comparison, network, or multiple-treatments meta-analysis: many names, many benefits, many concerns for the next generation evidence synthesis tool.

Authors:  Georgia Salanti
Journal:  Res Synth Methods       Date:  2012-06-11       Impact factor: 5.273

4.  Reduce dimension or reduce weights? Comparing two approaches to multi-arm studies in network meta-analysis.

Authors:  Gerta Rücker; Guido Schwarzer
Journal:  Stat Med       Date:  2014-06-18       Impact factor: 2.373

5.  Commentary: Heterogeneity in meta-analysis should be expected and appropriately quantified.

Authors:  Julian P T Higgins
Journal:  Int J Epidemiol       Date:  2008-10       Impact factor: 7.196

Review 6.  An overview of conducting systematic reviews with network meta-analysis.

Authors:  Deborah M Caldwell
Journal:  Syst Rev       Date:  2014-09-29

7.  An empirical comparison of Bayesian modelling strategies for missing binary outcome data in network meta-analysis.

Authors:  Loukia M Spineli
Journal:  BMC Med Res Methodol       Date:  2019-04-24       Impact factor: 4.615

8.  Evidence synthesis for decision making 4: inconsistency in networks of evidence based on randomized controlled trials.

Authors:  Sofia Dias; Nicky J Welton; Alex J Sutton; Deborah M Caldwell; Guobing Lu; A E Ades
Journal:  Med Decis Making       Date:  2013-07       Impact factor: 2.583

9.  Evidence synthesis for decision making 2: a generalized linear modeling framework for pairwise and network meta-analysis of randomized controlled trials.

Authors:  Sofia Dias; Alex J Sutton; A E Ades; Nicky J Welton
Journal:  Med Decis Making       Date:  2012-10-26       Impact factor: 2.583

10.  Characteristics of networks of interventions: a description of a database of 186 published networks.

Authors:  Adriani Nikolakopoulou; Anna Chaimani; Areti Angeliki Veroniki; Haris S Vasiliadis; Christopher H Schmid; Georgia Salanti
Journal:  PLoS One       Date:  2014-01-22       Impact factor: 3.240

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.