Loukia M Spineli1. 1. Midwifery Research and Education Unit, Hannover Medical School, Hannover, Germany.
Abstract
BACKGROUND: The unrelated mean effects (UME) model has been proposed for evaluating the consistency assumption globally in the network of interventions. However, the UME model does not accommodate multiarm trials properly and omits comparisons between nonbaseline interventions in the multiarm trials not investigated in 2-arm trials. METHODS: We proposed a refinement of the UME model that tackles the limitations mentioned above. We also accompanied the scatterplots on the posterior mean deviance contributions of the trial arms under the network meta-analysis (NMA) and UME models with Bland-Altman plots to detect outlying trials contributing to poor model fit. We applied the refined and original UME models to 2 networks with multiarm trials. RESULTS: The original UME model omitted more than 20% of the observed comparisons in both networks. The thorough inspection of the individual data points' deviance contribution using complementary plots in conjunction with the measures of model fit and the estimated between-trial variance indicated that the refined and original UME models revealed possible inconsistency in both examples. CONCLUSIONS: The refined UME model allows proper accommodation of the multiarm trials and visualization of all observed evidence in complex networks of interventions. Furthermore, considering several complementary plots to investigate deviance helps draw informed conclusions on the possibility of global inconsistency in the network. HIGHLIGHTS: We have refined the unrelated mean effects (UME) model to incorporate multiarm trials properly and to estimate all observed comparisons in complex networks of interventions.Forest plots with posterior summaries of all observed comparisons under the network meta-analysis and refined UME model can uncover the consequences of potential inconsistency in the network.Using complementary plots to investigate the individual data points' deviance contribution in conjunction with model fit measures and estimated heterogeneity aid in detecting possible inconsistency.
BACKGROUND: The unrelated mean effects (UME) model has been proposed for evaluating the consistency assumption globally in the network of interventions. However, the UME model does not accommodate multiarm trials properly and omits comparisons between nonbaseline interventions in the multiarm trials not investigated in 2-arm trials. METHODS: We proposed a refinement of the UME model that tackles the limitations mentioned above. We also accompanied the scatterplots on the posterior mean deviance contributions of the trial arms under the network meta-analysis (NMA) and UME models with Bland-Altman plots to detect outlying trials contributing to poor model fit. We applied the refined and original UME models to 2 networks with multiarm trials. RESULTS: The original UME model omitted more than 20% of the observed comparisons in both networks. The thorough inspection of the individual data points' deviance contribution using complementary plots in conjunction with the measures of model fit and the estimated between-trial variance indicated that the refined and original UME models revealed possible inconsistency in both examples. CONCLUSIONS: The refined UME model allows proper accommodation of the multiarm trials and visualization of all observed evidence in complex networks of interventions. Furthermore, considering several complementary plots to investigate deviance helps draw informed conclusions on the possibility of global inconsistency in the network. HIGHLIGHTS: We have refined the unrelated mean effects (UME) model to incorporate multiarm trials properly and to estimate all observed comparisons in complex networks of interventions.Forest plots with posterior summaries of all observed comparisons under the network meta-analysis and refined UME model can uncover the consequences of potential inconsistency in the network.Using complementary plots to investigate the individual data points' deviance contribution in conjunction with model fit measures and estimated heterogeneity aid in detecting possible inconsistency.
Entities:
Keywords:
Bland-Altman plot; consistency assumption; deviance information criterion; network meta-analysis
Systematic reviews with network meta-analysis (NMA) have been at the forefront of
evidence-based medicine over the past 2 decades.
The explosive rate of published systematic reviews with NMA from several health
care fields and a recent comprehensive review on the methodological advances of NMA
attest to the reception of this evidence synthesis design from the wide research
community.[1,2] NMA
has the advantage of providing a hierarchy of interventions for a specific research
question to assist the end users of systematic reviews in selecting the best
intervention for a condition. The intervention hierarchy results from modeling direct
evidence from the relevant clinical trials and indirect evidence for interventions never
compared in any clinical trial simultaneously. Consequently, NMA provides coherent
evidence for all possible comparisons of interventions under the investigated outcome.The credibility of the results from NMA strongly depends on the validity of the
consistency assumption that underlies this evidence synthesis tool. The consistency
assumption dictates the agreement of direct and indirect evidence for any pairwise
comparison in a closed loop of interventions (i.e., a path that starts and ends with the
same intervention).
The evaluation of the consistency assumption includes methods for local and
global detection of possible inconsistency.[2,5] The local evaluation is the most
prevalent in the published systematic reviews.[6,7] Among the methods for global
evaluation, the unrelated mean effects (UME) model, introduced by Dias et al.,
is the most frequently applied.
The UME model is particularly useful in complex networks, in which the
implementation of several statistical tests of inconsistency, such as the loop-specific approach,
may become cumbersome, challenging when loops are also informed by multiarm
trials and prone to multiplicity issues.A global evaluation using the UME model is achieved by comparing the Bayesian NMA model
with the Bayesian UME model using measures of model fit,
such as the deviance information criterion (DIC).
The model with the smaller DIC value by 3 or 5 units may be preferred regarding
model fit and complexity.
If the UME model fits the data better, this is evidence of possible inconsistency
in the network.
A scatterplot of the posterior mean deviance of the individual data points under
the UME model against the NMA model can reveal the trials with a higher than expected
posterior mean deviance.
These trials may help identify the loops with possible evidence of inconsistency.Furthermore, the inspection of the between-trial variance for substantial reductions also
offers valuable information on the suitability of the contrasted models. Suppose the
between-trial variance estimated from a model such as the UME that does not incorporate
the consistency assumption is substantially lower than that estimated from the NMA
model. This implies that the estimated between-trial variance had to increase for the
NMA model to fit well given the lack of consistent effects, thus suggesting potential
inconsistency.The presence of multiarm trials in the network may challenge the application of the UME
model. Suppose a comparison is informed by a multiarm trial alone. In that case,
selecting a different baseline intervention for that trial may omit this comparison from
the estimation process.
This was the case with the network of thrombolytic treatments that Dias et al.
considered to illustrate the UME model (figure 3 in Ref 5). A closed-loop of 3
interventions (SK, SK plus t-PA, and Acc t-PA) was informed by a multiarm trial and a
2-arm trial for 1 of the comparisons (SK versus SK plus t-PA). Of the 16 observed
comparisons, the UME model estimated 15 treatment effects, inevitably omitting 1 of the
comparisons in the multiarm trial (SK plus t-PA versus Acc t-PA). Considering a
different baseline intervention for this trial (e.g., SK plus t-PA) would have resulted
in the omission of a different comparison (in that case, SK versus Acc t-PA).The omitted comparisons in the multiarm trials may carry evidence of possible design
inconsistency in the network. The design inconsistency indicates disagreement in the
treatment effects across different designs (i.e., 2-arm and multiarm trials) for the
same comparison. Design inconsistency can be formally investigated using the
design-by-treatment interaction model.[12,13] Contrasting the NMA model with
the UME model regarding the treatment effect of all observed comparisons in the network
offers an informal, exploratory investigation of the design inconsistency.We aimed to propose a straightforward refinement of the UME model that accommodates the
multiarm trials properly and yields treatment effects for all observed comparisons in
networks with multiarm trials. Therefore, the proposed refinement allows the scrutiny of
all observed evidence in the network to draw informed conclusions about the possibility
of consistency. The article has the following structure. We first introduce 2 published
systematic reviews with NMA as motivating examples. Then we present the Bayesian
random-effects NMA and UME models—the latter as proposed by Dias et al.
(called the UME-Dias model). We follow by proposing a straightforward refinement
of the Bayesian random-effects UME model (called the refined UME model). We demonstrate
the refined UME model using the motivating examples. Finally, we conclude with a
discussion of the results, strengths, and limitations of the proposed refinement in
complex networks with multiarm trials.
Motivating Examples
We considered 2 motivating examples: 1) the network of antimanic drugs for the mean
change on mania rating scales,
and 2) the network of pharmacologic interventions for the exacerbation of
chronic obstructive pulmonary disease (COPD).
Both networks included multiarm trials. The corresponding articles analyzed
these networks in the standardized mean difference (SMD) and log odds ratio (OR)
scales. Both outcomes were harmful; therefore, a negative SMD or log OR favored the
first intervention in the comparison, and a positive SMD or log OR favored the
second intervention. There was information on the number of missing (participant)
outcome data in each arm of every trial for both networks. We excluded the missing
outcome data from the analysis for illustrative purposes. Namely, we subtracted the
number of missing outcome data from the number randomized in each arm of every
trial. Methods to properly account for aggregate missing outcome data in NMA have
been described elsewhere.[16,17]
Methods
Random-Effects NMA Model with Multiarm Trials
For a network of
trials comparing different sets of interventions, we have the
following information for the investigated outcome in each trial arm. For a
binary outcome, we collect the number of events,
, as reported in arm
of trial
out of the total randomized participants,
. We assume that
is sampled from a binomial distribution with an underlying
probability of an event,
. Then the underlying log odds in arm
of trial
is a function of the underlying log odds of the baseline arm,
, and the underlying log OR (the treatment effect),
,whereFor a continuous outcome, we extract the mean outcome,
, as reported in arm
of trial
alongside the standard deviation,
, as measured in the total randomized participants,
. We assume a normal distribution for
with an underlying mean outcome,
, and a variance,
that is commonly assumed known, even though it has been
estimated. Then, the underlying mean in arm
of trial
is a function of the underlying mean of the baseline arm,
, and the underlying SMD (the treatment effect),
,where
and
is the pooled standard deviation in that trial,with
being the number of arms in trial
.Under the random-effects model,
is assumed to follow a normal distribution with mean
(the consistency equation) and between-trial variance,
, assumed common within the network.
With
, we indicate the intervention studied in arm
of trial
, and with
, we indicate the reference intervention of the whole network.
The NMA model estimates the treatment effect of
comparisons with the selected reference intervention (i.e.,
with
and
) and uses the consistency equation to obtain the remaining
possible comparisons.Likewise, in a multiarm trial
, in which the consistency assumption is inherent, we estimate
the treatment effects of
comparisons with the baseline intervention of the trial. The
vector of random effects,
, is assumed to follow a multivariate normal distribution with
mean vector
and
variance-covariance matrix
(equation (10) in Dias et al.
), which is equivalent to conditional univariate normal distributions of
arm
given the arms from 2 to
(equation (11) in Dias et al.
),By assuming a common
in the whole network, the correlation between any 2 random
effects in the multiarm trial equals 0.5.
These random effects refer to contrasts with the baseline arm of the
multiarm trial. Choosing a different baseline arm will yield a different vector
.
Random-Effects UME Model of Dias and Colleagues
Contrary to the NMA model, the UME-Dias model does not pose consistency
equations. Therefore, the UME-Dias model comprises separate random-effects
pairwise meta-analyses for the observed comparisons.
The random-effects UME-Dias model also considers a shared between-trial
variance across the observed comparisons to borrow strength from comparisons
with many trials.[5,11] Hence, the estimation of between-trial variance is greatly improved.
Suppose the network includes a total of
observed comparisons with
. Then, under the random-effects model, the random effects
follow separate normal distributions with mean
(
,
and
) for the observed comparisons and shared between-trial
variance,
.In the absence of multiarm trials, the UME-Dias model estimates the treatment
effects of all observed comparisons. When the network includes multiarm trials,
the UME-Dias model estimates the same vector
with the NMA model in multiarm trial
. However, contrary to the NMA model, the UME model treats the
random effects in vector
as separate univariate normal distributions (se pp 651–2,
appendix in Ref 5),By making it similar to fitting separate pairwise meta-analyses to the data, the
random-effects UME-Dias model retains its simplicity.
However, of the
possible comparisons in that multiarm trial, the UME-Dias
model will not estimate a total of
comparisons. These comparisons do not include the baseline
arm. If these comparisons are not informed by any 2-arm trial and are not found
in the vector
of other multiarm trials in the network, they will be omitted
completely by the UME-Dias model. In the Bayesian framework, the posterior
distribution of
for the omitted comparisons will coincide with the prior distribution.
Hence, if a noninformative normal prior distribution with zero mean and
variance equal to 10,000 is assigned on
, the posterior standard deviation of
will be approximately 100 for the omitted comparisons.
Refined Random-Effects UME Model
The UME-Dias model does not properly accommodate the multiarm trials. The random
effects are inherently correlated in multiarm trials. Hence, as a matter of
principle, the conditional univariate normal distributions of the random effects
for the multiarm trials (equation (11) in Dias et al.
) should be maintained in the UME model. Dias et al.
also suggested accounting for the correlated effects in the multiarm
trials; however, the authors did not formally implement this model in their
book.In our proposed refinement of the UME-Dias model, we have maintained the
conditional univariate normal distributions of the random effects for the
multiarm trials. We have also developed an algorithm to automatically detect
pairwise comparisons in the multiarm trials that the UME-Dias model would omit.
When there is at least 1 omitted comparison, we perform another random-effects
NMA in the subset of multiarm trials, and we use the consistency equation to
obtain the summary treatment effect of the omitted comparisons. When the subset
of multiarm trials forms subnetworks rather than a fully connected network, we
perform random-effects NMA with consistency equations in each subnetwork
separately. To prevent the multiarm trials from contributing twice to the
estimation of
, we consider a different
for the subset of multiarm trials, indicated by
. This “companion” model does not share any common parameter
with the rest of the model. Therefore, using the multiarm trials twice does not
affect the estimation of the parameters outside the companion model.In the absence of multiarm trials, the refined UME model boils down to the
UME-Dias model. In the presence of multiarm trials, comparisons can also be
omitted when the fixed-effect model is considered.
In that case, the weighting approach proposed by Rücker and Schwarzer
can be used to prevent comparison omission. Dias et al.
discussed this weighting approach for the fixed-effect UME-Dias model in
a Bayesian context.
Model Implementation
For each network, we applied Bayesian random-effects NMA, the refined UME model,
and the UME-Dias model.
We considered a normal prior distribution with zero mean and variance
10,000 for the location parameters of the models, and we assigned a half-normal
prior distribution with scale parameter 1 on
and
. We considered 3 chains of different initial values, 100,000
iterations with 10,000 burn-in and thinning equal to 5 and 20 for the network of
antimanic drugs under the NMA and both UME models, respectively, and thinning
equal to 10 and 20 for the network for COPD under the NMA and both UME models,
respectively. We inspected the autocorrelation plots to define the values for
the thinning. Convergence was assessed using the Gelman-Rubin convergence
diagnostic, where
:
indicates a lack of convergence for the corresponding model node.
We applied JAGS via the R-package R2jags (statistical
software R, version 4.1.1) to run the models.[22-24]We tabulated the posterior median and posterior standard deviation of
, the posterior mean of the residual deviance (
), the DIC, and the number of effective parameters under all 3
models. A larger DIC by 5 units for the NMA model or
exceeding the number of data points under the NMA model
indicates a possible inconsistency in the network. We created scatterplots and
Bland-Altman plots on the posterior mean deviance contribution of the individual
points to detect possible data points where the compared models have a poor fit.
We considered the refined UME model to calculate the bias and limits of
agreement in the Bland-Altman plot. We also obtained the leverage plots
separately for the NMA and UME models to supplement our observations from the
scatterplot and Bland-Altman plots. We used forest plots to illustrate the
posterior mean and 95% credible interval of all observed comparisons under the
NMA, the refined UME, and UME-Dias models. All figures were created using the
R-package ggplot2.
For the network plots, we used the R-package pcnetmeta.
All functions related to the present article can be found as
supplementary material (Supplementary Material 1).
Results
Network of Antimanic Drugs
Figure 1A illustrates
the network of 13 antimanic drugs and placebo. There were 17 (27%) 3-arm trials
in the network. Of the 33 (36%) observed comparisons, 7 (21%) were informed
solely by at least 1 multiarm trial and contained no baseline intervention. The
UME-Dias model omitted these comparisons. The maximum value of
across the parameters of all 3 models was 1.03, thus
indicating convergence. All models yielded a posterior mean of residual deviance
that exceeded the total number of 141 trial arms. The NMA model provided the
largest
, followed by the UME-Dias and refined UME models
(
157.82, 149.73, and 146.67, respectively; Table 1). The DIC was
similar for the NMA and UME-Dias model but lower by 2.54 units for the refined
UME model (Table
1). The posterior median and 95% credible interval (CrI) of
were almost identical across the 3 models (Table 2). According
to the DIC and the estimated
, there was little to choose between a model with and without
consistency.
Figure 1
The network of antimanic drugs for the mean change on mania rating scales
(plot A). The network of pharmacologic interventions for
exacerbating chronic obstructive pulmonary disease
(plot B). Each node refers to an intervention, and each link
refers to a pairwise comparison. The size of the nodes is proportional
to the number of observed comparisons that include that node. The
thickness of the edge is proportional to the number of trials that
investigated that comparison. The colored intervention loops indicate
multiarm trials.
Table 1
Measures of Model Assessment for the NMA and UME Models
Model Assessment Measures
Network of Antimanic
Drugs
Network for COPD
NMA
UME
UME-Dias
NMA
UME
UME-Dias
DIC
264.28
261.54
264.08
92.15
90.47
90.56
pD
106.46
114.87
114.36
35.58
35.62
35.85
D¯
157.82
146.67
149.73
56.58
54.85
54.70
Total trial arms
141
50
COPD, chronic obstructive pulmonary disease;
, posterior mean of the residual deviance; DIC,
deviance information criterion; NMA, network meta-analysis;
, number of effective parameters; UME, unrelated
mean effects.
Table 2
Estimated between-Trial Standard Deviation (
) under the Compared Models
The network of antimanic drugs for the mean change on mania rating scales
(plot A). The network of pharmacologic interventions for
exacerbating chronic obstructive pulmonary disease
(plot B). Each node refers to an intervention, and each link
refers to a pairwise comparison. The size of the nodes is proportional
to the number of observed comparisons that include that node. The
thickness of the edge is proportional to the number of trials that
investigated that comparison. The colored intervention loops indicate
multiarm trials.Measures of Model Assessment for the NMA and UME ModelsCOPD, chronic obstructive pulmonary disease;
, posterior mean of the residual deviance; DIC,
deviance information criterion; NMA, network meta-analysis;
, number of effective parameters; UME, unrelated
mean effects.Estimated between-Trial Standard Deviation (
) under the Compared ModelsCOPD, chronic obstructive pulmonary disease; CrI, credible interval;
NMA, network meta-analysis; UME, unrelated mean effects.Overall, including the consistency equations yielded similar posterior mean
deviance contributions of the trial arms with discounting these equations and
almost identical deviance contributions for both UME models (Figure 2A). An exception
was 7 trial arms associated with larger deviance under the NMA model compared
with the UME models that greatly exceeded the expected deviance contribution
equal to 1. Of those trial arms, the arms of the 40th trial yielded a remarkably
large posterior mean deviance contribution under all models.
Figure 2
Scatterplot on the posterior mean deviance contributions of the trial
arms under the unrelated mean effects (UME) models (refined and Dias and
colleagues; y axis) and the network meta-analysis (NMA)
model (x axis) for the network of antimanic drugs (plot
A). The gray dotted lines refer to 1 posterior mean deviance
contribution. The Bland-Altman plot on the difference in the posterior
mean deviance contributions between the NMA and the UME models (refined
and Dias and colleagues) against the average posterior mean deviance
contributions of the compared models (plot B). Each data point
corresponds to a trial arm, indicated by a pair of numbers. The first
number refers to the trial ID, and the second number refers to the
trial’s arm, as placed in the analyzed data set (Supplementary Table S1).
Scatterplot on the posterior mean deviance contributions of the trial
arms under the unrelated mean effects (UME) models (refined and Dias and
colleagues; y axis) and the network meta-analysis (NMA)
model (x axis) for the network of antimanic drugs (plot
A). The gray dotted lines refer to 1 posterior mean deviance
contribution. The Bland-Altman plot on the difference in the posterior
mean deviance contributions between the NMA and the UME models (refined
and Dias and colleagues) against the average posterior mean deviance
contributions of the compared models (plot B). Each data point
corresponds to a trial arm, indicated by a pair of numbers. The first
number refers to the trial ID, and the second number refers to the
trial’s arm, as placed in the analyzed data set (Supplementary Table S1).In the Bland-Altman plot, the bias was slightly positive, indicating a tendency
for the NMA model to yield slightly larger posterior mean deviance contributions
on average than the refined UME model (Figure 2B). The trial arms associated
with larger deviance under the NMA model were found outside the upper limit of
agreements. These trial arms contributed 1 posterior mean deviance under the UME
models (Figure 2A). Two
of these points referred to 2-arm trials that were the sole contributors in
divalproex versus carbamazepine (the 28th trial) and lithium versus divalproex
(the 26th trial; Supplementary Table S1 in Supplementary Material 2). The
remaining outlying points referred to haloperidol and lithium compared only in
the multiarm trial (the 53rd trial), which included the omitted comparison of
paliperidone versus lithium (Supplementary Table S1 in Supplementary Material 2). Supplementary Figure S1 (in Supplementary Material 2) located
trials 28, 26, and 53 outside the red parabola (i.e.,
) of the leverage plot for the NMA model only (plot C) and
trial 40 (olanzapine versus lithium) outside the red parabola of the leverage
plot for all models. These points contributed to a DIC an amount larger than
3.Overall, there were similar posterior estimates and sufficient overlapping in the
95% CrIs of SMD for the NMA and either UME model except for the comparisons of
divalproex versus carbamazepine and of lithium versus divalproex and haloperidol
(Figure 3). These
comparisons corresponded to trials that contributed to the poor fit of the NMA
model (Figure 2;
Supplementary Figure S1), thus signaling possible inconsistency
in the network
Figure 3
A panel of forest plots on the standardized mean difference (SMD) for all
observed comparisons under the network meta-analysis (NMA), the refined
unrelated mean effects (UME) model, and the UME-Dias model. Results
refer to the posterior mean and 95% credible interval of the SMD. Gray
panels refer to the omitted comparisons. Red and black indicate weak and
strong evidence, respectively. Namely, the corresponding 95% credible
interval includes and excludes the null value, respectively.
A panel of forest plots on the standardized mean difference (SMD) for all
observed comparisons under the network meta-analysis (NMA), the refined
unrelated mean effects (UME) model, and the UME-Dias model. Results
refer to the posterior mean and 95% credible interval of the SMD. Gray
panels refer to the omitted comparisons. Red and black indicate weak and
strong evidence, respectively. Namely, the corresponding 95% credible
interval includes and excludes the null value, respectively.
Network of Pharmacologic Interventions for COPD
The network of 5 interventions (and their combinations) and placebo for COPD
included 5 (24%) multiarm trials: 2 three-arm and 3 four-arm trials (Figure 1B). Of the 15
(54%) observed comparisons, 6 (40%) were omitted from the UME-Dias model. All
models converged according to the
diagnostic: the maximum value across the parameters of all
models was 1.01. Both UME models yielded a similarly smaller posterior mean of
residual deviance (
) than the NMA model that exceeded the total number of 50 trial
arms. The DIC indicated little to choose between these models (Table 1). The
estimated τ was substantial in all models but slightly larger and less precise
under the NMA model (Table
2).Overall, both UME models yielded similar posterior mean deviance contributions
(Figure 4A). The
scatterplot indicated a poor fit of the NMA model for 2 trial arms that
exhibited substantial deviance under the NMA model and deviance close to 1 for
both UME models (Figure
4A). These trial arms were flagged as outliers in the Bland-Altman
plot as they exceeded the upper limit of agreement (Figure 4B). The 2 points referred to the
unique trial that compared formoterol with tiotropium (Supplementary Table S2 in Supplementary Material 2). This trial
was found outside the red parabola of the leverage plot for NMA (Supplementary Figure S2, plot C, in Supplementary Material 2)
alongside the 2-arm trial 1 (0 events in fluticasone; Supplementary Table S2). Trial 1 also contributed to the poor
fit of both UME models (Supplementary Figure S2, plots A and B) as both arms were
associated with a substantial posterior mean deviance (Figure 4A).
Figure 4
Scatterplot of the posterior mean deviance contributions of the trial
arms under the unrelated mean effects (UME) models (refined and Dias and
colleagues; y axis) and the network meta-analysis (NMA)
model (x axis) for the network of pharmacologic
interventions for chronic obstructive pulmonary disease (plot A). The
gray dotted lines refer to 1 posterior mean deviance contribution. The
Bland-Altman plot on the difference in the posterior mean deviance
contributions between the NMA and the UME models (refined and Dias and
colleagues) against the average posterior mean deviance contributions of
the compared models (plot B). Each data point corresponds to a trial
arm, indicated by a pair of numbers. The first number refers to the
trial ID, and the second number refers to the trial’s arm, as placed in
the analyzed data set (Supplementary Table S2).
Scatterplot of the posterior mean deviance contributions of the trial
arms under the unrelated mean effects (UME) models (refined and Dias and
colleagues; y axis) and the network meta-analysis (NMA)
model (x axis) for the network of pharmacologic
interventions for chronic obstructive pulmonary disease (plot A). The
gray dotted lines refer to 1 posterior mean deviance contribution. The
Bland-Altman plot on the difference in the posterior mean deviance
contributions between the NMA and the UME models (refined and Dias and
colleagues) against the average posterior mean deviance contributions of
the compared models (plot B). Each data point corresponds to a trial
arm, indicated by a pair of numbers. The first number refers to the
trial ID, and the second number refers to the trial’s arm, as placed in
the analyzed data set (Supplementary Table S2).The panel of forest plots illustrated almost identical results for the UME models
and, overall, similar results for the NMA and either UME model (Figure 5). For the
comparison of formoterol with tiotropium, the posterior estimate of OR differed
remarkedly between the NMA and UME models, which, in conjunction with the
substantial deviance contribution of the corresponding trial to the poor fit of
the NMA model (Figure
4), may suggest possible inconsistency in the network.
Figure 5
A panel of forest plots on the odds ratio (OR) for all observed
comparisons under the network meta-analysis (NMA), refined unrelated
mean effects (UME) model, and the UME-Dias model. The results refer to
the posterior mean and 95% credible interval of the OR. Gray panels
refer to the omitted comparisons. Red and black indicate weak and strong
evidence, respectively. Namely, the corresponding 95% credible interval
includes and excludes the null value, respectively.
A panel of forest plots on the odds ratio (OR) for all observed
comparisons under the network meta-analysis (NMA), refined unrelated
mean effects (UME) model, and the UME-Dias model. The results refer to
the posterior mean and 95% credible interval of the OR. Gray panels
refer to the omitted comparisons. Red and black indicate weak and strong
evidence, respectively. Namely, the corresponding 95% credible interval
includes and excludes the null value, respectively.
Discussion
The refined UME model and the UME-Dias model gave the same conclusions regarding
possible inconsistency in both examples. Both models elucidated the same data points
that contributed to possible inconsistency in the investigated networks. They also
yielded similar measures of model fit and almost identical estimated
. However, proper accommodation of the multiarm trials and
visualization of all observed comparisons were possible only with the refined UME
model, which are the main strengths of this model.The present study considered a series of complementary plots to thoroughly
investigate the individual data points’ deviance contribution, which laid the
foundation for signaling possible inconsistency in both examples. Specifically, the
scatterplot flagged the data points with higher posterior mean deviance than
expected for the NMA model. The Bland-Altman plot complemented the scatterplot by
detecting the outlying trial arms among those with substantial deviance contribution
as they laid outside the 95% limits of agreement. Furthermore, this plot offers
further exploratory insights that are not obvious by looking at the scatterplot. For
instance, the lack of randomness in the scattered points and a nonzero bias may
indicate a possible mismatch between the direct and NMA evidence that questions the
whole evidence base. The leverage plot for NMA revealed that the outlying data
points were found outside the red parabola, thus contributing substantially to the
DIC (an amount larger than 3) and the model’s poor fit. The panel of forest plots
pinpointed poor overlapping in the 95% CrI of the treatment effects of comparisons
informed by the outlying trials, showing the consequences of possible
inconsistency.The panel of forest plots should be used to aid model critique and highlight the
issues caused when potential inconsistency is detected rather than scrutinize the
relative effects between the NMA and UME models. A comparison of the posterior mean
of the treatment effects obtained via the NMA model with those obtained via the UME
model has been criticized as an inappropriate method to evaluate inconsistency.
This is because an NMA estimate is an amalgamation of direct and indirect
evidence.In the present study, we elaborated on the random-effects UME model for 2 reasons.
First, the random-effects model is more appropriate than the fixed-effect model in
systematic reviews, in which clinical and methodological heterogeneity should be
expected and may manifest as statistical heterogeneity.
Second, the proper accommodation of multiarm trials and related
parameterization issues are relevant in the random-effects model.We have developed an algorithm to automatically detect the omitted comparisons (if
any) in the network and incorporate them in the refined UME model. In Supplementary Material 1, we provide user-defined functions in R to
run the refined UME model in 1 step and obtain the necessary plots of the present
work. The user can employ the following effect measures: OR for binary outcomes,
mean difference, SMD, and ratio of means for continuous outcomes.The refined UME model is not immune to different parameterization, which comprises
our work’s major limitation. The choice of parameterization will affect the derived
omitted comparisons, because selecting a different baseline intervention for the
multiarm trials will derive different omitted comparisons.[5,11] Suppose a connected network
comprises an ABC, an AC, and an AB trial. Selecting intervention A to be the
baseline arm in the ABC trial will yield the BC comparison as omitted. Selecting
intervention B or C as the baseline arm in the ABC trial will not yield omitted
comparisons. Suppose the network did not include an AB trial. Selecting intervention
C as the baseline arm in the ABC trial will yield the AB as an omitted comparison.
However, regardless of the selected parameterization, the refined UME model will
estimate all observed comparisons of the network, contrary to the UME-Dias
model.Another limitation, common to the UME models, is that different parameterizations of
the multiarm trials may affect the estimates and possibly the measures of model fit.
Suppose there is also a BC trial in the example above, and we are interested
in the BC comparison. Since the multiarm and 2-arm trials inform all comparisons,
there are no omitted comparisons. However, selecting a different baseline arm will
lead to one comparison being informed solely by the corresponding 2-arm trial. When
A is the baseline arm in the ABC trial, the BC trial supports only the BC
comparison. When B or C is the baseline arm, both ABC and BC trials contribute to
the estimation of BC. The extent to which different parameterizations lead to
different conclusions may also depend on the extent of between-trial variance.
Different parameterizations may not affect the conclusions if the between-trial
variance is low.Useful clinical decisions can be made based only on models that assume the underlying
evidence is consistent to ensure coherent estimates for a proper incremental
assessment of benefits and costs. Models like the UME that synthesize evidence
without the consistency assumption are useful for assessing the feasibility of this
assumption. When the consistency assumption is not deemed feasible, the whole
evidence base should be called into question. Decisions should not be based on
results from any syntheses that do not properly account for the reasons for this
inconsistency.
Conclusion
The proposed refinement of the UME model handles multiarm trials properly, and it
yields treatment effects for all observed comparisons. A thorough inspection of the
deviance contribution of the individual data points in conjunction with visualizing
the posterior summaries of all observed comparisons under the NMA and refined UME
models can aid our conclusions about possible global inconsistency in the network.
In the presence of inconsistency, we should not be making inferences based on any of
the models because they do not adequately describe the totality of the evidence
available.Click here for additional data file.Supplemental material, sj-7z-1-mdm-10.1177_0272989X211068005 for A Revised
Framework to Evaluate the Consistency Assumption Globally in a Network of
Interventions by Loukia M. Spineli in Medical Decision MakingClick here for additional data file.Supplemental material, sj-docx-2-mdm-10.1177_0272989X211068005 for A Revised
Framework to Evaluate the Consistency Assumption Globally in a Network of
Interventions by Loukia M. Spineli in Medical Decision Making
Authors: Sofia Dias; Nicky J Welton; Alex J Sutton; Deborah M Caldwell; Guobing Lu; A E Ades Journal: Med Decis Making Date: 2013-07 Impact factor: 2.583
Authors: Adriani Nikolakopoulou; Anna Chaimani; Areti Angeliki Veroniki; Haris S Vasiliadis; Christopher H Schmid; Georgia Salanti Journal: PLoS One Date: 2014-01-22 Impact factor: 3.240