Literature DB >> 25810269

Estimating the Expected Value of Sample Information Using the Probabilistic Sensitivity Analysis Sample: A Fast, Nonparametric Regression-Based Method.

Mark Strong1, Jeremy E Oakley2, Alan Brennan1, Penny Breeze1.   

Abstract

Health economic decision-analytic models are used to estimate the expected net benefits of competing decision options. The true values of the input parameters of such models are rarely known with certainty, and it is often useful to quantify the value to the decision maker of reducing uncertainty through collecting new data. In the context of a particular decision problem, the value of a proposed research design can be quantified by its expected value of sample information (EVSI). EVSI is commonly estimated via a 2-level Monte Carlo procedure in which plausible data sets are generated in an outer loop, and then, conditional on these, the parameters of the decision model are updated via Bayes rule and sampled in an inner loop. At each iteration of the inner loop, the decision model is evaluated. This is computationally demanding and may be difficult if the posterior distribution of the model parameters conditional on sampled data is hard to sample from. We describe a fast nonparametric regression-based method for estimating per-patient EVSI that requires only the probabilistic sensitivity analysis sample (i.e., the set of samples drawn from the joint distribution of the parameters and the corresponding net benefits). The method avoids the need to sample from the posterior distributions of the parameters and avoids the need to rerun the model. The only requirement is that sample data sets can be generated. The method is applicable with a model of any complexity and with any specification of model parameter distribution. We demonstrate in a case study the superior efficiency of the regression method over the 2-level Monte Carlo method.
© The Author(s) 2015.

Entities:  

Keywords:  Bayesian decision theory; Monte Carlo methods; computational methods; economic evaluation model; expected value of sample information; generalized additive model.; nonparametric regression

Mesh:

Year:  2015        PMID: 25810269      PMCID: PMC4471064          DOI: 10.1177/0272989X15575286

Source DB:  PubMed          Journal:  Med Decis Making        ISSN: 0272-989X            Impact factor:   2.583


Health economic decision-analytic models are used to estimate the expected net benefits of competing decision options. The true values of the input parameters of such models are rarely known with certainty, and uncertainty in model parameters typically results in decision uncertainty. This may motivate decision makers to consider options for further data collection alongside the adoption of new technologies or to delay adoption until after data collection.[1,2] The value of learning an input parameter (or a group of input parameters) can be quantified by its partial expected value of perfect information (partial EVPI).[3-6] However, we are unlikely to be able to collect perfect information, and it is more useful to quantify the value of specific research designs that will inform the decision-making problem. The value of reducing, rather than eliminating, uncertainty through the collection of data is captured by the expected value of sample information (EVSI).[4,7,8] The EVSI for any particular data collection exercise will depend not only on the study design in question but also on the decision context.[8] Important factors include whether there are costs associated with either delaying or reversing adoption decisions,[1,9,10] whether the adoption decision will be fully implemented[11] and whether the proposed study extends across jurisdictions.[12,13] Some of these factors arise in moving from EVSI per patient to population EVSI, but others also arise in estimating EVSI per patient. While recognizing these issues, we do not discuss them further in this article but concentrate on the problem of computing per-person EVSI within a single jurisdiction under the assumption of costless reversibility with perfect implementation and with no delay in either the study or the adoption. The concept of EVSI was first discussed in the health economics literature well over a decade ago.[4,7,14,15] Despite this, very few research funding or design decisions are informed by EVSI. This, at least in part, reflects the computational burden of calculating EVSI via generic Monte Carlo sampling-based methods. For example, in a recently published cost-effectiveness study, the authors noted that to compute EVSI without assuming an approximation that the model was linear, it would have taken 7.5 days.[16] In another example, the proposed EVSI analysis would have taken 37.5 days.[17] Clearly, computation times of this order are prohibitive. The reason for the high computational cost of EVSI analysis is that, unless the model is of a certain form or unless certain parametric assumptions are made, a nested 2-level Monte Carlo scheme is required. In this scheme, plausible data sets are generated in an outer loop, and then conditional on each data set, samples are generated from the posterior distribution of the parameters in an inner loop. The model is run with each inner loop set of parameters to estimate the expected net benefits, conditional on the data sets that have been simulated in the outer loop. The computational cost arises primarily due to the repeated evaluation of the model within the inner loop but also due to the burden of repeated sampling in the inner loop. If the aim is to search over the study design space, then this problem is further compounded because EVSI itself must be repeatedly calculated. Another difficult arises with the 2-level Monte Carlo approach if the prior distribution of the model parameters is not conjugate to the data likelihood. In this case, generating the inner-loop samples will typically require Markov chain Monte Carlo (MCMC), and the repeated application of MCMC for each sampled data set adds to the computational burden. A fast approximation scheme for the inner-loop step has been proposed,[18,19] but this method requires repeated evaluation of partial derivatives of the log posterior density function and therefore considerable time and effort writing the necessary computer code on the part of the analyst. Computationally cheaper single-loop approaches are sometimes available, but these rely either on the model being of a certain form[7,20,21] or on assumptions of Normality of the mean incremental net benefits.[1] Fast single-loop methods also exist for computing partial EVPI for single parameters,[22,23] but these have not yet been extended to the computation of EVSI. In this article, we present a method for calculating per-patient EVSI that avoids the nested 2-level scheme, requiring only the single set of sampled model inputs and corresponding model outputs (i.e., net benefits) that is generated in a standard probability sensitivity analysis (PSA). The method is based on a nonparametric regression of the net benefits on data samples that are generated conditional on the sampled input parameters in the PSA and follows closely the nonparametric regression method for computing partial EVPI described in Strong et al.[24] The method makes no assumptions regarding the form of the model, does not require the use of MCMC, and does not require that the parameter prior is conjugate to the data likelihood. All that is required is that the data likelihood can be sampled (but not necessarily evaluated). The article is structured as follows. In the second section, we introduce the method and describe its general application. In the third section, we demonstrate the method in the case study model that was used for illustrative purposes in Ades et al.,[7] and in the fourth section, we present results. We conclude with a short discussion.

Method

EVSI is the expected difference between the value of the optimal decision based on some sample of data, informative for some subset of inputs, and the value of the decision made only with prior information.[3,4,7] To express this, we first introduce some notation. We assume that we are faced with D decision options, indexed d = 1, . . ., D, and have built a model NB(d, θ) that aims to predict the net benefit of decision option d given a vector of p input parameter values θ = (θ1, . . ., θ). The true values of the input parameters are assumed to be unknown, and we represent beliefs about the input parameters via their joint distribution p(θ). We index a sample drawn from the joint distribution of the parameters with a bracketed superscript, θ(, for sample draws n = 1, . . ., N. We envisage that we can collect data that will be informative for some subset of parameters. We consider the (as yet uncollected) data as a vector of random variables and denote this as uppercase X. We use lowercase x for some arbitrary realized (or sampled) vector of values from the distribution of X. We use the bracketed superscript notation to index a sample of data vectors x(, n = 1, . . ., N. We denote the expectation over the joint distribution of θ as . We denote the expectation over the distribution of the data X as and over the posterior distribution of θ given data X as . The expected value of our optimal decision, made only with current information is If we had data X that were informative for (some subset of) the inputs, then the optimal decision would be that with the greatest net benefit, after averaging over the joint distribution of the inputs conditional on the data, p(θ|X). The expected net benefit would be But, since X is uncollected, we must average over possible data sets, The distribution of X can be obtained by the marginalization of p(X, θ) = p(X|θ)p(θ), which suggests a straightforward Monte Carlo sampling scheme for X, that is, sample first a value θ* from the prior p(θ) and then sample X from the data likelihood p(X|θ = θ*). Note that the data likelihood will depend on the design of the study under consideration, and we return to this point when we discuss our case study. The EVSI is then the difference between equation (3) and equation (1), At this point, we note that we can reexpress (4) as The reason for the reexpression will become apparent when we discuss Monte Carlo sampling schemes for estimating EVSI.

The Monte Carlo Approach to Computing EVSI

A probabilistic sensitivity analysis (PSA) takes N samples from the joint distribution of the input parameters, {θ(1), . . ., θ(}, and generates a corresponding set of N net benefits {NB(d, θ(1)), . . ., NB(d, θ()} for each decision option d. From this, the usual Monte Carlo solution to the second term in equation (4) is The first term in equation (4) requires more work, and unless there are analytic solutions to the expectations, the usual approach is to use a nested 2-level Monte Carlo method.[7] Here, the estimator is given by where θ( are samples drawn from the posterior distribution of θ|x(, and x( are generated by first sampling θ( from p(θ) and then x( from p(X|θ = θ(). Subtracting equation (6) from equation (7) results in the 2-level Monte Carlo EVSI estimator However, if we arrange our sampling scheme such that it reflects equation (5), we obtain instead Here, both terms in the EVSI expression are evaluated using the same sampled values of θ, and hence we will obtain an estimator with smaller Monte Carlo error than if we were to use equation (8). The first problem with the nested 2-level scheme is the requirement to evaluate the net benefit function (i.e., to run the model) at each iteration of the inner loop, resulting in J×K model evaluations. If the model is slow to run and/or if J and K are large (to obtain adequate precision), then the scheme will be computationally burdensome. A second potential problem is the requirement to sample from the posterior distribution of the input parameters, conditional on the sampled data, that is, obtaining the j = 1, . . ., J samples θ( from each p(θ|x() in the inner loop. If p(X|θ) and p(θ) are conjugate, then the posterior distribution will be of a known form, and sampling from it will be straightforward. However, if conjugate forms are not appropriate, we may be required to resort to MCMC to generate the j = 1, . . ., J samples θ( from p(θ|x(). The MCMC step must be repeated for each of the k = 1, . . ., K sampled data values, and this will add considerably to the computational burden. Setting up the MCMC sampler (e.g., via writing BUGS code[25]) and checking the MCMC chain(s) for convergence also requires investment in modeler time. We note at this point that in some restricted cases, we can avoid entirely the inner-loop Monte Carlo step. If the model is linear or multilinear (i.e., of “sum-product” form) in the parameters, and if the parameters are independent of one another (and retain this independence after updating with data), and if we can analytically compute the posterior expectations of the parameters given the data, then we can simply “plug in” the expected parameter values into the net benefit equation to obtain the expected net benefit.[7,21,26]

Nonparametric Regression Method

The problem with the 2-level Monte Carlo scheme is the need to compute the inner expectation in the first term in equation (4) via Monte Carlo. Not only does this require J model runs for each outer loop, but it is this step that requires the potentially problematic sampling from the conditional distribution p(θ|X). We therefore propose to estimate this expectation as follows. Consider generating a random parameter vector θ from p(θ) and a random data sample X conditional on θ and then evaluating NB(d, θ). We recognize that we can express the observed NB(d, θ) as a sum of the conditional expectation that we require and a mean-zero error term, where the error ϵ is a function of both X and θ. To see why ϵ has zero mean, we rearrange to give and then take expectations with respect to both X and θ, giving noticing that the first term in the right-hand side of equation (11) is a function only of θ and the second term a function only of X (θ having been integrated out). Then, by the law of total expectation (or “tower rule”), , and hence . Note that this holds regardless of the distribution of θ and regardless of the relationship between θ and X. Next, we recognize that the expectation can be thought of as an unknown function of X. We denote this function g(d, X), and substituting this into equation (10) gives In some instances, the data X will be high dimensional (e.g., censored time-to-event data in a study that measures survival), and if this is so, we write the function g in terms of some low-dimensional summary statistic of the data T(X) = {T1(X), . . ., T(X)}, giving We discuss choice of summary statistic in the next section. Last, for each decision option d = 1, . . ., D, we treat the net benefits in the PSA sample, NB(d, θ() as “noisy” data through which we can learn about the target function g{d, T(x()} for k = 1, . . ., K. Thus, we can think of this as D regression problems. However, we immediately recognize that the target functions have unknown form, and we have no desire to impose any particular form. We could begin by fitting a standard linear model, with power and interaction terms to model the nonlinearity between the net benefits and the data, but we choose instead to adopt the more flexible “nonparametric” regression approach offered by the generalized additive model (GAM). GAM models assume that the expectation of the dependent variable is a smooth but usually unknown function of the independent variable, which is exactly what we need here. The standard linear model is a special case of the GAM model in which the functional form of the expectation of the dependent variable is specified a priori. When we adopt a GAM model, we usually represent the unknown underlying smooth function as some form of spline, a common choice being the cubic spline. In the simplest case, a univariate cubic spline represents an arbitrary smooth single-input function as a series of short cubic polynomials joined piecewise such that the function is twice-differentiable at the “knots” (i.e., join points). The same spline can also be represented as the weighted sum of a series of predetermined “basis functions” that extend over the whole range of the function input. Simple univariate cubic splines have natural extensions to higher dimensions and to a regression framework, where the spline parameters (i.e., the basis function weights) are estimated from noisy data. For an introduction to GAM models, see Hastie and Tibshirani[27] or Wood.[28] Returning now to the problem of estimating g{d, T(X)}, we obtain the necessary “data” for the GAM regression analysis as follows. We assume we have at our disposal a PSA sample of size K. This consists of a set of k = 1, . . ., K samples from the distribution of the input parameters θ(, and k = 1, . . ., K corresponding evaluations of the model NB(d, θ(), for each decision option d = 1, . . ., D. The dependent variable for regression d is the net benefit NB(d, θ() for decision option d. To generate realizations of the independent variable (which is common to all d = 1, . . ., D regression analyses), we sample, for each value θ(, a data set x( from the likelihood p(X|θ = θ() and, from x(, calculate the summary statistic T(x(). At no point do we attempt to derive or sample from the potentially difficult posterior distribution of the parameters, p(θ|x(), and at no point do we rerun the economic model. This is why the method is fast and simple. We note here that EVSI is invariant to the reexpression of net benefits as incremental net benefits, relative to some chosen “baseline” option. Under this reexpression, the (incremental) net benefit of the baseline option is zero. This reduces the number of regression equations from D to D− 1.

Choice of Summary Statistic T(·)

Study data x may be scalar or vector valued and may be informative for one or more economic model parameters. In the simplest case, we have scalar data that are informative for a single economic model parameter. Here we choose T(x) = x. If x is vector valued, but we still expect data only to update a single model parameter θ, and then we choose T(x) to be a sample estimator for θ. This leads to quite natural summary statistics. So, for example, if we wish to calculate the expected value of a 2-arm, binary outcome trial to update beliefs about an odds ratio, then we our choice of T(x) would be the sample odds ratio. If we wish to update beliefs about q > 1 economic model parameters {θ1, . . ., θ}, then we would calculate q summary statistics T(x) = {T1(x), . . ., T(x)}, where each T(x) is a sample estimator for θ. For example, if we wish to calculate the expected value of a study to learn about the shape and scale parameters of a Weibull distribution, {θ1, θ2}, and x are censored time-to-event data, then we would choose {T1(x), T2(x)} to be the sample estimates derived from a Weibull survival model. In the case of q > 1, we write the vector of scalar summary statistics as T(x) = {T1(x), . . ., T(x)}, and fit the GAM model NB (d, θ) = g{d, T(x)} +ϵ, where g is now a multivariate smoothing function.[28]

Hypothetical Nonparametric Regression Example

To give an example, we imagine a hypothetical net benefit function NB(θ) = θ2 (for clarity, we have dropped the decision option index d in this example). The parameter θ represents a proportion (e.g., of people in the population who have a certain characteristic), and current knowledge about the proportion is expressed via a Beta(40,200) distribution. We want to know the value of doing a study with 500 participants to learn about the proportion. The number of people in the study with the characteristic of interest, x, is modeled using a Binomial(θ, 500) distribution. The PSA sample comprises samples {θ(1), . . ., θ(} with the corresponding samples from the net benefit function {NB(θ(1)), . . ., NB(θ()}. For each sample θ(, we generate a sample of data x( from X|θ( ~ Binomial (θ(, 500). The data here are scalar, and we therefore choose T(x) = x. We fit a GAM regression of NB(θ) on T(x) and extract the fitted values, which are our estimates of . In this hypothetical example, we can calculate analytically, so can compare the GAM regression values with their true counterparts. Figure 1 shows a scatter plot of sampled values of the incremental net benefit, NB(θ) = θ2, versus sampled values of T(x). The two lines show the posterior expected net benefit as a function of the sampled values of T(x). The solid line shows the GAM model fitted values, and the dashed line shows the analytically calculated values.
Figure 1

Hypothetical example. Generalized additive model (GAM) model fitted values of the posterior expected incremental net benefit versus analytic values. The lines representing the GAM fitted and analytic values are almost indistinguishable.

Hypothetical example. Generalized additive model (GAM) model fitted values of the posterior expected incremental net benefit versus analytic values. The lines representing the GAM fitted and analytic values are almost indistinguishable.

Regression Diagnostics

As with all regression analysis, it is important to check assumptions. Most important, we require that there is no structure in the residuals (e.g., a U-shaped or S-shaped pattern) since this would suggest unmodeled structure in the target function g and therefore bias in the fitted values. We note that for the purposes of calculating EVSI, we are seeking to estimate only the posterior mean net benefits. We do not require the posterior variance of the net benefits for the EVSI computation and therefore whether the residuals have equal variance and follow a particular distribution is of secondary importance.[29] In contrast to the estimator for the EVSI, the estimator for the Monte Carlo standard error of the EVSI given in the online Appendix does rely on the net benefits having approximately equal variance and approximate Normality if the number of rows of the PSA is small. However, if the size of the PSA is large, then the standard error estimator can be justified on large sample results, even in the absence of Normality of the net benefits (see Appendix).[28]

EVSI Calculation

After fitting a GAM model for each decision option d, we then extract the regression model fitted values. The fitted values are estimates of g{d, T(x()}, k = 1, . . ., K, our target quantity. We denote the GAM fitted values for decision option d as , and the estimated EVSI is then given by Note that we choose rather than as the second term in the EVSI estimator, following expression (5) rather than expression (4). By choosing this as our estimator, we exploit the positive correlation between the two terms in equation (14) and hence estimate the EVSI with increased precision. The sampling scheme for the GAM regression-based EVSI is given in Box 1.
Box 1

Generalized Additive Model (GAM) Regression-Based EVSI Algorithm

Generate a PSA sample of size K: for k = 1, . . ., K do  Sample θ(k) from the distribution of the parameters, p(θ)  Evaluate the economic model to obtain (incremental) net benefits NB(d, θ(k)) end forGiven the PSA sample, simulate data samples: for k = 1, . . ., K do  Generate a data sample x(k) from p(X|θ(k))  Calculate summary statistic T(x(k)) end forFit regression models and calculate EVSI:  Regress net benefits NB(d, θ(k)) from the PSA on T(x(k)) for each d  Extract GAM model fitted values for each d  Calculate EVSI via equation (14).

PSA, probability sensitivity analysis; EVSI, expected value of sample information.

Generalized Additive Model (GAM) Regression-Based EVSI Algorithm PSA, probability sensitivity analysis; EVSI, expected value of sample information. Because we are averaging over k, we can think of this as a single-loop Monte Carlo method. The size of K will determine the precision of the estimate of the EVSI, and a method for estimating the standard error of the GAM-based approximation is given in the Appendix.

Case Study: Ades (2004) Decision Tree Model

Model

Our case study is based on the model that was used for illustrative purposes in Ades et al.[7] The decision problem has two options: d = 1 (standard care) and d = 2 (new treatment) and can be represented by a simple decision tree (Figure 2). There are 11 parameters in the model, which we write as the vector θ = (L, Q, λ). Parameter definitions and distributions are given in Table 1. The output of the model is the net benefit for each decision option in monetary units. The algebraic form of the model is given in equations (15) and (16), with some components of θ being redundant in each net benefit function.
Figure 2

Decision tree model. From Ades et al.,[7] copyright © 2004, Society for Medical Decision Making. Reprinted by Permission of SAGE Publishers.

Table 1

Case Study Parameter Distributions

DescriptionParameterMeanDistribution
Mean remaining lifetimeL 30Constant
QALY after critical event, per yearQE 0.6405logit(QE) ∼N(0.6, 1/6)
QALY decrement due to side effectsQSE 1Constant
Cost of critical eventCE $200,000Constant
Cost of treatmentCT $15,000Constant
Cost of treatment side effectsCSE $100,000Constant
Probability of critical event, no treatmentPC 0.15Beta(15,85)
Probability of treatment side effectsPSE 0.25Beta(3,9)
Odds ratio, (PT/(1 −PT ))/(PC / (1 −PC))OR 0.2636log(OR) ∼N(–1.5, 1/3)
Probability of critical event on treatmentPT 0.0440[derived from OR and PC]
Monetary value of 1 QALYλ $75,000Constant

QALY, quality-adjusted life year. Adapted from Ades et al.[7] Copyright © 2004. Reprinted with permission from SAGE Publications.

Decision tree model. From Ades et al.,[7] copyright © 2004, Society for Medical Decision Making. Reprinted by Permission of SAGE Publishers. Case Study Parameter Distributions QALY, quality-adjusted life year. Adapted from Ades et al.[7] Copyright © 2004. Reprinted with permission from SAGE Publications. The model is multilinear in the parameters, and all parameters are independent. Thus, the expectation of the net benefit, , is equal to the net benefit equation evaluated at the parameter expectations, . This is generally the case for decision tree models with independent parameters but not for Markov models or individual-level simulation models. For our case study, we consider the same 3 data collection scenarios presented in Ades et al.[7]—that is, data collection to inform the probability of side effects (P), the quality of life after critical event (Q), and the treatment effect size (OR). For each scenario, we calculated EVSI using 3 methods. First, we replicated the method presented in Ades et al. The method relies on the model being of multilinear form with independent parameters and that there are analytic solutions (or good approximations) to the posterior expectations of the parameters, conditional on simulated data. Hence, the method only requires a single-loop Monte Carlo scheme to evaluate the outer expectation in the first term in equation (4). Second, we used the 2-level Monte Carlo scheme outlined earlier, with an MCMC inner loop. This method does not rely on the model being multilinear with independent parameters or that there are analytic solutions to the posterior expectations of the parameters. However, it has high computational cost. Third, we used the GAM regression method presented earlier. As with the Ades et al. method, the GAM method uses Monte Carlo to evaluate the outer expectation in the first term in equation (4). For consistency, we use K to denote the size of the outer expectation Monte Carlo loop when reporting all 3 methods. Because all 3 methods use Monte Carlo to estimate the outer expectation in equation (4), there will be a Monte Carlo sampling error that tends to zero as the outer-loop sampling size increases. For each estimated EVSI, we calculated the Monte Carlo standard error using the methods presented in the Appendix. We repeated each analysis with a range of values of K to demonstrate the relationship between K and the Monte Carlo standard error. We also recorded the total CPU time required to undertake the EVSI computation to compare the efficiency of each method at different values of K. For the Ades et al.[7] method, we chose values of K equal to 104, 105, and 106. For the MCMC-based method, we chose an inner-loop sample size of J = 104 after an initial exploration to determine an adequate sample size to achieve stability of the inner-loop estimates. We then chose values of K equal to 104, 105, and 106. Values of K greater than this required prohibitively long runtimes. For the GAM-based method, we chose values of K equal to 104, 105, and 106.

Data Collection Scenario 1: EVSI for the Probability of Side Effects (P)

To reduce uncertainty about P, we considered the value of undertaking an observational study of n = 60 patients on the new treatment. The number of participants observed to experience a side effect is assumed to follow a Binomial(P, 60) distribution.

Method 1—single-loop method presented in Ades et al

In method 1, we replicated the single-loop method used in the case study in Ades et al.[7] First, we drew k = 1, . . ., K samples from the Beta(3, 9) prior distribution for P. Next, for each sampled value , we generated a sample of data x( from a Binomial(, 60) distribution. Due to conjugacy, the posterior distribution has the closed form P|x( ~ Beta (3 +x(, 69 −x(), with expectation . Because the economic model is multilinear in the parameters, we can calculate for each data set k the exact expected net benefit by plugging into equations (15) and (16) the posterior expectation , along with the expected values for the remaining uncertain parameters, , , and . Finally, we estimated the EVSI by

Method 2—2-level nested Monte Carlo/MCMC sampling scheme

In method 2, we implemented the generic 2-level Monte Carlo scheme outlined earlier. In an outer loop, we drew k = 1, . . ., K samples from the Beta(3, 9) prior distribution for P. For each value , we generated a sample of data x( from a Binomial(, 60) distribution. Conditional on each simulated trial data value x(, we then ran an inner-loop of size J = 104. At each run of the inner loop, we sampled a vector of parameter values θ( from the posterior distribution p(θ|x() and evaluated the model net benefit equation NB(d, θ(). Finally, we calculated EVSI via equation (9). Because x( is only informative for the parameter P, and P is independent of all other model parameters, drawing from p(θ|x() reduced in this case to drawing from the posterior distribution p(P|x() and the prior distributions for the remaining parameters. Although the posterior distribution p(P|x() has a known form in this example, we did not assume that this was so and implemented the inner loop in OpenBUGS.[25] At each inner-loop step, we discarded the first 1000 MCMC samples as a burn-in.

Method 3—GAM regression

In method 3, we implemented the GAM regression scheme outlined earlier. First, we generated a PSA sample of size K. We calculated the incremental net benefit for each PSA sample. Next, for each parameter vector θ( in the PSA sample, we generated a sample of data x( from a Binomial(, 60) distribution. Because the data in this case are scalar, the summary statistic is T(x() = x(. We regressed the incremental net benefits on T(x(), extracted the model-fitted values, and estimated the EVSI via equation (14).

Data Collection Scenario 2: EVSI for Quality of Life after Critical Event (Q)

To reduce uncertainty about Q, we considered the value of undertaking an observational study of n = 100 patients who have experienced a critical event. We assume that, conditional on Q, the sample mean of the logit transform of the quality of life reported in a single data collection exercise is Normally distributed with expectation logit(Q) and variance σ2/n, where σ2, the population variance, is assumed known and equal to 2 (see Ades et al.[7] for details). As in scenario 1, we replicated the single-loop method used in the case study in Ades et al.[7] First, we drew k = 1, . . ., K samples from the Normal(0.6, 1/6) prior distribution for logit(Q). Next, for each sampled value logit(Q)(, we generated a sample of data x( from a Normal{logit(Q)(, 1/50} distribution. Due to conjugacy, the posterior distribution of logit(Q)|x( is Normal{(0.6 × 6 +x(× 50)/(50 + 6),1/(50 + 6)}. The posterior expectation can then be estimated from the posterior distribution of the logit-transformed parameter, p(logit(Q)|x(), using a Taylor series method (described in full in Ades et al.). Because the economic model is multilinear in the parameters, we can calculate for each data set k the expected net benefits by plugging into equations (15) and (16) the posterior expectation , along with the expected values for the remaining uncertain parameters, , and . We estimated EVSI as described for scenario 1 using equation (17). In method 2, we implemented the 2-level Monte Carlo scheme outlined earlier. In an outer loop, we drew K samples from the Normal(0.6, 1/6) prior distribution for logit(Q). For each value logit(Q)(, we generated a sample of data x( from a Normal{logit(Q)(, 1/50} distribution. Conditional on each simulated trial data value x(, we then ran an inner loop of size J = 104. At each run of the inner loop, we sampled a vector of parameter values θ( from the posterior distribution p(θ|x() and evaluated the model net benefit equation NB(d, θ(). Finally, we calculated EVSI via equation (9). Because x( is only informative for the parameter Q, and Q is independent of all other model parameters, drawing from p(θ|x() reduced in this case to drawing from the posterior distribution p(Q|x() and the prior distributions for the remaining parameters. The posterior distribution p(Q|x() does not have a standard form, and we therefore implemented the inner loop in OpenBUGS.[25] At each inner-loop step, we discarded the first 1000 MCMC samples as a burn-in. In method 3, we implemented the GAM regression scheme outlined earlier. First, we generated a PSA sample of size K. We calculated the incremental net benefit for each PSA sample. Next, for each parameter vector θ( in the PSA sample, we generated a sample of data x( from a Normal{logit(Q)(, 1/50} distribution. Because the data are scalar, the summary statistic is T(x() = x(. We regressed the incremental net benefits on T(x(), extracted the model-fitted values, and estimated the EVSI via equation (14).

Data Collection Scenario 3: EVSI for the Treatment Effect Size (OR)

To reduce uncertainty about the treatment effect size parameter OR, we consider the value of undertaking a randomized controlled trial with n = 200 patients allocated to the new treatment, and n = 200 patients allocated to standard care. We assume that, conditional on P and P, where logit(P) = logit(P) + log(OR), the number of critical events is x ~ Binomial (P, 200) in the new treatment group and x ~ Binomial (P, 200) in the standard care group. Again, we replicated the single-loop method used in the case study in Ades et al.[7] The scheme for updating the model parameters conditional on the sampled data is rather more complex than in scenarios 1 and 2. We give a brief outline of the method here, and the reader is referred to the original study for full details. First, we drew k = 1, . . ., K samples from the Beta(15, 85) prior distribution for P, and k = 1, . . ., K samples from the Normal(–1.5, 1/3) prior distribution for log(OR). For each k, we calculated . Next, for each value , we generated a sample of data from a Binomial(, 200) distribution and, from each value , a sample of data from a Binomial(, 200) distribution. From and , we calculated the log odds ratio for simulated study k. This was assumed to follow a Normal distribution with expectation equal to log(OR)( and variance equal to . These assumptions result in conjugacy and therefore a Normal posterior distribution for log(OR) given the sampled data. The posterior expectation of log(OR) was added to the prior expectation for logit(P) to give a value of logit(P) that reflected the new knowledge of the treatment effect derived from the simulated trial. The posterior expectation of was derived from the expectation of using a Taylor series approximation. Finally, the expected net benefits were obtained by plugging into equations (15) and (16) the posterior expectations , along with the prior expected values for the remaining uncertain parameters, , , and . We then estimated EVSI as described for scenario 1 using equation (17). In method 2, we implemented the 2-level Monte Carlo scheme outlined earlier. In an outer loop, we generated samples of data as in method 1. Conditional on each simulated trial data vector, we ran an inner loop of size J = 104. Because we require that x( is informative only for the parameter OR and not for P, we used the inner-loop step to sample parameter values from its posterior distribution p(P|x() and values of P from its prior distribution. We evaluated the model net benefit equation NB(d, θ() at each inner loop run and calculated EVSI via equation (9). The posterior distribution p(P|x() does not have a standard form, and we therefore implemented the inner loop in OpenBUGS.[25] At each inner-loop step, we discarded the first 1000 MCMC samples as a burn-in. In method 3, we implemented the GAM regression scheme outlined earlier. First, we generated a PSA sample of size K. We calculated the incremental net benefit for each PSA sample. Next, for each parameter vector θ( in the PSA sample, we generated a sample of data comprising from a Binomial(, 200) distribution and from a Binomial(, 200) distribution. Data samples in this case are vector valued, , and we therefore calculated the sample log odds ratio statistic . We regressed the incremental net benefits on T(x(), extracted the model-fitted values, and estimated the EVSI via equation (14).

Results

Table 2 shows the EVSI values, standard errors, and timings for the 3 data collection scenarios calculated by the Ades et al.[7] method, the 2-level Monte Carlo/MCMC method, and the GAM regression method. For comparison, Ades et al. report values of $5550, $1880, and $3260 for scenarios 1, 2, and 3, respectively, using the single-loop method with a sample size of 105. Partial EVPI values for P, and OR (the parameters updated in scenarios 1, 2, and 3) are $6280, $2090, and $3890, respectively. Estimated EVSI Values and CPU Run Times for the Three Case Study Scenarios EVSI, expected value of sample information; GAM, generalized additive model. Partial EVPI values (an upper limit on EVSI) are $6280, $2090, and $3890 for scenarios 1, 2, and 3, respectively. There is good agreement between all 3 methods in scenarios 1 and 2. In scenario 3, the most precise EVSI estimates obtained using the MCMC and GAM methods are in agreement with each other and are approximately $200 lower than the most precise estimate obtained using the Ades et al.[7] method. The analytic method for computing the inner conditional expectation is exact for scenario 1 but not for scenarios 2 and 3. In scenarios 2 and 3, a Taylor series approximation is used to derive the expectation of a logit-transformed parameter from the expectation of the parameter itself, and in scenario 3, the log odds ratio obtained for each simulated study is assumed to be Normally distributed. These approximations are not required in either the 2-level method or the regression method. The assumption in the analytic method that the study log odds ratios are Normally distributed may in particular be problematic. The assumption is reasonable if the underlying probabilities are close to 0.5 but not when probabilities are close to 0 or 1. In our case study, the mean probability of a critical event is on the new treatment and on standard care. To explore the robustness of the assumption of Normality, we generated samples from X ~ Binom(0.044, 200) and X ~ Binom(0.15, 200) and, for each pair of samples, calculated the log odds ratio. A Normal QQ plot of the sampled log odds ratios in Figure 3 shows that the assumption of Normality does not hold.
Figure 3

Normal QQ plot for samples of a study log odds ratio with P = 0:044, P = 0:15, and n = n = 200.

Normal QQ plot for samples of a study log odds ratio with P = 0:044, P = 0:15, and n = n = 200.

Precision and Computational Efficiency

For comparisons within each method, the Monte Carlo standard error scales in proportion to K−1/2 (as expected), and the computation time scales approximately in proportion to K. For the same level of precision, the 2-level Monte Carlo method requires 104 (i.e., the inner loop size) times as many samples as does the Ades et al.[7] analytic method. This holds across all 3 scenarios. The 2-level method is roughly 5 orders of magnitude slower than the Ades et al. method for the same value of K, primarily due to the requirement for 104 model evaluations at each outer-loop run. For scenario 1, the standard errors obtained using the GAM method are approximately 20% larger than those of the analytic method, given the same number of model runs. Thus, to achieve the same precision, the number of runs would need to be increased by approximately 1.22 = 1.44 times (this was confirmed empirically; results not shown). For scenario 2, the standard errors obtained using the GAM method are approximately 4 times larger than those of the analytic method, given the same number of model runs. Thus, to achieve the same precision, the number of runs would need to be increased by approximately 16 times. For scenario 3, the standard errors obtained using the GAM method are approximately 60% larger than those of the analytic method, given the same number of model runs. Here, to achieve the same precision, the number of runs would need to be increased by approximately 2.6 times. The precision of the GAM estimate, as well as being related to the sample size, is related to the uncertainty in the model parameters that are not updated by the data. This uncertainty propagates through to the model output (the net benefit) in the PSA sample and causes the variability in the net benefit that is modeled by the error term in the regression. In the case of scenario 2, the GAM estimate requires relatively more samples (for some given precision) than does the GAM estimate for scenario 1 or 3. This suggests that in scenario 2, the uncertainty in the parameters not updated by data (i.e., P and log(OR)) is large. This is confirmed by the relatively high partial EVPI values for P and log(OR). In terms of computational speed, the GAM method is roughly an order of magnitude slower than the analytic method but roughly 4 orders faster than the 2-level method.

Discussion

Our key idea is that instead of estimating, for each decision option, the posterior expected net benefits via a computationally burdensome repeated (inner-loop) Monte Carlo step, we estimate the functional relationship between the posterior expected net benefit and the simulated data via a nonparametric regression.

Strengths and Limitations

The value of the nonparametric regression method over the 2-level Monte Carlo approach is 2-fold: it is straightforward to implement, requiring less detailed mathematical thinking on the part of the analyst for any particular application, and is several orders of magnitude faster for any given precision. Due to the method’s ease of use and the fact that it only requires the PSA sample and not the model itself, we would suggest that the regression method is applicable in even the simplest of modeling contexts. In common with other established methods for computing EVSI, we must be able to generate sample data sets, conditional on samples from the prior distribution of the model parameters. For complex study designs, this may not be straightforward. For our method, we must also be able to summarize the sampled data in either a scalar, or low-dimensional, summary statistic. Again, for complex study designs, this may not always be easy.

How This Fits with Existing Literature

One option for computing EVSI is to assume that the incremental net benefit is Normally distributed with parameters that are known functions of study sample size. Under this assumption, EVSI can be calculated using fast analytical methods. This “parametric” approach is most appropriate in settings in which cost-effectiveness analysis is undertaken alongside a single 2-arm randomized controlled trial (RCT). In this setting, the mean incremental net benefit is derived directly from individual-level costs and effects, and the central limit theorem can be used to justify the assumption of Normality. The method has been explored in theoretical investigations[9-13] and applied in real clinical decision problems.[30] However, the approach is not straightforward for decision problems with more than 2 options and may not be appropriate when a more complex decision-analytic model has been used to estimate incremental net benefit. A nonlinear cost-effectiveness model with non-Normal input parameters may generate notably non-Normal incremental net benefits. It may be difficult to predict the relationship between the size of a proposed study that informs some particular subset of parameters and the net benefits of a range of competing decision options. Deriving the parameters of the Normal distribution(s) that the parametric approach requires may therefore be difficult. The nonparametric regression approach that we propose has some similarities to the model emulation method proposed in Oakley,[31] and in one sense, the GAM model can be viewed as an emulator for the posterior expectation of the net benefit, conditional on the data. The important difference between the 2 approaches is that in Oakley, the net benefit function itself is emulated. Emulating the net benefit function allows for the rapid evaluation of a slow economic model, but it does not address the problem of how to sample from a difficult posterior distribution of the parameters conditional on the data. Our method also has some similarities to the spline-based approach for computing partial EVPI that has been proposed by Madan et al.[21] Here, a spline is used to approximate the conditional expectation of some predefined subfunction of the net benefit function, given some sampled value of a parameter of interest. Although the method by Madan et al is shown to perform well, it requires algebraic manipulation of the net benefit function to identify the appropriate subfunctions for the spline approximation. This may be difficult in complex models.

Implications for Practice and for Research

At present, commissioners and funders of research are not using EVSI methods as part of their day-to-day decision-making processes (although there are exceptions to this[30,32,33]). However, it is becoming more common for early economic evaluation models to be requested by research funders to establish the potential benefits of proposed new interventions. One barrier to implementation of EVSI at this stage has been the complexity of calculation, and we hope that our method can help to remove this barrier. We do recognize, though, that there are other important barriers to the widespread use of EVSI, including a lack of understanding of value of information methods, a mistrust of mathematical models, and a requirement for trialists to work within a frequentist hypothesis test-based framework. Throughout the article, we have assumed that the decision maker’s utility is equal to net benefit and that the decision problem is to maximize net benefit over a set of discrete treatment options. This is the typical decision problem faced by government agencies such as the National Institute for Health and Care Excellence (NICE), but the type of decision problem faced by the pharmaceutical industry is somewhat different. Here, utility is profit, and under value-based pricing, the decision problem is to choose price to maximize profit, subject to the constraint that additional health benefits do not cost more than the funder’s willingness to pay. EVSI from an industry perspective has been explored by Willan[34] and by Willan and Eckermann,[35] as well as more recently within a value-based pricing context by Breeze and Brennan.[36] Our nonparametric regression method for computing EVSI will apply equally well in this context. Further research could extend to testing the nonparametric regression method in other cost-effectiveness models, including Markov cohort models and more complex patient-level models. An important feature of our method is that all variation in the net benefit function that is not due to the data is taken up by the error term in the regression analysis and “averaged out.” Thus, any variation in the net benefit that arises due to poor convergence of a patient-level model is also averaged out in the regression.[24] This means that, to calculate EVSI for a patient-level model (in which patients do not interact), only a single patient needs to be “run” through the model at each evaluation of the PSA. We look forward to more research in this area.
Table 2

Estimated EVSI Values and CPU Run Times for the Three Case Study Scenarios

Sample SizeEVSI (SE), $
Mean CPU Time, s
Outer (K)Inner (J)TotalScenario 1Scenario 2Scenario 3
Ades et al[7] (2004) method
1041045660 (107)1955 (38)3121 (79)0.1
1051055543 (34)1872 (12)3279 (26)0.2
1061065565 (11)1884 (3.7)3245 (8.0)1.2
Two-level Monte Carlo method
1041041085464 (105)1871 (37)2967 (80)4456
1051041095562 (34)1892 (12)3049 (26)43,303
10610410105569 (11)1886 (3.7)3031 (8.1)424,686
GAM regression method
1041045334 (130)2047 (163)3117 (137)0.1
1051055534 (42)1846 (51)3020 (41)0.7
1061065580 (13)1861 (16)3035 (13)8.1

EVSI, expected value of sample information; GAM, generalized additive model. Partial EVPI values (an upper limit on EVSI) are $6280, $2090, and $3890 for scenarios 1, 2, and 3, respectively.

  32 in total

1.  An efficient method for computing single-parameter partial expected value of perfect information.

Authors:  Mark Strong; Jeremy E Oakley
Journal:  Med Decis Making       Date:  2012-12-28       Impact factor: 2.583

2.  Need for speed: an efficient algorithm for calculation of single-parameter expected value of partial perfect information.

Authors:  Mohsen Sadatsafavi; Nick Bansback; Zafar Zafari; Mehdi Najafzadeh; Carlo Marra
Journal:  Value Health       Date:  2013-01-26       Impact factor: 5.725

Review 3.  Optimal global value of information trials: better aligning manufacturer and decision maker interests and enabling feasible risk sharing.

Authors:  Simon Eckermann; Andrew R Willan
Journal:  Pharmacoeconomics       Date:  2013-05       Impact factor: 4.981

4.  Expected value of sample information for Weibull survival data.

Authors:  Alan Brennan; Samer A Kharroubi
Journal:  Health Econ       Date:  2007-11       Impact factor: 3.046

5.  Expected value of sample information calculations in medical decision modeling.

Authors:  A E Ades; G Lu; K Claxton
Journal:  Med Decis Making       Date:  2004 Mar-Apr       Impact factor: 2.583

6.  Valuing Trial Designs from a Pharmaceutical Perspective Using Value-Based Pricing.

Authors:  Penny Breeze; Alan Brennan
Journal:  Health Econ       Date:  2014-09-09       Impact factor: 3.046

7.  Expected value of sample information for multi-arm cluster randomized trials with binary outcomes.

Authors:  Nicky J Welton; Jason J Madan; Deborah M Caldwell; Tim J Peters; Anthony E Ades
Journal:  Med Decis Making       Date:  2013-10-01       Impact factor: 2.583

Review 8.  What is the clinical effectiveness and cost-effectiveness of cytisine compared with varenicline for smoking cessation? A systematic review and economic evaluation.

Authors:  Joanna Leaviss; William Sullivan; Shijie Ren; Emma Everson-Hock; Matt Stevenson; John W Stevens; Mark Strong; Anna Cantrell
Journal:  Health Technol Assess       Date:  2014-05       Impact factor: 4.014

9.  Are value of information methods ready for prime time? An application to alternative treatment strategies for NSTEMI patients.

Authors:  Seamus Kent; Andrew Briggs; Simon Eckermann; Colin Berry
Journal:  Int J Technol Assess Health Care       Date:  2013-10       Impact factor: 2.188

10.  Strategies for efficient computation of the expected value of partial perfect information.

Authors:  Jason Madan; Anthony E Ades; Malcolm Price; Kathryn Maitland; Julie Jemutai; Paul Revill; Nicky J Welton
Journal:  Med Decis Making       Date:  2014-01-21       Impact factor: 2.583

View more
  32 in total

1.  Using Evidence from Randomised Controlled Trials in Economic Models: What Information is Relevant and is There a Minimum Amount of Sample Data Required to Make Decisions?

Authors:  John W Stevens
Journal:  Pharmacoeconomics       Date:  2018-10       Impact factor: 4.981

2.  The Curve of Optimal Sample Size (COSS): A Graphical Representation of the Optimal Sample Size from a Value of Information Analysis.

Authors:  Eric Jutkowitz; Fernando Alarid-Escudero; Karen M Kuntz; Hawre Jalal
Journal:  Pharmacoeconomics       Date:  2019-07       Impact factor: 4.981

3.  Cost-Effectiveness of Treatments for the Management of Bone Metastases: A Systematic Literature Review.

Authors:  Lazaros Andronis; Ilias Goranitis; Sue Bayliss; Rui Duarte
Journal:  Pharmacoeconomics       Date:  2018-03       Impact factor: 4.981

4.  Metamodeling for Policy Simulations with Multivariate Outcomes.

Authors:  Huaiyang Zhong; Margaret L Brandeau; Golnaz Eftekhari Yazdi; Jianing Wang; Shayla Nolen; Liesl Hagan; William W Thompson; Sabrina A Assoumou; Benjamin P Linas; Joshua A Salomon
Journal:  Med Decis Making       Date:  2022-06-23       Impact factor: 2.749

5.  Personalization of Medical Treatment Decisions: Simplifying Complex Models while Maintaining Patient Health Outcomes.

Authors:  Christopher Weyant; Margaret L Brandeau
Journal:  Med Decis Making       Date:  2021-08-20       Impact factor: 2.749

6.  Calculating the Expected Value of Sample Information in Practice: Considerations from 3 Case Studies.

Authors:  Anna Heath; Natalia Kunst; Christopher Jackson; Mark Strong; Fernando Alarid-Escudero; Jeremy D Goldhaber-Fiebert; Gianluca Baio; Nicolas A Menzies; Hawre Jalal
Journal:  Med Decis Making       Date:  2020-04-16       Impact factor: 2.583

7.  Implementing Generalized Additive Models to Estimate the Expected Value of Sample Information in a Microsimulation Model: Results of Three Case Studies.

Authors:  Dustin J Rabideau; Pamela P Pei; Rochelle P Walensky; Amy Zheng; Robert A Parker
Journal:  Med Decis Making       Date:  2017-11-09       Impact factor: 2.583

8.  Probabilistic threshold analysis by pairwise stochastic approximation for decision-making under uncertainty.

Authors:  Takashi Goda; Yuki Yamada
Journal:  Sci Rep       Date:  2021-10-04       Impact factor: 4.379

9.  Computing the Expected Value of Sample Information Efficiently: Practical Guidance and Recommendations for Four Model-Based Methods.

Authors:  Natalia Kunst; Edward C F Wilson; David Glynn; Fernando Alarid-Escudero; Gianluca Baio; Alan Brennan; Michael Fairley; Jeremy D Goldhaber-Fiebert; Chris Jackson; Hawre Jalal; Nicolas A Menzies; Mark Strong; Howard Thom; Anna Heath
Journal:  Value Health       Date:  2020-05-27       Impact factor: 5.725

10.  Antimicrobial-impregnated central venous catheters for preventing neonatal bloodstream infection: the PREVAIL RCT.

Authors:  Ruth Gilbert; Michaela Brown; Rita Faria; Caroline Fraser; Chloe Donohue; Naomi Rainford; Alessandro Grosso; Ajay K Sinha; Jon Dorling; Jim Gray; Berit Muller-Pebody; Katie Harron; Tracy Moitt; William McGuire; Laura Bojke; Carrol Gamble; Sam J Oddie
Journal:  Health Technol Assess       Date:  2020-11       Impact factor: 4.106

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.