| Literature DB >> 29696154 |
Thuva Vanniyasingam1,2, Caitlin Daly1, Xuejing Jin1, Yuan Zhang1, Gary Foster1,2, Charles Cunningham3, Lehana Thabane1,2,4,5,6.
Abstract
OBJECTIVES: This study reviews simulation studies of discrete choice experiments to determine (i) how survey design features affect statistical efficiency, (ii) and to appraise their reporting quality. OUTCOMES: Statistical efficiency was measured using relative design (D-) efficiency, D-optimality, or D-error.Entities:
Keywords: Discrete choice experiment; Relative D-efficiency; Relative D-error; Statistical efficiency; Systematic survey
Year: 2018 PMID: 29696154 PMCID: PMC5898574 DOI: 10.1016/j.conctc.2018.01.002
Source DB: PubMed Journal: Contemp Clin Trials Commun ISSN: 2451-8654
Studies investigating the number of choice tasks, attributes, and attribute levels
| Author, Year | Outcome of interest | Method to create design | Design setting | Distribution of Priors of parameter estimates | Choice sets | Alternatives | Attributes | Attribute levels | Results |
|---|---|---|---|---|---|---|---|---|---|
| # Choice tasks | |||||||||
| Vanniyasingam, 2016 [ | Relative | . | . | no priors | 2–20 | 2–5 | 2–20 | 2–5 | Generally, as the number of choice tasks increases, relative D-efficiency increases |
| # Attributes | |||||||||
| Vanniyasingam, 2016 [ | Relative | . | . | no priors | 2–20 | 2–5 | 2–20 | 2–5 | Generally, increasing# attributes, decreases relative D-efficiency (not monotonically) designs with a small# of alternatives and large number of attributes could not be created. |
| # Attribute levels | |||||||||
| Vanniyasingam, 2016 [ | Relative | . | . | no priors | 2–20 | 2–5 | 2–20 | 2–5 | Generally, increasing# attribute levels, decreases relative D-efficiency designs yield higher D-efficiency measures when the# attribute levels match the number of alternatives Generally, binary attributes perform best across all other designs |
| Graβhoff, 2013 [ | Efficiency | . | . | β1 = 0, β2 = 1 | . | 3 | 1–7 | Unrestricted quantitative (continuous) and qualitative (binary) attributes | Design-optimality was achieved when two alternatives were identical or differed only in the (unrestricted) quantitative variable, while the third alternative varies in all of the qualitative components. |
Studies investigating the number of alternatives on statistical efficiency
| # Alternatives | |||||||||
|---|---|---|---|---|---|---|---|---|---|
| Author, Year | Outcome of interest | Method to create design | Design setting | Distribution of Priors of parameter estimates | Choice sets | Alternatives | Attributes | Attribute levels | Results |
| Vermeluen, 2010 [ | Db-error | 1: Best choice experiment | 1: Partial best choice experiment | Parameter estimates follow a normal distribution with mean priors | 9 | 4, 5, 6 | 5 | 3 three-level, 2 two-level attributes | |
| Vanniyasingam, 2016 [ | Relative D-efficiency | Random allocation | . | No priors | 2–20 | 2–5 | 2–20 | 2–5 | |
Studies investigating the incorporation of choice behaviour on statistical efficiency
| Author, Year | Outcome of interest | Method to create design | Design setting | Sample size | Choice sets | Altern-atives | Attri- | Attribute levels | Results |
|---|---|---|---|---|---|---|---|---|---|
| Crabbe, 2012 [ | Local D-error | Individually adapted sequential Bayesian designs (IASB) with covariates incorporated IASB designs, no covariates single nearly orthogonal designs, no covariates | Choice behaviour is influenced by 2 covariates Choice behaviour is NOT influenced by 2 (irrelevant) covariates | 25, 250 | 16 | 3 | 3 | 3 | Across all design settings and sample sizes: Despite IASB designs incorporating two IASB designs with two relevant covariates perform better (in terms of D-efficiency) in comparison to IASB designs with to irrelevant covariates, holding everything else constant. |
| Donkers, 2003 [ | Average percentage change in D-error | Design incorporates the proportion of the population selecting y = 1, which varies from 2.5%, 5%, 10%, 15%, and 50% of the population. Results of D-error compared to random sampling from population. | . | Sample selection is dependent on proportion that selects Y = 1 | 2 | 2 | 1 binary, 1 continuous | As the proportion of the population selecting Y = 1 increases from 2.5% to 50%, D-efficiency improves. As more individuals select 1, the magnitude of the reduction in D-error decreases (in comparison to when a random sample is used). The highest reduction in D-error (or improvement in D-efficiency) is when only 2.5% of the population selects y = 1. Above results are consistent when binary attribute (x = 1) is distributed 10% or 50% of the time within the DCE. | |
| Donkers, 2003 [ | Average percentage change in D-error | Design incorporates the proportion of the population selecting y = 1, which varies from 2.5%, 5%, 10%, 15%, and 50% of the population. Results of D-error compared to random sampling from population. | . | Sample selection is dependent on: y only y and x x only | 2 | 2 | 1 binary, 1 continuous; | Type of sample selection (y only, y and x, x only) Designs with sample selection on both y and x yields higher statistical efficiency than designs with sample selection on y only or x only, where y is the outcome, and x is an attribute. |
Studies investigating Bayesian priors on statistical efficiency
| Author, Year | Yu, 2009 [ | Vermeulen, 2010 [ | Bliemer, 2010 [ |
|---|---|---|---|
| Outcome of interest | Relative local D-efficiency | Db-error | D-error and percentage change in D-error |
| Describe the scenario | 8 different designs, each compared within 5 different parameter spaces/design settings. | Comparing four designs within 3 settings for designs varying in alternatives and variance priors of parameters | Misspecification of prior parameter values |
| Method to create design | Models 1–3: Mixed logit semi-Bayesian d-optimal design | Model 1: Best choice experiment | Model 1: MNL |
| Design setting | Parameters were drawn from a normal distribution: | Setting 1: Partial best choice experiment | Setting 1: MNL model |
| Heterogeneity prior | Model 1 = 1.5 × 18; Model 2 = 18; | . | . |
| Distribution of Priors of parameter estimates | Model 1–3: Normal distribution with fixed mean, covariance I8 | Parameter estimates follow a normal distribution with mean priors: | Settings 1–3: Assumed true value of parameters: β0 = −0.5, |
| Choice sets | 12 | 9 | 12 |
| Alternatives | 3 | 4, 5, 6 | 2 |
| Attributes | 4 | 5 | 3 |
| Attribute levels | 3 | 3 three-level, 2 two-level attributes | 2 three-level attributes, 1 four-level attribute |
| Results | Across all 5 design settings: Mixed logit model designs performed substantially better than designs that ignored respondent heterogeneity Comparing Semi-Bayesian designs (Models 1–3): Overspecifying the heterogeneity prior (Model 1) does not have too large of a negative impact on efficiency Underspecifying the heterogeneity prior (Model 3) has a greater loss in efficiency in comparison to overspecifying it (Model 1) Results remain consistent across other design setting such as: 2 × 3 × 4/2/24 and 2 × 2 × 3/3/12 | Parameter priors: | D-errors of designs with misspecified priors were higher than designs with correctly specified priors (from scenario. |
Studies investigating methods to create DCE designs on statistical efficiency
| Author, Year | Vermeulen, 2011 [ | Bliemer, 2010 [ | Vermeulen, 2008 [ | Vermeulen, 2010 [ | Vermeulen, 2010 [ |
|---|---|---|---|---|---|
| Outcome | Db-error | D-error | db-error | Db-error | Relative D-efficiencies |
| Describe the scenario | Comparing different designs to create DCEs for 2 settings: full rank- and partial rank-order choice-based conjoint experiments | Comparing three types of designs against each other and an orthogonal design. | Comparing different designs to create DCEs in 2 settings: a presence and absence of 'no-choice' alternative in DCEs | Comparing different designs to create DCEs in 3 settings and with varying alternatives and variance priors | Comparing semi-Bayesian D-optimal best-worst design with 6 benchmark designs |
| Choice sets | 9 | 9, 12 | 16 | 9 | 9 |
| Alternatives | 4 | 2,3 | 2 and "no choice" alternative | 4,5, 6 | 4 |
| Attributes | 5 | 3,4 | 3 | 5 | 5 |
| Attribute levels | 3322 | 3241 | 3221 | 3222 | 3222 |
| Method to create design | Design: 1. Bayesian D-optimal ranking 2. D-optimal choice 3. Balanced overlap 4. Near-orthogonal 5. Random | Design: 1. MNL design 2.Cross-sectional mixed logit design (heterogeneity prior = 0), Panel mixed logit design Priors: fixed parameters, priors equal to the mean | Design: 1. MNL model 2. Extended no-choice MNL 3. Nested no-choice MNL 4. Model-robust | Design: 1. Best choice 2. Partial rank-order conjoint 3: Best-worst choice 4: Orthogonal | Design: |
| Design setting | Setting: 1. Full rank-order choice-based conjoint experiments 2. Partial rank-order choice-based conjoint experiments | Setting: 1. MNL 2. Cross-sectional mixed logit 3. Panel mixed logit model 4. Orthogonal (within alternatives) design | Setting: 1. Extended no-choice multinomial logit model 2: Nested no-choice multinomial logit model | Setting: 1. Partial best choice experiment 2. Rank-order conjoint experiment 3. Best-worst experiment | Setting: |
| Priors | Settings 1–3: Assumed priors correspond to true parameter values: | Priors for each setting: | Coefficients come from an 8-dimensional normal distribution with | ||
| Results | D-opt rank > D-opt. choice > Near-orthogonal > Random > Balanced overlap | Models estimated using designs specifically generated for that model outperform designs generated for different mode forms. | Models estimated using designs specifically generated for that model outperform designs generated for different mode forms. | 1)Models estimated using designs specifically generated for that model outperform designs generated for different mode forms | 1. Design 1 > 2, 4, 5, 6, 7 |
Comment: The greater than sign “>” indicates which method performed better than another method in terms of statistical efficiency.
Reporting items of simulations studies
| Author, Year | Protocol | Primary outcome | Clear aim | Number of failures | Software | Random number generator or starting seed | Rationale for creating designs | Methods for creating designs | Scenarios: Total number of designs | Scenarios: Range of design characteristics explored | Method to evaluate each scenario | Distribution used to simulate data* |
|---|---|---|---|---|---|---|---|---|---|---|---|---|
| Vermeulen, 2011 [ | 0 | 1 | 1 | 0 | 1 | 0 | 1 | 1 | 1 | 1 | 1 | 0 |
| Yu, 2009 [ | 0 | 1 | 1 | 0 | 0, 1 | 0 | 1 | 1 | 1 | 1 | 1 | 1 |
| Bliemer, 2010 [ | 0 | 1 | 0 | 0 | 1 | 0 | 1 | 1 | 1 | 1 | 1 | 1 |
| Crabbe, 2012 [ | 0 | 1 | 1 | 0 | 1 | 0 | 1 | 1 | 1 | 1 | 1 | 0 |
| Vermeulen, 2010 [ | 0 | 1 | 1 | 0 | 1 | 0 | 1 | 1 | 1 | 1 | 1 | 0 |
| Vermeulen, 2008 [ | 0 | 1 | 1 | 0 | 0 | 0 | 1 | 1 | 1 | 1 | 1 | 0 |
| Vanniyasingam, 2016 [ | 0 | 1 | 1 | 1 | 1 | 0 | 1 | 1 | 1 | 1 | 1 | 1 |
| Graβhoff, 2013 [ | 0 | 1 | 0 | 0 | 0 | 0 | 1 | 1 | 1 | 1 | 1 | 1 |
| Donkers, 2003 [ | 0 | 1 | 1 | 0 | 0 | 0 | 1 | 1 | 1 | 1 | 1 | 0 |
Comment: 1 = reported; 0 = unclear/not reported for each column.
*1 = the chosen design characteristics are motivated by real-world scenario (previous literature referenced, etc) OR by other simulation study scenarios, 0 = not motivated by other studies.