Literature DB >> 27660322

Prescriber preferences for behavioural economics interventions to improve treatment of acute respiratory infections: a discrete choice experiment.

Cynthia L Gong1, Joel W Hay1, Daniella Meeker2, Jason N Doctor1.   

Abstract

OBJECTIVE: To elicit prescribers' preferences for behavioural economics interventions designed to reduce inappropriate antibiotic prescribing, and compare these to actual behaviour.
DESIGN: Discrete choice experiment (DCE).
SETTING: 47 primary care centres in Boston and Los Angeles. PARTICIPANTS: 234 primary care providers, with an average 20 years of practice. MAIN OUTCOMES AND MEASURES: Results of a behavioural economic intervention trial were compared to prescribers' stated preferences for the same interventions relative to monetary and time rewards for improved prescribing outcomes. In the randomised controlled trial (RCT) component, the 3 computerised prescription order entry-triggered interventions studied included: Suggested Alternatives (SA), an alert that populated non-antibiotic treatment options if an inappropriate antibiotic was prescribed; Accountable Justifications (JA), which prompted the prescriber to enter a justification for an inappropriately prescribed antibiotic that would then be documented in the patient's chart; and Peer Comparison (PC), an email periodically sent to each prescriber comparing his/her antibiotic prescribing rate with those who had the lowest rates of inappropriate antibiotic prescribing. A DCE study component was administered to determine whether prescribers felt SA, JA, PC, pay-for-performance or additional clinic time would most effectively reduce their inappropriate antibiotic prescribing. Willingness-to-pay (WTP) was calculated for each intervention.
RESULTS: In the RCT, PC and JA were found to be the most effective interventions to reduce inappropriate antibiotic prescribing, whereas SA was not significantly different from controls. In the DCE however, regardless of treatment intervention received during the RCT, prescribers overwhelmingly preferred SA, followed by PC, then JA. WTP estimates indicated that each intervention would be significantly cheaper to implement than pay-for-performance incentives of $200/month.
CONCLUSIONS: Prescribing behaviour and stated preferences are not concordant, suggesting that relying on stated preferences alone to inform intervention design may eliminate effective interventions. TRIAL REGISTRATION NUMBER: NCT01454947; Results. Published by the BMJ Publishing Group Limited. For permission to use (where not already granted under a licence) please go to http://www.bmj.com/company/products-services/rights-and-licensing/

Entities:  

Keywords:  antibiotic prescribing; conjoint analysis; discrete choice; revealed preference; stated preference

Year:  2016        PMID: 27660322      PMCID: PMC5051402          DOI: 10.1136/bmjopen-2016-012739

Source DB:  PubMed          Journal:  BMJ Open        ISSN: 2044-6055            Impact factor:   2.692


In this discrete choice experiment, prescribers were asked about which interventions they felt would be most effective in reducing their rates of inappropriate antibiotic prescriptions for acute respiratory infections; such information can be valuable when developing clinical quality improvement interventions. This is one of few studies in the healthcare literature that not only elicits prescriber preferences using a discrete choice experiment but also compares these stated preferences to actual prescribing behaviour as observed in a randomised controlled trial. Results indicate that stated and revealed preferences are not concordant, suggesting that clinical quality improvement programmes should not rely solely on clinician input, but instead combine expert-driven and clinician-driven approaches. The discrete choice experiment may not necessarily capture true preference, but instead be a reflection of convenience and ease of use of a specific clinical quality improvement intervention. Stated preferences for this group of healthcare providers may not necessarily reflect those of a national sample of providers.

Introduction

According to the Centers for Disease Control and Prevention, up to 50% of antibiotics are not optimally prescribed, leading to an estimated 2 million illnesses and 23 000 deaths due to antibiotic resistance alone.1 Most antibiotics in the USA are prescribed for acute respiratory tract infections (ARIs), and about half of these prescriptions are issued to patients with non-bacterial diagnoses.2 3 Despite attempts to curb inappropriate antibiotic prescribing through interventions such as physician and patient education, electronic clinical decision support and financial incentives, these have only resulted in modest reductions in antibiotic prescribing rates for non-bacterial ARIs.4 Clinical quality improvement interventions frequently rely upon changing clinicians' practice behaviours, such as reducing orders for inappropriate treatments or diagnostic tests. The consensus-recommended best practices in design and implementation of quality improvement interventions include some component of ‘local participatory’ approaches, that is, working with frontline staff, including the target population.5–7 This engagement may directly or indirectly influence the intervention design, but it is unclear if this practice yields the optimal design. For example, theory and survey results suggest that physicians are likely to indicate preferences for direct financial incentives (bonuses).8–10 However, studies of effectiveness have shown mixed results of direct incentives in practice.10 11 This raises questions regarding whether stated preferences for intervention features elicited in a participatory process should be incorporated into design. One approach to changing prescribing behaviour applies ideas from the behavioural sciences, using social cues and subtle changes in the clinic environment to influence clinical decision-making.12 13 In fact, the UK government implements behavioural economics ‘nudges’ into its health services, and a recent study in the UK found that social norm feedback from England's Chief Medical Officer was highly effective in reducing inappropriate antibiotic prescribing at a low cost.14–16 In the USA, the ‘Use of behavioral economics and social psychology to improve treatment of acute respiratory infections (BEARI)’ study, a multisite cluster randomised controlled trial (RCT), applied behavioural techniques and assessed the impact of various behavioural interventions on the rates of inappropriate antibiotic prescribing in various practices in Illinois, Massachusetts and Southern California.17 18 As part of the BEARI study, a discrete choice experiment (DCE) was conducted to elicit prescribers' stated preferences to evaluate prescriber preferences for one intervention compared to another. The use of discrete methods to elicit preferences for various programmes and interventions in healthcare has significantly increased in recent years.19–22 These methods have been used in the context of health system reforms or quality improvement programmes, and healthcare policies, programmes, services, incentives and interventions.9 20 While they provide valuable feedback regarding factors that should be considered for a given programme or intervention, few studies in healthcare have evaluated how stated preferences compare to real-life behaviour. Those who have assessed external validity have generally found that stated preferences are consistent with actual decision-making behaviour on an aggregate level, while individual-level concordance is limited.23 24 Previous studies evaluating an individual's decisions for vaccination or disease screening indicated a positive predictive value for DCEs of 85%, but that the negative predictive value was only 26%; however, the authors noted that the majority of people opt for preventive healthcare in these situations, yielding the overall predictive values of the DCE.21 25 26 Furthermore, only one study has specifically considered physician decision-making—the ultimate driver of quality and cost of care.27 Based on these studies, DCE-elicited responses may generally reflect real-life behaviours. However, few studies have tested behaviour in a RCT and compared the results to DCE-elicited preferences for the same group of individuals (particularly physicians) to validate the DCE responses.28 Thus, the objective of this study was to elicit prescriber preferences for different behavioural economics interventions to reduce antibiotic prescribing, and compare these to actual prescribing behaviour as revealed in the BEARI study.

Methods

Design

Details for the BEARI study have been described elsewhere.17 Briefly, this was a multicentre trial conducted in Illinois, Massachusetts and Southern California to determine whether behavioural economic interventions could influence prescriber behaviour. The interventions implemented in the BEARI study included: (1) Suggested Alternatives (SA), which used computerised clinical decision support to suggest over-the-counter and prescription non-antibiotic treatment choices to clinicians prescribing an antibiotic for an ARI; (2) Accountable Justifications (JA), which prompted clinicians to enter an explicit justification when prescribing an antibiotic for an ARI that would then appear in the patient's electronic health record; and (3) Peer Comparison (PC), which was an email sent periodically to each prescriber that compared his/her rate of inappropriate antibiotic prescribing relative to top-performing peers.17 All clinicians received an education module reviewing ARI diagnosis and treatment guidelines, and each was assigned to 0, 1, 2 or all 3 interventions. Interventions were in place for 18 months at each practice site. Upon completion of the study, all prescribers were asked to complete a computerised exit survey, which assessed prescribers' level of satisfaction with each of the interventions, and included the DCE. A DCE is used to determine an individual's preferences for specific alternatives given a specific scenario, rather than observing an individual's behaviour in real markets to determine his/her revealed preferences.29 The individual is presented with a pair of choices with varying levels of attributes and is asked to choose the preferred alternative based on the attributes presented (refer to Supplementary materials for details). In the DCE, participants were presented with the following scenario, followed by a set of 10 choice pairs from which prescribers had to choose a preferred alternative that they felt would most likely reduce their ARI antibiotic prescribing: Suppose that your Health Care Organization implements various measures to reduce overprescribing of antibiotics for acute respiratory infections (e.g. Acute bronchitis, Acute Pharyngitis, Sinusitis, etc.). Assume implementation with an electronic health record system with electronic prescribing options and physician monitoring that can be automatically enabled or disabled. You will be provided with a short patient handout that discusses antibiotic overprescribing for acute respiratory infections. The five attributes included for each alternative were JA, SA, PC, pay-for-performance ($100, $200, or $0) and additional time spent with the patient (5 or 0 min) (table 1).
Table 1

Attributes and descriptions

AttributeDescription
SAIf you enter a target ARI diagnosis into the patient electronic health record, the EHR will prominently display a list of appropriate non-antibiotic prescription and non-prescription treatment alternatives to antibiotics
JAIf you order an antibiotic for a patient with an ARI, a new EHR screen will pop up that asks you to enter a short written justification for why the antibiotic prescription was necessary. What you write in this screen becomes part of the patient's permanent medical record, and therefore visible to other providers
PCEach other week, you and your peers will receive in your work mailboxes (or email) an updated ranking comparing your antibiotic prescribing rate to the top-performing (ie, ‘best’) decile of your clinical peers
P4PIf you are able to reduce your rate of antibiotic prescribing for all of your patients with ARI to the lowest 10th centile of your peers, you will receive a predetermined monthly reward payment of either $100 or $200
ATYour productivity management record system will change the RVU/visit time allowed for each patient with ARI to increase by 5 min per visit as you respond to the patient's concerns regarding the alternative treatments for their ARI diagnosis. If you achieve the antibiotic prescription reduction goal of reducing your antibiotic prescriptions by 50%, this 5 min per visit time will be preserved

ARI, acute respiratory tract infection; AT, additional time; EHR, Electronic Health Record; JA, Accountable Justifications; P4P, pay-for-performance; PC, Peer Comparison; RVU, relative value units; SA, Suggested Alternatives.

Attributes and descriptions ARI, acute respiratory tract infection; AT, additional time; EHR, Electronic Health Record; JA, Accountable Justifications; P4P, pay-for-performance; PC, Peer Comparison; RVU, relative value units; SA, Suggested Alternatives. These attributes and their levels were chosen based on interviews with prescriber focus groups in Los Angeles. Each attribute was toggled on or off in the choice pairs presented to respondents (figure 1). Respondents were asked to evaluate a total of 10 choice pairs which were assigned using a fractional factorial design with orthogonality of main effects.30
Figure 1

DCE treatment scenarios. DCE, discrete choice experiment.

DCE treatment scenarios. DCE, discrete choice experiment. A pilot survey was administered to a small number of prescribers in Los Angeles and Boston who were not involved in the BEARI study and thus not exposed to any of the interventions, and to a subset of prescribers in Chicago who did receive exposure to the BEARI interventions. The DCE was then administered as part of the exit survey for all prescribers involved in the BEARI study.

Data analysis

Baseline provider characteristics were collected at the start of the study, and information about the clinic environment and patient population mix were collected during the exit survey. Descriptive statistics were performed on these data. Preference data from the DCE were analysed using multinomial logit and mixed logit models drawn 500 times to obtain a robust output. In addition, the models were run on the subset of responders who completed 6 or more of the 10 choice tasks, and the results compared to those of the entire sample. The impact of intervention assignment on stated preference was evaluated by including intervention assignment as an explanatory variable in the mixed logit model, and interacting intervention assignment on each programme. Willingness-to-pay (WTP) for each intervention type was calculated by taking the ratio of the estimated coefficient for a given attribute to the cost coefficient, for example: . This calculation was performed for each intervention to quantify the monetary value of each attribute included in the DCE.31 All data were analysed in Stata V.12.0 (StataCorp. Stata Statistical Software: Release 12. StataCorp LP: College Station, Texas, 2011).

Results

Details for the outcomes of the BEARI study have been described elsewhere.18 Briefly, the BEARI study found that the most significant reduction in the rates of inappropriate antibiotic prescribing was achieved with JA and PC, when compared to the education control. SA yielded a non-significant reduction in inappropriate antibiotic prescribing rates.18 These results suggest that PC and JA may be the most effective interventions to influence prescribing. Of the 253 prescribers recruited in the BEARI RCT study, n=234 responses were received for the DCE survey, although only 157 provided responses to all 10 treatment scenarios. Of the responders, 26 were exposed to all three BEARI interventions, 35 to SA and JA, 32 to SA and PC, and 23 to JA and PC in the BEARI RCT. Prescribers were primarily medical doctors with an average of 20 years of clinical practice, with more men in the control group (table 2).
Table 2

Baseline demographics

Total (n=234)Control (n=24)Treatment (n=210)
CharacteristicsNumber (%)Number (%)Number (%)
Age (years, mean)48.547.848.6
Gender
 Female137 (58.5%)12 (50.0%)125 (59.5%)
 Male81 (34.6%)11 (45.8%)70 (33.3%)
 No response16 (6.8%)1 (4.2%)15 (7.1%)
Years of practice (mean)19.919.021.2
Professional credentials
 Medical doctor174 (74.4%)18 (75.0%)156 (74.3%)
 Nurse practitioner24 (10.3%)2 (8.3%)22 (10.5%)
 Physician assistant13 (5.6%)2 (8.3%)11 (5.2%)
 Doctor of osteopathy6 (2.6%)1 (4.2%)5 (2.4%)
 No response17 (7.3%)1 (4.2%)16 (7.6%)
Specialty
 Internal medicine126 (53.5%)14 (58.3%)112 (53.3%)
 Paediatrics23 (9.8%)2 (8.3%)21 (10%)
 Family/general practice28 (12.0%)3 (12.5%)25 (11.9%)
 Geriatrics3 (1.3%)03 (1.3%)
 Rheumatology1 (0.43%)01 (0.43%)
 Preventive medicine1 (0.43%)01 (0.43%)
 Infectious disease1 (0.43%)01 (0.43%)
 Midwife1 (0.43%)01 (0.43%)
 No response/other50 (21.4%)5 (20.8%)45 (21.4%)
Baseline demographics Results of the mixed logit model are presented in table 3.
Table 3

BEARI study mixed logit regression results

Exposure groupOR (95% CI)CoefficientSE (95% CI)p Valuep Value*
Full sample (n=239)<0.001
 Pay-for-performance1.007 (1.006 to 1.008)0.0070.001 (0.006 to 0.008)0.000
 SA2.108 (1.727 to 2.573)0.7460.102 (0.546 to 0.945)0.000
 JA1.092 (0.85 to 1.404)0.0880.128 (−0.162 to 0.339)0.490
 PC1.446 (1.184 to 1.766)0.3690.102 (0.169 to 0.568)0.000
 Additional time1.159 (1.105 to 1.216)0.1470.024 (0.1 to 0.195)0.000
Controls (n=24)0.414
 Pay-for-performance1.006 (1.003 to 1.009)0.0060.002 (0.003 to 0.009)0.000
 SA2.074 (1.302 to 3.303)0.7290.237 (0.264 to 1.195)0.002
 JA0.924 (0.512 to 1.67)−0.0790.302 (−0.67 to 0.513)0.794
 PC1.563 (1.01 to 2.421)0.4470.223 (0.01 to 0.884)0.045
 Additional time1.085 (0.99 to 1.191)0.0820.047 (-0.01 to 0.174)0.082
SA (n=135)<0.001
 Pay-for-performance1.007 (1.006 to 1.009)0.0070.001 (0.006 to 0.009)0.000
 SA2.199 (1.626 to 2.973)0.7880.154 (0.486 to 1.09)0.000
 JA1.009 (0.719 to 1.418)0.0090.173 (-0.33 to 0.349)0.957
 PC1.552 (1.173 to 2.054)0.4400.143 (0.159 to 0.72)0.002
 Additional time1.192 (1.117 to 1.274)0.1760.034 (0.11 to 0.242)0.000
JA (n=121)<0.001
 Pay-for-performance1.006 (1.005 to 1.008)0.0060.001 (0.005 to 0.008)0.000
 SA1.955 (1.504 to 2.54)0.6700.134 (0.408 to 0.932)0.000
 JA1.451 (1.061 to 1.984)0.3720.160 (0.059 to 0.685)0.020
 PC1.429 (1.097 to 1.861)0.3570.135 (0.093 to 0.621)0.008
 Additional time1.174 (1.095 to 1.259)0.1600.036 (0.09 to 0.23)0.000
PC (n=101)<0.001
 Pay-for-performance1.007 (1.005 to 1.009)0.0070.001 (0.005 to 0.009)0.000
 SA2.589 (1.819 to 3.685)0.9510.180 (0.598 to 1.304)0.000
 JA1.079 (0.684 to 1.703)0.0760.233 (−0.38 to 0.532)0.744
 PC1.285 (0.901 to 1.832)0.2500.181 (−0.105 to 0.605)0.167
 Additional time1.200 (1.103 to 1.305)0.1820.043 (0.098 to 0.266)0.000

Results indicate the OR of each alternative relative to the control group, broken down by exposure group (full sample, controls, SA, JA and PC).

*p Value for significance across all coefficients for that sample.

BEARI, use of behavioral economics and social psychology to improve treatment of acute respiratory infections; JA, Accountable Justifications; PC, Peer Comparison; SA, Suggested Alternatives.

BEARI study mixed logit regression results Results indicate the OR of each alternative relative to the control group, broken down by exposure group (full sample, controls, SA, JA and PC). *p Value for significance across all coefficients for that sample. BEARI, use of behavioral economics and social psychology to improve treatment of acute respiratory infections; JA, Accountable Justifications; PC, Peer Comparison; SA, Suggested Alternatives. χ2 statistic indicates that coefficients were significantly different from one another (p<0.001). Overall, prescribers tended to prefer any of the presented alternatives to prescriber education and guideline review, regardless of exposure group. Prescribers did not prefer PC as much as other interventions for reducing inappropriate antibiotic prescriptions. Instead, prescribers consistently preferred SA regardless of which intervention they were actually exposed to, including those in the control group. These results match those found in the pilot study; those who were not exposed to any of the BEARI interventions also strongly preferred SA (OR=2.75, p=0.003). Results were similar in the pilot group exposed to the interventions, though slightly less pronounced (OR=2.07, p=0.018) (see online supplementary table S1). These trends remained consistent whether prescribers were exposed to just one, two or all of the interventions, although the magnitude and significance of preference varied compared to the results by individual intervention (see online supplementary table S2). In contrast, PC was not strongly preferred as an intervention most likely to reduce ARI antibiotic prescribing, though still statistically significant among controls (OR=1.56, p=0.045), and those exposed to SA (OR=1.55, p=0.002) and JA (OR=1.43, p=0.008). In fact, those who were exposed to PC had a non-significant preference for this intervention (OR=1.285, p=0.167), even though it was highly effective in reducing inappropriate antibiotic prescribing rates in the BEARI experiment. Overall, prescribers were either indifferent to or did not prefer JA (although this was non-significant), unless they were exposed to that intervention in the study, in which case there was a significant preference for the intervention (OR=1.45, p=0.02). In contrast, the pilot study showed that JA was non-preferred among non-exposed prescribers (OR=0.68, p=0.049), though preferred among those exposed (OR=1.79, p=0.247). Pay-for-performance was marginally preferred among prescribers across all groups (control and treatment) with an OR=1.007, implying that each additional $100 in financial incentives would increase the probability of preferring pay-for-performance by 70%. This suggests that clinicians feel that every $100 increase in compensation would improve their inappropriate antibiotic prescribing rates by 70% relative to the control group. These results also generally held true in the pilot study. Additional office time was more strongly preferred than pay-for-performance, although this preference was non-significant in the control group (OR=1.085, p=0.082). Among those exposed to SA, JA and PC, additional office time was highly significantly preferred, with OR=1.19, 1.17 and 1.2, respectively (p<0.05). These results imply that overall, clinicians felt that additional time spent with a patient to form an appropriate diagnosis and treatment plan would be more effective than financial incentives to reduce inappropriate antibiotic prescribing. When adjusted for intervention in the interacted model, trends in stated preferences remained stable (see online supplementary table S3). Those exposed to JA did not favour SA as strongly as those exposed to SA or PC (OR=1.889, p<0.001, vs OR=2.085, p<0.001 and OR=2.406, p<0.001, respectively), as in the base case model. In addition, JA remained non-preferred even when stratified by exposure group, except for those exposed to JA in the BEARI trial. WTP estimates indicate the value of each intervention to prescribers (table 4).
Table 4

WTP estimates

WTPMonthlyAnnually
Controls
 SA−$120.52−$1446.28
 JA$12.98$155.81
 PC−$73.84−$886.09
 Additional time−$13.55−$162.61
SA
 SA−$108.21−$1298.56
 JA−$1.29−$15.48
 PC−$60.38−$724.56
 Additional time−$24.17−$290.04
JA
 SA−$109.21−$1310.46
 JA−$60.65−$727.81
 PC−$58.16−$697.90
 Additional time−$26.10−$313.23
PC
 SA−$129.43−$1553.13
 JA−$10.34−$124.06
 PC−$34.07−$408.80
 Additional time−$24.79−$297.43

Negative values indicate the value of that particular alternative to the prescriber; thus, the absolute value is the amount the prescriber is willing to accept; a positive number implies that the prescriber would be willing to give up that dollar amount rather than use the alternative presented.

JA, Accountable Justifications; PC, Peer Comparison; SA, Suggested Alternatives; WTP, willingness-to-pay.

WTP estimates Negative values indicate the value of that particular alternative to the prescriber; thus, the absolute value is the amount the prescriber is willing to accept; a positive number implies that the prescriber would be willing to give up that dollar amount rather than use the alternative presented. JA, Accountable Justifications; PC, Peer Comparison; SA, Suggested Alternatives; WTP, willingness-to-pay. SA was equivalent to $1299–$1553 in financial incentives per year, whereas PC was only worth $408–$886 per year. Each minute of additional office time for patient education was only worth an average $265 per year. These estimates imply that prescribers could be paid an average of $1400, $650 or $265 per year to achieve the same results as SA, PC and additional time, respectively. These results also imply that SA is equivalent to an additional 5–6 min of office time, and PC 3 min. In contrast, prescribers in the control group would rather sacrifice $13 of pay per month (or $156 annually) than be required to enter a justification for every inappropriate antibiotic prescription (JA). Compared to the results of the DCE, the exit survey yielded an overwhelmingly ambivalent attitude towards the BEARI study as a whole, and towards each individual intervention. While 60% of prescribers agreed that providing feedback on clinician performance and using electronic decision support tools are effective ways to improve quality care, 42% of prescribers felt that the BEARI interventions were neither useful nor useless in improving antibiotic prescribing practices. When asked about the usefulness of the PC emails, 37% of prescribers responded neutrally, while 14% felt the information was not at all useful, and only 7% found it very useful. For SA and JA, 30% of prescribers responded neutrally, ∼20% felt the interventions were not at all useful and only 10% found the interventions to be very useful.

Discussion

Overall, the DCE-elicited preferences did not reflect actual behaviour as revealed in the RCT of prescriber interventions to alter antibiotic prescribing. Although the DCE overwhelmingly favoured SA as the most effective method for reducing inappropriate antibiotic prescriptions across treatment and controls, this intervention yielded similar antibiotic prescribing rates to controls throughout the duration of the BEARI trial.18 On the other hand, most prescribers generally felt that JA was ineffective in reducing their ARI antibiotic prescribing, despite its significant impact on inappropriate antibiotic prescription rates in the trial. PC was generally significantly preferred to provider education, although this preference was not as strong as that for SA. Regardless of the intervention group to which participants were assigned in the trial, the general trends in stated preference were similar across all groups in the DCE survey, while the trial showed that the trajectory of antibiotic prescribing rates for JA and PC relative to the control group consistently decreased compared to SA. Differences between stated and revealed preferences may be due to a number of factors. Prescribers did not get to choose their intervention in the randomised trial; therefore, prescribing behaviour may not represent the prescriber's true ‘revealed preference’. Another consideration is that in the DCE task, prescribers may not be choosing the intervention that best reduces inappropriate antibiotic prescribing; instead, the stated preference might be for the intervention that is the least inconvenient for the prescriber. The DCE results are also reflected in the exit survey, which indicated that the majority of prescribers rated the usefulness of each intervention as 1, 2 or 3 on a scale of 1–5 (1 being not at all useful, 5 being very useful). Only 10% of prescribers found SA to be very useful (rating of 5 on a scale of 1–5), while just 8% and 6% found JA and PC to be very useful interventions, respectively. Given the lack of enthusiasm in the survey responses, it is unsurprising that even SA, the most preferred intervention according to the DCE, was only worth up to $1500 annually to clinicians, and the other interventions even less. SA may be the most preferred intervention overall because it is a clinical decision support tool, as opposed to a socially motivated intervention such as JA or PC. The stated preference for SA is consistent with design recommendations that if Electronic Health Record (EHR) alerts are used, they should include actionable information.32 33 On the other hand, socially motivated interventions force the prescriber to conform to ‘social norms’ due to a concern about his/her social reputation and/or an awareness of the social norm of cooperation.34 Rather than applying social pressure, SA bears no social consequences if the prescriber chooses to ignore the alternatives presented. In contrast, prescribers' dislike of JA is consistent with findings describing the phenomenon of ‘alert fatigue’, and relatively poor efficacy of alerts that do not include actionable recommendations.33 35 However, JA in BEARI was actually highly effective in the RCT, and differs from similar studies that do not include peer accountability.33 One limitation of the BEARI study is that there were many incomplete responses to the DCE task. However, results for the entire sample were compared to results for those who completed six or more choice tasks, and there were no significant differences. In addition, geographic variation in prescribing habits and patient attitudes may have affected the rate at which inappropriate antibiotics were prescribed, regardless of the presence of a behavioural economic intervention. However, it is assumed that such attitudes and behaviours would have been balanced among all treatment groups after randomisation. Responses may have been different had the DCE been administered before the RCT took place, since prescribers would not have actually experienced the interventions prior to being asked about them. The pilot DCE was administered to prescribers who did not receive any of the interventions, yet it showed similar statistically significant stated preferences to those elicited from prescribers who participated in the trial. This, along with the fact that actual intervention assignment did not affect the overall trend of stated preferences, suggests that prescribers appear to express the same preferences regardless of exposure to the BEARI interventions. Finally, contrary to consensus regarding the inclusion of an opt-out option in DCE to reflect a respondent's choice in real life, no opt-out option was included.36 37 However, one study empirically evaluated the effect of an opt-out option on attribute preference, and found that while attribute value estimates differed, there were no notable differences in the relative order of the attributes (when compared to each other).36 Based on their findings, the authors recommended that an opt-out option always be included in a DCE if non-participation is an option in real life as well. In BEARI, prescribers did not have the option to opt out of participation, and it is likely that if the BEARI interventions were implemented in practice, prescribers would also be unable to opt out. Compared to previous studies attempting to evaluate the external validity of DCEs in healthcare, this study provides stronger evidence that stated and revealed preferences can be non-concordant. Specifically, revealed preferences were determined through a RCT, and the same individuals were then asked to complete the DCE. Participants were not given a choice as to which intervention they wished to be implemented in clinic; instead, real-life behaviour was captured in the trial. Prescribers were also from various types of practice settings across the country, increasing the external validity of the results, rather than eliciting responses from a single practice site. Previous studies did not use an experimental design to determine revealed preferences, and thus, the results are not as robust. Ultimately, had the BEARI trial relied upon the DCE preferences of the clinicians unexposed to the interventions to inform intervention selection, the least effective intervention (SA) would have been adopted and one of the effective interventions (JA or PC) would have been rejected. Relying on the exit survey responses would have led to none of the interventions being adopted. Our findings are consistent with recommendations that quality improvement interventions combining local participatory approaches with expert-driven approaches are likely to be most effective.38 In particular, approaches that engage targets in the implementation strategy rather than the design and development of interventions may optimally ensure that interventions are deployed in a pragmatic way without being influenced by the stated preferences of clinicians.

Conclusion

This study is part of a growing body of literature comparing stated and revealed preferences in a healthcare setting. Consistent with similar studies evaluating the external validity of DCE-stated preferences, this study showed that they are not concordant, and suggests that quality improvement interventions should not rely on frontline staff input alone. Future work should continue to test the external validity of DCE preferences in healthcare and the factors that contribute to differences between stated and revealed preferences.
  30 in total

Review 1.  From understanding health care provider behavior to improving health care: the QUERI framework for quality improvement. Quality Enhancement Research Initiative.

Authors:  L V Rubenstein; B S Mittman; E M Yano; C D Mulrow
Journal:  Med Care       Date:  2000-06       Impact factor: 2.983

Review 2.  A behavioral model of clinician responses to incentives to improve quality.

Authors:  Anne Frølich; Jason A Talavera; Peter Broadhead; R Adams Dudley
Journal:  Health Policy       Date:  2006-04-19       Impact factor: 2.980

3.  Asymmetric paternalism to improve health behaviors.

Authors:  George Loewenstein; Troyen Brennan; Kevin G Volpp
Journal:  JAMA       Date:  2007-11-28       Impact factor: 56.272

4.  Comparing welfare estimates from payment card contingent valuation and discrete choice experiments.

Authors:  Mandy Ryan; Verity Watson
Journal:  Health Econ       Date:  2009-04       Impact factor: 3.046

5.  Automating complex guidelines for chronic disease: lessons learned.

Authors:  Saverio M Maviglia; Rita D Zielstorff; Marilyn Paterno; Jonathan M Teich; David W Bates; Gilad J Kuperman
Journal:  J Am Med Inform Assoc       Date:  2003 Mar-Apr       Impact factor: 4.497

6.  Balancing health care evidence and art to meet clinical needs: policymakers' perspectives.

Authors:  Louise E Parker; Mona J Ritchie; Joann E Kirchner; Richard R Owen
Journal:  J Eval Clin Pract       Date:  2009-12       Impact factor: 2.431

7.  Using stated preference modeling to forecast the effect of medication attributes on prescriptions of alcoholism medications.

Authors:  Tami L Mark; Joffre Swait
Journal:  Value Health       Date:  2003 Jul-Aug       Impact factor: 5.725

8.  Conducting discrete choice experiments to inform healthcare decision making: a user's guide.

Authors:  Emily Lancsar; Jordan Louviere
Journal:  Pharmacoeconomics       Date:  2008       Impact factor: 4.981

Review 9.  Discrete choice experiments in health economics: a review of the literature.

Authors:  Esther W de Bekker-Grob; Mandy Ryan; Karen Gerard
Journal:  Health Econ       Date:  2010-12-19       Impact factor: 3.046

10.  Provision of social norm feedback to high prescribers of antibiotics in general practice: a pragmatic national randomised controlled trial.

Authors:  Michael Hallsworth; Tim Chadborn; Anna Sallis; Michael Sanders; Daniel Berry; Felix Greaves; Lara Clements; Sally C Davies
Journal:  Lancet       Date:  2016-02-18       Impact factor: 79.321

View more
  8 in total

Review 1.  Behavioral Economics and Ambulatory Antibiotic Stewardship: A Narrative Review.

Authors:  Alexandra R Richards; Jeffrey A Linder
Journal:  Clin Ther       Date:  2021-10-23       Impact factor: 3.637

Review 2.  Systematic review and validity assessment of methods used in discrete choice experiments of primary healthcare professionals.

Authors:  Gregory Merlo; Mieke van Driel; Lisa Hall
Journal:  Health Econ Rev       Date:  2020-12-09

Review 3.  Application of discrete choice experiments to enhance stakeholder engagement as a strategy for advancing implementation: a systematic review.

Authors:  Ramzi G Salloum; Elizabeth A Shenkman; Jordan J Louviere; David A Chambers
Journal:  Implement Sci       Date:  2017-11-23       Impact factor: 7.327

4.  The clinician crowdsourcing challenge: using participatory design to seed implementation strategies.

Authors:  Rebecca E Stewart; Nathaniel Williams; Y Vivian Byeon; Alison Buttenheim; Sriram Sridharan; Kelly Zentgraf; David T Jones; Katelin Hoskins; Molly Candon; Rinad S Beidas
Journal:  Implement Sci       Date:  2019-06-14       Impact factor: 7.327

5.  Exploring determinants of health provider choice and heterogeneity in preference among outpatients in Beijing: a labelled discrete choice experiment.

Authors:  Jingrong Zhu; Jinlin Li; Zengbo Zhang; Hao Li; Lingfei Cai
Journal:  BMJ Open       Date:  2019-04-03       Impact factor: 2.692

6.  Effectiveness of interventions targeting antibiotic use in long-term aged care facilities: a systematic review and meta-analysis.

Authors:  Magdalena Z Raban; Claudia Gasparini; Ling Li; Melissa T Baysari; Johanna I Westbrook
Journal:  BMJ Open       Date:  2020-01-09       Impact factor: 2.692

7.  Community stakeholder preferences for evidence-based practice implementation strategies in behavioral health: a best-worst scaling choice experiment.

Authors:  Nathaniel J Williams; Molly Candon; Rebecca E Stewart; Y Vivian Byeon; Meenakshi Bewtra; Alison M Buttenheim; Kelly Zentgraf; Carrie Comeau; Sonsunmolu Shoyinka; Rinad S Beidas
Journal:  BMC Psychiatry       Date:  2021-02-04       Impact factor: 3.630

Review 8.  It's about the patients: Practical antibiotic stewardship in outpatient settings in the United States.

Authors:  Alpesh N Amin; E Patchen Dellinger; Glenn Harnett; Bryan D Kraft; Kerry L LaPlante; Frank LoVecchio; James A McKinnell; Glenn Tillotson; Salisia Valentine
Journal:  Front Med (Lausanne)       Date:  2022-07-27
  8 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.