Literature DB >> 34074902

Evaluations of Continuing Health Provider Education Focused on Opioid Prescribing: A Scoping Review.

Abhimanyu Sud1, Graziella R Molska2, Fabio Salamanca-Buentello3.   

Abstract

PURPOSE: Continuing health provider education (HPE) is an important intervention supported by health policy to counter the opioid epidemic; knowledge regarding appropriate program design and evaluation is lacking. The authors aim to provide a comprehensive understanding of evaluations of opioid-related continuing HPE programs and their appropriateness as interventions to improve population health.
METHOD: In January 2020, the authors conducted a systematic search of 7 databases, seeking studies of HPE programs on opioid analgesic prescribing and overdose prevention. Reviewers independently screened the titles and abstracts of all studies and then assessed the full texts of all studies potentially eligible for inclusion. The authors extracted a range of data using categories for evaluating complex programs: the use of theory, program purpose, inputs, activities, outputs, outcomes, and industry involvement. Results were reported in a narrative synthesis.
RESULTS: Thirty-nine reports on 32 distinct HPE programs met inclusion criteria. Of these 32, 31 (97%) were U.S./Canadian programs and 28 (88%) were reported after 2010. Measurements of changes in knowledge and confidence were common. Performance outcomes were less common and typically self-reported. Most studies (n = 27 [84%]) used concerns of opioid-related harms at the population health level to justify the educational intervention, but only 5 (16%) measured patient- or population-level outcomes directly related to the educational programs. Six programs (19%) had direct or indirect opioid manufacturer involvement.
CONCLUSIONS: Continuing HPE has been promoted as an important means of addressing population-level opioid-related harms by policymakers and educators, yet published evaluations of HPE programs focusing on opioid analgesics inadequately evaluate patient- or population-level outcomes. Instead, they primarily focus on self-reported performance outcomes. Conceptual models are needed to guide the development and evaluation of continuing HPE programs intended to have population health benefits.
Copyright © 2021 by the Association of American Medical Colleges.

Entities:  

Mesh:

Substances:

Year:  2022        PMID: 34074902      PMCID: PMC8781229          DOI: 10.1097/ACM.0000000000004186

Source DB:  PubMed          Journal:  Acad Med        ISSN: 1040-2446            Impact factor:   7.840


Opioid-related harms, at epidemic levels throughout the United States and Canada, have worsened in the context of the COVID-19 pandemic. [1-3] While an increasing proportion of deaths is due to nonmedical opioids (e.g., heroin, fentanyl, fentanyl analogues) and to polysubstance use, prescribed opioids continue to contribute to this epidemic. [4,5] In response, numerous prescribing-focused policies have been implemented. These include prescription drug monitoring programs (PDMPs), [6] regulatory authority investigations of prescribers, and the development and dissemination of clinical practice guidelines. [7,8] Health provider education (HPE) has been central to the implementation of many of these interventions, and a variety of major policies, as well as national media reporting, have identified HPE as a pivotal strategy for countering the opioid epidemic. [9-12] The central role of HPE is driven by a recognition that, heretofore, HPE on pain has been deficient in terms of both quantity and quality across the health professions education continuum. [13,14] A compounding factor is the notable history of pharmaceutical industry involvement, which has influenced prescriber behavior not just through commercial promotions but also through heavy investments in HPE. [15,16] Although HPE is considered an important policy intervention for countering the opioid epidemic, there is a gap in understanding the appropriate design of these programs, how they are evaluated, and the manner in which program components affect outcomes. [17,18] This gap persists despite at least 2 systematic reviews that have assessed the effectiveness of HPE for opioid prescribing. [19,20] The role of educational theory informing the development, delivery, and evaluation of these programs is particularly unclear—as is how educational theory should inform population health policy.

Objective and Research Questions

The objective of this scoping review is to provide a comprehensive understanding of evaluations of continuing HPE focused on opioid prescribing. The following 4 related questions guided our research: What explicit educational, evaluative, or health policy frameworks; models; or theories are used to inform evaluative reports about opioid continuing HPE? What evidence-based program components are included in the interventions? What outcome types do these studies report? What is the role of the pharmaceutical industry in opioid continuing HPE?

Method

Two frameworks for scoping reviews [21,22] and a guide for conducting systematic reviews in medical education [23,24] informed this review. Additionally, we adapted a framework for applying theory to synthesize evidence from systematic reviews. [25] We used 2 related sets of knowledge on continuing HPE to synthesize the results. We used a review of systematic reviews on the effectiveness of continuing medical education to categorize evidence-based best practices in continuing HPE [26] and we applied an outcomes framework for planning and assessing continuing HPE to determine outcome types. [27,28] To facilitate consideration of HPE as health policy, we categorized these education outcomes as implementation, effectiveness, and impact outcomes according to frameworks for complex interventions (Figure 1). [29] We developed an a priori protocol for this review and made no amendments to the protocol.
Figure 1

Mapping Moore and colleagues’ continuing health provider education outcome levels (left of figure) [27,28] to complex intervention outcome evaluation types (right of figure). [29] Reprinted with permission from Moore DE Jr, Chappell K, Sherman L, and Vinayaga-Pavan M. A conceptual framework for planning and assessing learning in continuing education activities designed for clinicians in 1 profession and/or clinical teams. Med Teach. 2018;40:904–913.

Mapping Moore and colleagues’ continuing health provider education outcome levels (left of figure) [27,28] to complex intervention outcome evaluation types (right of figure). [29] Reprinted with permission from Moore DE Jr, Chappell K, Sherman L, and Vinayaga-Pavan M. A conceptual framework for planning and assessing learning in continuing education activities designed for clinicians in 1 profession and/or clinical teams. Med Teach. 2018;40:904–913.

Inclusion/exclusion criteria

Participants.

We included studies of practicing health professionals involved in opioid prescribing for chronic pain, including physicians, nurses/nurse practitioners, dentists, and pharmacists. We excluded studies of continuing HPE whose primary participants were trainees (e.g., medical residents) or non–health professionals (e.g., patients, family members).

Interventions.

We included studies of educational programs relating to opioid analgesic or naloxone (opioid antagonist) prescribing. We excluded studies of opioid agonist and antagonist therapies, which are typically prescribed for the management of opioid use disorder rather than for pain. We used a broad and widely accepted definition of continuing education (CE) that encompasses interventions ranging from live, group-based lectures (“traditional CE”) to asynchronous online self-learning (e.g., archived webinars) to one-to-one, point-of-care learning (e.g., academic detailing). [30] The evaluations had to focus primarily on the educational intervention and include any kind of evaluative learning outcome. [26] We excluded evaluations of larger complex interventions that, while incorporating education, did not report specific evaluations of the continuing HPE component. We did not include any criteria about comparators.

Types of studies.

We included any kind of evaluative study design, such as quasi-experimental pre–post designs, prospective and retrospective observational designs, randomized controlled trials, and qualitative studies. We did not include systematic reviews or other evidence syntheses, nor did we include nonevaluative articles such as commentaries, editorials, or conceptual reports. We excluded studies focused only on needs assessments or program development that did not also incorporate a program evaluation. Meeting abstracts were not included in our review, but we used such records to identify possible full-length reports that would meet our inclusion criteria. To find these, we conducted manual searches and contacted authors of relevant meeting abstracts.

Context.

We did not set any limits in terms of timing, geography, or language given that problematic opioid prescribing has been a long-standing issue and a global phenomenon. [31] Based on our previous research, we anticipated that the majority of reports would likely be from the United States and written in English. [32] We included only studies of prescribers who worked in outpatient, ambulatory settings. We excluded studies based in hospital settings (e.g., perioperative clinical encounters) because priorities and practices of opioid prescribing in these contexts differ substantially from ambulatory settings where the majority of long-term opioid prescribing occurs. Additionally, more organizational (specifically noneducational) factors influence opioid prescribing in these inpatient settings. We excluded studies focused only on palliative and end-of-life care as these are different treatment contexts with different educational priorities.

Search strategy

Our search strategy followed the 3-step approach recommended by the Joanna Briggs Institute, [33] and we consulted with a professional librarian. The first step involved an initial limited search of MEDLINE, referenced against a set of high-relevance sentinel studies identified from 2 recent systematic reviews. [19,20] We then analyzed the words and language used in the titles, abstracts, and index terms of relevant studies to refine our search strategy (see Supplemental Digital Appendix 1 at http://links.lww.com/ACADMED/B127 for full database searches). The second step was to conduct a comprehensive search in the following databases: Embase (Embase Classic + EMBASE) Cumulative Index to Nursing and Allied Health (EBSCOHost) MEDLINE (Medline-in-Process, Medline Epubs Ahead of Print) PsycINFO Dissertation Abstracts Education Resources Information Centre (ProQuest) Cochrane Trials and International Trials Registers and Cochrane Database of Systematic Reviews and Protocols (Wiley). We used medical subject headings (MeSH) and textword terms to search for the concepts of “opioids” and “continuing health professional education.” We searched all databases from inception to January 20, 2020. The final step entailed a manual search of the references of the included studies to identify any additional relevant records.

Study selection

Search results were first uploaded into EndNote (Clarivate Analytics, Philadelphia, Pennsylvania), and duplicates were removed using a systematized process. [34] We then uploaded these records to the Covidence (Melbourne, Victoria, Australia) online data management platform for screening. Four reviewers (including A.S. and G.R.M.) independently screened titles and abstracts in duplicate and resolved most conflicts. When they could not reach an agreement, the duplicate reviewers discussed the record with a third reviewer, and we made a final decision by consensus. Next, 3 reviewers (including A.S. and G.R.M.) conducted a full-text review of the remaining records in duplicate, tracking their reasons for excluding studies using Covidence. Again, these reviewers discussed conflicts and consulted the third reviewer when they could not agree, and we made a final decision by consensus. We followed the Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) extension for scoping reviews for manuscript preparation, [35] and we created a PRISMA flow diagram to present the number of records identified, screened, assessed for eligibility, and included or excluded (along with the reasons they were excluded).

Data extraction and charting

We developed a data extraction form using Excel software (Version 16.4, Microsoft, Redmond, Washington). We adapted a logic model for medical education programs, [36] and we applied both (1) Moore and colleagues’ outcomes framework for continuing HPE [27,28] and (2) the evidence-based continuing HPE practices from Cervero and Gaines’ review of systematic reviews. [26] We extracted the following data from each study: Theory (explicit, implicit, or missing); Program purpose (including the primary problem, target audience, and needs assessment); Inputs (i.e., faculty, funding, registration fees, regulatory involvement, accreditation [whether learners earned CE credit], promotion/recruitment, program development); Activities (i.e., duration, setting, synchronicity, group size [< 20 small, ≥ 20 large], number of interventions, number of exposures, interactivity); Outputs (i.e., number of programs, location of programs, external factors identified [such as incentives for participation]); Outcomes (i.e., participation, satisfaction, knowledge, confidence, attitudes, competence, performance, patient health, population health); and Industry involvement. Two of us (A.S. and G.R.M.) tested the form on 3 studies to ensure all relevant results were extracted, and we made adjustments to the form accordingly.

Results

After removing 638 duplicates, the database search yielded 3,075 records. Of these, 2,886 were considered irrelevant and removed after title and abstract screening. We screened the full text of 189 articles, and we excluded 149 (Figure 2; see Supplemental Digital Appendix 2 at http://links.lww.com/ACADMED/B128 for the 149 articles excluded and the reasons for their exclusion). At the data extraction phase, we excluded 2 articles that were reviews of policies regarding opioid-related continuing HPE, rather than evaluations of specific programs. [37,38] At this stage, we also identified an additional study to include that provided more information about the delivery and development of a program described in another included study. Of the remaining 39 records, [39-77] 7 reported different outcomes for programs covered in other reports, so we analyzed these with the primary record (and the denominators for the data reported below are 32 unless otherwise stated).
Figure 2

PRISMA [35] flow diagram presenting the number of records identified, screened, assessed for eligibility, and included or excluded (along with the reasons they were excluded). Abbreviations: PRISMA, Preferred Reporting Items for Systematic Reviews and Meta-Analyses; CINAHL, Cumulative Index to Nursing and Allied Health; ERIC, Education Resources Information Centre.

PRISMA [35] flow diagram presenting the number of records identified, screened, assessed for eligibility, and included or excluded (along with the reasons they were excluded). Abbreviations: PRISMA, Preferred Reporting Items for Systematic Reviews and Meta-Analyses; CINAHL, Cumulative Index to Nursing and Allied Health; ERIC, Education Resources Information Centre.

Program descriptions

In total, we analyzed 39 evaluations of 32 continuing HPE programs from 28 distinct author groups. The reports spanned 1983 to 2020, though reports on 28 programs (88%) were published during or after 2010 (Appendix 1). The vast majority reported on U.S. or Canadian programs: 25 (78%) were based in the United States, 6 (19%) in Canada, and 1 (3%) in Turkey. [64] Physician participants were the primary learning targets of 28 (88%) of the programs and were the exclusive targets of 17 (53%). Only 2 of the programs that targeted exclusively physicians [50,66] (6% of the 32) focused on naloxone prescribing while the rest focused on opioid analgesics. Of the 32 programs, 9 (28%) also focused on nurse practitioners, but never exclusively. Physician assistants and dentists were included in, respectively, 2 programs [49,75] (6%) and 1 program [40] (3%). Four programs (13%) focused on only pharmacists and/or pharmacy technicians [41,55-57,68,69] and 2 other programs (6%) included pharmacists together with physicians. [40,66] Of the 6 programs that included pharmacists, 4 focused on naloxone prescribing or distribution. [55-57,66,68,69] Opioid-related continuing HPE targeted practitioners in a variety of settings, including community primary care (11 programs [34%]), community pharmacies (4 [12%]), academic internal medicine practices (2 [6%]), providers in veterans and military health settings (2 [6%]), and the United States Indian Health Service (1 program [3%]). Reports on 10 programs (31%) did not describe a targeted setting.
Most programs (n = 19 [59%]) were delivered via traditional live group learning, including both large groups (15 programs [47%]) and small groups (4 [12%]). These were typically delivered as standalone programs, but we also found examples of delivery at the workplace [44-51,53,59,61,64,68] or in the context of a conference. [39,52,53,65,67,69,72] Some of these traditional live group programs also provided an alternative online individual asynchronous version [39,47,48,52,55-57,66,69]; none of these blended in-person and online programs. One program, [67] which was the first to report any kind of virtual delivery of education, did blend group learning with cases emailed monthly to participants after the in-person intervention. Four programs (12%) [41,54,58,75] were delivered asynchronously online only and 1 (3%) [63] was delivered only synchronously online; the evaluations for all 4 of these were published recently (between 2016 and 2020). We found 4 examples of academic detailing [44-46,49,50,61] and 2 of audit and feedback [59,70]—all of which were delivered at the workplace. Articles on 9 of the 32 programs (28%) reported professional (medical or pharmacy) regulator involvement in the development or delivery of the program. Most of these programs involved traditional live group learning, though we also noted examples of other delivery modes. Of these 9 programs, 3 mandated physician participation in the program (e.g., due to identified inappropriate prescribing), [62,63,73,74] and another mixed mandated and voluntary participants. [61] None of the programs with pharmacist regulator involvement mandated pharmacist involvement. The single study evaluating the effect of an educational program for United States Indian Health Service providers [63] required participation, but by mandate of the employer, not of the medical regulator. Articles on more than half of the programs (n = 17 [53%]) reported accreditation. Only 6 (19%) reported the specific number of credits participants could earn, which ranged from 0.2 to 20. In a further 6 programs (19%), the opioid-related continuing HPE was embedded within larger complex interventions. These included the following: education embedded in a quality improvement program focused on the use of an electronic medical record (EMR)-based protocol, [51] a public health detailing program embedded in a jurisdiction-wide opioid crisis campaign, [61] an educational program that was part of a statewide campaign, which also included a media strategy, the release of prescribing guidelines, the formation of public health coalitions, and private industry–led public education campaigns, [53] the distribution of workplace naloxone toolkits together with the delivery of educational programs for pharmacists, [41,55-57] and a statewide approach to facilitating naloxone utilization, delivered with pharmacist and pharmacy technician training in naloxone. [68]

Evaluation design and evidence-based components

Articles on most programs (n = 27 [84%]) reported that population-level concerns of opioid-related harms driven by inappropriate prescribing were used to justify the continuing HPE program and its evaluation (Appendix 2). Studies of 5 programs (16%) reported, instead, focusing on performance-level problems such as practices relating to opioid prescribing and chronic pain. Articles on 3 of these programs [42-46,73,74] published outcomes in the 1980s and 1990s, before the recognition of widespread, population-level harms from opioids. A more recent study (from 2015) was based in Turkey, [64] where the epidemiology of opioid prescribing and related harms is distinct from that of the United States and Canada. The final of these 5 studies [65] focused on end-of-life and palliative care, rather than only chronic pain management.
Articles on only 6 programs (19%) identified an explicit model for the development or evaluation of the educational intervention. One set of authors [42,43] identified the Transtheoretic Model of Behavior Change as the primary orienting model for their intervention. This program also applied Moore and colleagues’ framework, [27] as did another program, using traditional live group learning. [72] Two programs applied the Theory of Planned Behavior, [41,55] and one of these also referenced the Kirkpatrick model in describing program development. [55-57] The single example of an explicitly interprofessional program [52] reported the use of frameworks for interprofessional education. [78-80] One intervention entailing academic detailing [47,48] applied the Medical Research Council’s framework for complex interventions for both program development and program evaluation. [81] Finally, 1 traditional live, small-group program with regulatory involvement was designed around models of physician behavior change. [73] While the studies of the other programs did not explicitly reference any theories, models, or frameworks for program development or evaluation, we could identify some implicit models. For example, 1 program [51] used quality improvement and change management as the overarching framework for its intervention. Likewise, an academic detailing intervention [44-46] applied a combination of communication theory and behavioral science for development. Another 2 programs [58,77] used a clear outcomes-based approach to evaluation congruent with Moore and colleagues’ framework, [27] focusing on the change in clinician knowledge, competence, confidence, and performance. Similarly, an additional program discussed an approach to physician behavior change driven by changes in confidence and knowledge. [66] In terms of evaluation, studies of more than half of the programs (n = 17 [53%]) used quasi-experimental pre–post designs. Of these 17, only 2 included comparator groups. [52,77] Evaluations of 6 programs [53,65,68,69,74,76] (19%) measured only postintervention outcomes without any baseline measures. The studies of 2 programs (6%) examining the effects of the educational interventions on physician prescribing used retrospective cohort studies, [50,60] and another one (3%) used a time series analysis. [59] We identified 5 controlled trials [42-46,49,70,75] one of which [49] did not report clear randomization between groups. Two of the 4 randomized controlled trials were among the earlier studies published in the 1980s and 1990s. [42-46] Both of the audit and feedback studies [59,70] and 2 of the 4 academic detailing studies [44-46,49] used designs that allow for greater causal inference, namely controlled trials and time series analysis. Articles on 4 programs (12%) reported use of all the evidence-based components of interventions likely to improve physician performance and patient health. [42,43,60,67,75] These programs also tended to use evaluation designs that allow for greater causal inference (randomized controlled trials and a retrospective cohort study). Studies on 20 of the programs (62%) reported interactivity, and studies on a different 20 (62%) reported use of multiple methods. About a third of the programs (12 [38%]) included more than 1 educational exposure. The duration of the interventions ranged from 13 minutes to 780 minutes; most reported under 300 minutes. The longer programs tended to be traditional live group learning, while the academic detailing interventions and several of the virtual-only programs tended to be the shortest. Studies of 6 programs (19%), including the 2 audit and feedback programs and 1 program that included only mailed information, did not specify the duration of the intervention. [50,59,64,70,73,74,76] All the evaluations of the programs reported participation as an evaluation outcome, but articles on most (n = 22 [69%]) did not include satisfaction. Of the 32 distinct programs, evaluation measures of 16 of them (50%) included changes in knowledge, 18 (56%) included changes in confidence, and a different 18 (56%) included changes in attitudes. The studies of only 5 programs (16%) included measures of competence. One of these was a study of an educational program to improve enrollment in a PDMP that included self-reported changes in competence. [58] Performance, evaluated in 23 (72%) of programs, was much more commonly measured; however, 13 of these 23 studies included self-reported performance changes. Of the 8 programs for which objective performance changes (e.g., in opioid prescribing rates) were measured, 6 used either randomized controlled or retrospective cohort studies. The evaluation of only 1 of these 8 reported an implementation outcome besides participation. [44] The articles on 3 additional programs reported objective performance changes—not of the program participants but of the entire jurisdiction where the intervention was delivered. [61,62,74] Evaluations of 6 programs [49,51,55-57,59,68,70] (19%) reported patient health outcomes; however, the patient outcomes for one of these programs was not directly connected to patients cared for by program participants. [68] The studies of 3 programs (9%), all evaluating traditional live group learning programs for physicians or pharmacists, reported community health outcomes. [53,62,68] Yet, none of these outcomes related directly to program participants’ patients or communities; rather, programs reported overall patterns of opioid-related harms in the jurisdictions and during the time in which the intervention took place.

Industry involvement

The evaluations of 6 programs (19%), all of which were published between 2016 and 2018, reported either direct or indirect industry involvement in the development or evaluation of the continuing HPE intervention. All of these were opioid analgesic education programs; none were naloxone training programs. Three were developed under the opioid analgesic risk evaluation and mitigation strategy (REMS), which is funded by opioid manufacturers, and all 3 of these reported accreditation. [39,54,77] Another program used content from a REMS program and received an educational grant from a pharmaceutical manufacturer. [52] A further program, not related to REMS, also received an educational grant from a pharmaceutical manufacturer. [58] The sixth program was sponsored by a pharmaceutical manufacturer and a private insurer, and the study authors were compensated directly by the pharmaceutical company. [70]

Discussion

Summary of findings

The published literature on opioid-related continuing HPE makes little explicit use of theory from any domain to guide either development or evaluation. This paucity is emblematic of the general field of continuing HPE, in which the development and use of theory, frameworks, and models are considered to be in their infancy. [26] This situation, though, is common to improvement programs even outside of HPE. As Davidoff and colleagues note, Despite the potential value of theory, a striking feature of many improvement efforts is the tendency of its practitioners to move straight to implementation, skipping the critical working out of the programme theory; for example, sometimes only the source of the problem is identified but not an accompanying theory of change. [82] This assertion accurately characterizes the bulk of studies included in our review. Without attending to theory, opioid-related continuing HPE program developers miss opportunities for expanding and optimizing impact and for contributing to the advancement of the field—all of which should be considered priorities in the context of the public health crisis that the programs claim to address. [82] Despite the paucity of theory to guide continuing HPE development and delivery, the extant literature provides substantial knowledge about components of programs that are likely to drive changes in practice behaviors and patient outcomes. [26] Similarly, conceptual frameworks making use of this published knowledge have been developed and widely disseminated. [27] Continuing HPE programs focused on opioids have so far made only moderate use of these best practices, which include interactivity, multiple methods, and multiple exposures. This underuse is associated with minimal reporting about patient- or population-level health outcomes. Again this problem is not specific to continuing HPE related to opioid analgesics but reflects the continuing HPE and health policy fields. [83-86] Some reasons for this dearth may include a lack of financial resources, poor data sources, or lack of methodological expertise. [87,88] Yet these omissions are hard to justify when the primary motivation for the development and support of the opioid-related continuing HPE programs is the mitigation of population-level harms resulting from opioid use. This weakness is nowhere more clear than in the dates of publication of the included studies: the vast majority were published from 2010 onward and none were published between 2000 and 2009. These dates indicate that continuing HPE focused on opioid analgesics became a topic of scientific and evaluative value only after population-level opioid-related harms were more widely acknowledged. Among the few programs that explicitly referenced theory, the most frequently mentioned frameworks were the Transtheoretic Model and the Theory of Planned Behavior. This behavior change orientation was also implicit in a number of the included studies. For example, Leong and colleagues claimed, There are many stages in the outcomes of CME [continuing medical education] training for physicians. The initial stage is for the physician to assimilate the new knowledge. The second stage is that the physician be convinced enough of the new information that they are willing to change their practice. Only then can the third stage of implementation of the new material be realized and improvements in patient care be achieved. [65] This implicit theory—that knowledge acquisition will lead to behavior change which will, in turn, result in improved patient care—appears to justify the assumption that measures of changes in knowledge or clinical performance are sufficient to presume patient- and population-level effects. Yet, this stance is not well supported either by good theory or strong evidence. [26] The few studies that did report objective measures of performance change tended not to measure implementation outcomes. As is common in other outcomes-based program evaluations, these studies seemed to treat the educational interventions as “black boxes”; that is, they aimed to determine primarily if the interventions worked but did not provide a good framework or data for understanding why or how the interventions worked or did not work. [89] That such frameworks are lacking was evident in many of the included studies examining patient- and population-level outcomes. Three of these studies [53,62,68] reported evaluative outcomes that did not link directly to program participants but instead to ecological trends in the jurisdictions in which the intervention was delivered. The U.S. Food and Drug Administration (FDA) identified this same serious methodological flaw in its review of the Extended-Release/Long-Acting Opioid Prescribing REMS program, a major U.S. continuing HPE initiative aimed at reducing opioid analgesic harms. [38] The pharmaceutical industry is known to have played an insidious role in the origin of the contemporary opioid epidemic, particularly by promoting opioid analgesic use, often via educational programs for prescribers. [90] Our review has demonstrated the persistent, though not always direct, role of the pharmaceutical industry in setting the U.S. and Canadian opioid-related continuing HPE agenda. This presence is more noticeable in the opioid analgesic studies than in the naloxone studies. Four of the opioid analgesic studies were evaluations of REMS-compliant programs. In addition to the poor evaluation of these programs from the U.S. FDA (mentioned above), the opioid REMS program has been criticized for program and blueprint developers’ ties to industry, for the nature and orientation of the content of the compliant programs, and for the possibility that the connections to industry are obscured upon delivery. [91]

Recommendations for practice

Based on the findings of this review, we have 2 primary recommendations for the practice of continuing HPE programs (especially those that aim to effect improvements at the population health level): (1) adopting evaluation frameworks for complex interventions and (2) using conceptual models that link clinical practice and population health.

Evaluation frameworks for complex interventions.

Continuing HPE interventions meet many of the criteria of so-called “complex interventions.” They involve the actions of people, involve a complex chain of steps, are embedded in social systems shaped by context, and are open systems subject to change. [92,93] The following are examples of methods from the evaluation sciences that may be useful for understanding the effects of complex interventions such as continuing HPE programs: the development of program theories, the use of logic models, and the determination of types of outcomes—and their relationship to one another. [36] Though others have long identified the need for sequential or staged evaluation for continuing HPE, [94] the field has been slow to move in this direction. However, there is evidence that this trend may be changing. Moore and colleagues’ recent renewal of a previous outcomes-based framework categorizes outcome types as “summative,” “performance,” and “impact.” [28] Their perspective aligns well with the effectiveness and impact outcomes commonly used in the evaluative sciences but, importantly, overlooks implementation outcomes, which are vital for the evaluation of complex interventions. [95] Knowledge from the evaluative sciences, which has been successfully applied to a wide range of population health programs, including ones involving other forms of education, [96] can and should be considered for continuing HPE programs, including those focused on opioid prescribing. Applying this literature may enhance the reach and effectiveness of these programs, enabling them to improve population health as intended.

Conceptual models that link clinical practice and population health.

Models focused on the behavior changes of individuals may be insufficient for meeting the need to translate continuing HPE to population health outcomes. Even widely used population-focused behavior change models (e.g., socioecological models [97]) may not be sufficient for this purpose, particularly because the effects of continuing HPE programs on patients and populations must be mediated first by clinician behavior change. As health systems have become increasingly concerned with population health outcomes, interest in the development of appropriate conceptual models for connecting clinical practice to population health has grown. [98,99] One promising paradigm is clinical population medicine (CPM), which Orkin defines as “the conscientious, explicit, and judicious application of population health approaches to care for individual patients and design of health care systems.” [100] In discussing population health improvement through a proposed new model of health service delivery in the province of Ontario, Canada, Orkin delineates the notion of a population of individuals within a defined clinical environment (e.g., a health team or a physician’s practice) against the notion of a population in a given community or jurisdiction, regardless of members, affiliation with or access to clinical services. [101] He argues that CPM, besides maintaining its traditional focus on the clinical environment, also focuses on the community population, and he describes various means of achieving this change in orientation. While a comprehensive framework for population-focused continuing HPE is currently wanting, such a framework could serve the educational needs of CPM or similar paradigms that explicitly conceptualize, develop, and examine the connections between clinical practice and population health. Our review emphasizes the ongoing needs to incorporate evidence-based practices into continuing HPE programs and to mitigate industry influence, both of which have already been broadly characterized in the existing literature. While the HPE community does not yet have good evidence that educational best practices (e.g., interactivity, multiple methods, multiple exposures) can drive population health improvements, we do have good evidence that such practices are associated with improvements in participants’ knowledge and performance—and, to some extent, improvements in the health of individual patients. Given, first, the absence of conceptual models for connecting individual patient health outcomes to population health outcomes and, second, the assumption of implicit models that improvements in patient health outcomes lead, in aggregate, to improvements in population health outcomes, we believe it is imperative that continuing HPE follow the best available evidence for driving intended outcomes. Industry participation in medical practice, specifically continuing HPE, has long been problematized for the potential misalignments between commercial interests and patient and population welfare. [102,103] In describing recommended continuing HPE policy that would mitigate this risk, Brennan and colleagues have suggested creating central repositories of industry contributions that would be managed by academic medical centers independently of industry. [104] Indeed, the U.S. FDA attempted a variation of this model in enacting its opioids REMS program. Given the significant deficiencies identified in the program and its ongoing connection to a number of opioid education programs, it is clear that, as Brennan and colleagues state, “other ways of funding CME will have to be identified.” [104] Public funds from local, state, and national governments are the most evident source to support continuing HPE that aligns funder and population health interests. In our review, we identified a few programs that benefitted from the direct engagement of government and public health agencies at both the development and delivery stages. [49,53,61-63,68] These could serve as models for the development of population-focused continuing HPE. Publicly funded development and evaluation grants may also indirectly support opioid-related continuing HPE. Importantly, effective continuing HPE programs must consider sustainability and ultimately attain independence from public support to maintain their long-term relevance and impact.

Limitations

We note several limitations to this review. We did not conduct a systematic search of the gray literature for reports evaluating opioid-related continuing HPE programs. Given that some relevant programs may be conducted by health agencies or similar institutions, evaluations may be published as institutional reports rather than scientific studies. However, considering the more pragmatic nature of such gray literature, we presume that few publications will have more extensively used theories, models, or frameworks than the evaluations captured in this review. Furthermore, while we extracted data on the types of outcomes, we did not formally evaluate the effectiveness or impact of the included programs. Given the heterogeneity of the types of outcomes and outcome measures, a formal synthesis of these outcomes would be challenging. We note, though, that other methods (e.g., a realist synthesis) could be of value to accomplish this. Other recently published reviews [19,20] have already aimed to study the effectiveness and impact of opioid HPE programs especially vis-à-vis other opioid epidemic interventions.

Conclusions

Despite both the recognition of patient- and population-level harms resulting from opioids and also the clear casting of continuing HPE as an appropriate intervention to address these harms, programs have only occasionally followed best practices, have rarely applied educational or other theories, and have seldom included methodologically sound evaluations of effects on patient or population health outcomes. Direct or indirect pharmaceutical industry involvement in the development and evaluation of opioid-related continuing HPE programs persists, which constitutes an important source of potential bias that may impair the ability of these programs to benefit patients and populations. Further conceptual models are urgently needed to guide the development and evaluation of continuing HPE programs, especially in the case of those intended to have population health impacts.

Acknowledgments:

Ms. Elizabeth Uleryk provided professional librarian support for this study, including designing and executing the database search strategy. Ms. Melanie Eliot and Dr. Pamela Sabioni contributed to title and abstract screening. Drs. Whitney Berta, Kate Hodgson and Pamela Sabioni provided helpful comments on earlier drafts of the report.
  84 in total

1.  BEME guide no. 3: systematic searching for evidence in medical education--part 2: constructing searches.

Authors:  Alex Haig; Marshall Dozier
Journal:  Med Teach       Date:  2003-09       Impact factor: 3.650

2.  Opioid Prescribing for Chronic Pain--Achieving the Right Balance through Education.

Authors:  Daniel P Alford
Journal:  N Engl J Med       Date:  2016-01-28       Impact factor: 91.245

3.  Trends in naloxone prescriptions prescribed after implementation of a National Academic Detailing Service in the Veterans Health Administration: A preliminary analysis.

Authors:  Mark Bounthavong; Michael A Harvey; Daina L Wells; Sarah J Popish; Julianne Himstreet; Elizabeth M Oliva; Chad L Kay; Marcos K Lau; Priyanka P Randeria-Noor; Andrea G Phillips; Melissa L D Christopher
Journal:  J Am Pharm Assoc (2003)       Date:  2017-01-11

4.  Interprofessional education in chronic non-cancer pain.

Authors:  Michael Allen; Tanya Macleod; Beverley Zwicker; Marco Chiarot; Carol Critchley
Journal:  J Interprof Care       Date:  2011-03-22       Impact factor: 2.338

5.  Developing and evaluating complex interventions: the new Medical Research Council guidance.

Authors:  Peter Craig; Paul Dieppe; Sally Macintyre; Susan Michie; Irwin Nazareth; Mark Petticrew
Journal:  Int J Nurs Stud       Date:  2012-11-15       Impact factor: 5.837

6.  Using a Logic Model to Assist in the Planning, Implementation, and Evaluation of Educational Programs.

Authors:  Elaine Van Melle
Journal:  Acad Med       Date:  2016-10       Impact factor: 6.893

7.  Clinical Population Medicine: Integrating Clinical Medicine and Population Health in Practice.

Authors:  Aaron M Orkin; Aamir Bharmal; Jenni Cram; Fiona G Kouyoumdjian; Andrew D Pinto; Ross Upshur
Journal:  Ann Fam Med       Date:  2017-09       Impact factor: 5.166

8.  Development and delivery of a pharmacist training program to increase naloxone access in Kentucky.

Authors:  Emma Palmer; Steve Hart; Patricia R Freeman
Journal:  J Am Pharm Assoc (2003)       Date:  2017-02-01

9.  Moving opioid misuse prevention upstream: A pilot study of community pharmacists screening for opioid misuse risk.

Authors:  Mark A Strand; Heidi Eukel; Siri Burck
Journal:  Res Social Adm Pharm       Date:  2018-07-17

10.  Physicians report adopting safer opioid prescribing behaviors after academic detailing intervention.

Authors:  Mary Jo Larson; Cheryl Browne; Ruslan V Nikitin; Nikki R Wooten; Sarah Ball; Rachel Sayko Adams; Kelly Barth
Journal:  Subst Abus       Date:  2018-05-04       Impact factor: 3.716

View more
  1 in total

1.  How the Suboxone Education Programme presented as a solution to risks in the Canadian opioid crisis: a critical discourse analysis.

Authors:  Abhimanyu Sud; Matthew Strang; Daniel Z Buchman; Sheryl Spithoff; Ross E G Upshur; Fiona Webster; Quinn Grundy
Journal:  BMJ Open       Date:  2022-07-12       Impact factor: 3.006

  1 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.