Literature DB >> 34983585

Iowa Implementation for Sustainability Framework.

Laura Cullen1, Kirsten Hanrahan2, Stephanie W Edmonds2, Heather Schacht Reisinger3,4, Michele Wagner2.   

Abstract

BACKGROUND: An application-oriented implementation framework designed for clinicians and based on the Diffusion of Innovations theory included 81 implementation strategies with suggested timing for use within four implementation phases. The purpose of this research was to evaluate and strengthen the framework for clinician use and propose its usefulness in implementation research.
METHODS: A multi-step, iterative approach guided framework revisions. Individuals requesting the use of the framework over the previous 7 years were sent an electronic questionnaire. Evaluation captured framework usability, generalizability, accuracy, and implementation phases for each strategy. Next, nurse leaders who use the framework pile sorted strategies for cultural domain analysis. Last, a panel of five EBP/implementation experts used these data and built consensus to strengthen the framework.
RESULTS: Participants (n = 127/1578; 8% response) were predominately nurses (94%), highly educated (94% Master's or higher), and from across healthcare (52% hospital/system, 31% academia, and 7% community) in the USA (84%). Most (96%) reported at least some experience using the framework and 88% would use the framework again. A 4-point scale (1 = not/disagree to 4 = very/agree) was used. The framework was deemed useful (92%, rating 3-4), easy to use (72%), intuitive (67%), generalizable (100%), flexible and adaptive (100%), with accurate phases (96%), and accurate targets (100%). Participants (n = 51) identified implementation strategy timing within four phases (Cochran's Q); 54 of 81 strategies (66.7%, p < 0.05) were significantly linked to a specific phase; of these, 30 (55.6%) matched the original framework. Next, nurse leaders (n = 23) completed a pile sorting activity. Anthropac software was used to analyze the data and visualize it as a domain map and hierarchical clusters with 10 domains. Lastly, experts used these data and implementation science to refine and specify each of the 75 strategies, identifying phase, domain, actors, and function. Strategy usability, timing, and groupings were used to refine the framework.
CONCLUSION: The Iowa Implementation for Sustainability Framework offers a typology to guide implementation for evidence-based healthcare. This study specifies 75 implementation strategies within four phases and 10 domains and begins to validate the framework. Standard use of strategy names is foundational to compare and understand when implementation strategies are effective, in what dose, for which topics, by whom, and in what context.
© 2021. The Author(s).

Entities:  

Keywords:  Domain mapping; Framework; Healthcare improvement; Implementation; Strategies; Sustainability

Mesh:

Year:  2022        PMID: 34983585      PMCID: PMC8725573          DOI: 10.1186/s13012-021-01157-5

Source DB:  PubMed          Journal:  Implement Sci        ISSN: 1748-5908            Impact factor:   7.327


This study updated a widely used typology with 75 implementation strategies arranged within four phases and 10 domains, to promote effective implementation of evidence-based practices by clinician leaders. This study uniquely identified 10 domains that offer guidance for selecting implementation strategies and a bridge to the strategies’ potential mechanism of action. Implementation strategies included in the framework are unique to the implementation step within the EBP process, avoiding confusion among implementation, EBP process steps, and project management.

Background

Adoption and sustained use of evidence-based practice (EBP) remains elusive [1-6]. The gap between research and practice is well known and a primary focus of implementation science. Adoption and sustainability begin when selecting, timing, operationalizing, and evaluating implementation strategies for use in practice. Clinicians are critical team members who must be brought into the EBP process early, before designing the practice change and implementation planning [7, 8]. Clinicians are challenged to select and use implementation strategies to improve clinical and operational outcomes. A large number of strategies for a variety of clinician users have been compiled [9-19]. Unfortunately, nurses and other clinicians continue to rely heavily on education and information sharing [20] as primary strategies for implementation [9-19], despite the lack of specificity about when or how to best use an informational approach [21]. Foundational to improving the science is the call to create a common language for implementation strategies [22-27]. Implementation strategies specifically target adoption, implementation, sustainability, and scale-up of EBP change [28]. In the Iowa Model (see Fig. 1), and other EBP process models, implementation strategies are explicitly differentiated from the EBP steps related to applying evidence to make decisions about care:
Fig. 1

Iowa Model revised: Evidence-based practice to promote excellence in healthcare

Identifying an issue or opportunity, Stating the purpose, Forming a team, Assembling, appraising, and synthesizing the body of evidence, Designing and piloting the practice change, Integrating and sustaining the practice change, and Dissemination. Iowa Model revised: Evidence-based practice to promote excellence in healthcare Implementation strategies are also differentiated from project management processes, methods, skills, and knowledge which surround and support all EBP steps and are applied to achieve project goals (e.g., creating a charter, organizing, scheduling). To move implementation forward, scientists must take several additional steps [21, 25]: (1) confirm and reconcile these compilations of strategies, (2) define and specify strategies [29-31], (3) identify the link between the strategy and their mechanism of action [32], (4) describe how to bundle strategies, (5) operationalize strategies for use and testing [33], and (6) establish clarity in outcomes influenced by strategies [34-37], all while keeping implementation strategies actionable and feasible for use in clinical settings. To this end, the need and recommendations for specifying elements of individual implementation strategies have been proposed [29–31, 36, 38]. Discrete strategies must be named, conceptually defined, and operationalized with further detail, so that each can be executed, measured, and compared in meaningful ways [31, 33]. Needed among these specifications is guidance matching local organizational needs with the strategy having the best fit to address that need [7, 33, 39, 40]. Further development and specification of an implementation framework to guide clinicians and researchers in strategy selection, while achieving sustainable outcomes, therefore, is needed. An application-oriented implementation framework, the Implementation Strategies for EBP guide (referred to as Iowa implementation framework in this paper; Fig. 2) [10], while designed as a framework for frontline nurses, is relevant for use by all clinicians and for research. The Iowa implementation framework was based on Diffusion of Innovations theory [40] to work synergistically within the Iowa Model, a well-known and frequently used EBP process model developed based on the same theory [8]. We have been leading EBP work since the 1980s and developed the implementation framework to fill a gap in the EBP process. Unlike other implementation strategy typologies, the strategies included in the Iowa implementation framework focus exclusively on the implementation step within the EBP process (see Fig. 1) and are arranged for clinicians as users to lead adoption among clinical teams. Users have requested this framework as a resource for implementing EBP within their healthcare settings, classroom teaching, and workshop presentations, and the most frequent request is for use in academic coursework. The framework includes a list of 81 implementation strategies with suggested timing for use within four implementation phases and targeting clinicians or the health system within which they work. The framework was developed to guide selection of strategies for leading EBP improvements and hardwire system changes. The strategies were identified and organized using implementation literature and decades of practical experience leading EBP improvements. Over the past 7 years, the guide has been requested from our organization’s website over 5000 times from 51 countries and all US states, cited over 100 times, and translated into several languages. The framework has been cited as supporting EBP change and organizational EBP programming, and as supporting the need for research or affirming the phased, yet iterative nature of implementation and the need to address organizational support [41-47]. Strategies have been operationalized with definitions, actionable procedures, and examples, to promote effective use and improve implementation outcomes in a monograph entitled Evidence-Based Practice in Action [33]. Despite the dissemination of these resources, the science has evolved and the need to promote evidence-based healthcare to improve quality and value continues to grow.
Fig. 2

Implementation strategies for evidence-based practice

Implementation strategies for evidence-based practice The purpose of this study was to evaluate and revise the Iowa implementation framework based on user feedback and emerging implementation science. Specific aims were: Determine user perspectives of usefulness of the framework Evaluate the typology of implementation strategies Evaluate the timing of implementation strategies within phases Identify domains for related implementation strategies Use study data, emerging science, and expert consensus to revise the framework The goal was to evaluate and strengthen the framework to be feasible and actionable and improve the effective use of implementation strategies by clinicians and healthcare leaders responsible for promoting delivery of evidence-based healthcare, as well as promote its use among implementation researchers.

Methods

A four-step, iterative mixed-methods approach guided framework evaluation and revisions: first—evaluate framework usability; second—identify and specify strategies; third—identify strategy domains; and fourth—revise and finalize the framework. Participants and procedures for each step are detailed below. Data collection began after the Institution Review Board determined the study was not human subjects’ research.

Evaluate framework usability

Participants

The framework has been available on the healthcare organization’s website for the previous 7 years. The website provides a mechanism to submit an electronic form to request permission for use, which generates an automated email reply with the requested resources. We recruited individuals who had completed the online request for the original Iowa implementation framework. The list was cleaned of duplicates and inactive email addresses. An initial invitation to participate included an imbedded link to the survey. Respondents with unfinished surveys received automated reminder emails at 2 weeks and 1 month.

Procedures

We developed the survey to identify what aspects of the original implementation framework were useful. Participants were invited to complete an online survey using QualtricsXM. The survey had sections exploring respondents’ use of and expertise with using the framework, evaluation of the framework (usability, generalizability, adaptability, accuracy, placement of strategies within the framework, and feedback on implementation phases and the target for each strategy), recommendations for updates and comments, and demographics. In the survey, we provided a list of the 81 names of the implementation strategies and asked respondents to select at least one of four phases in which each strategy should be used: (1) create awareness and interest, (2) build knowledge and commitment, (3) promote action and adoption, and (4) pursue integration and sustained use. The survey also queried respondents about seven potential new implementation strategies identified by the authors (i.e., advisory boards, cultural broker, disincentives, learning collaborative, revise professional roles, social media influencer, and visit other sites). These potential strategies were identified from reports about dissemination and implementation and experience leading EBP changes in the practice setting. Respondents were asked to recommend if any of the new implementation strategies should be added and, if so, the phase and target for each. They were also asked to suggest other new strategies. The survey took approximately 40 min to complete. See Supplemental Appendix A for the full survey. We report frequencies and percentages to describe the demographics of the survey respondents and their beliefs about usability, likeliness to use in the future, and expertise with the framework. Cochran’s Q analyses were used to determine the difference in proportions for participants’ responses to the phases for each of 81 implementation strategies included in the survey. A p-value of 0.05 was used to determine if respondents were significantly more likely to place an implementation strategy in one of the four phases. QualtricsXM [48] was used for descriptive statistics and SAS [49] was used to compute Cochran’s Q analyses.

Identify and specify strategies

An expert consensus panel was made up of three nurse researchers with expertise in the Iowa implementation framework, a nurse leader with clinical experience using and mentoring clinicians in the use of the implementation framework, and an implementation science expert with no prior experience with the Iowa implementation framework. The expert panel reviewed survey data about the strategies in the original framework, newly proposed strategies that survey participants considered and participants suggested additions, and determined 85 strategies to include in the pile sort method based on the strategies in the original framework, removing duplicates and separating discrete strategies and review of the survey results. To build consensus regarding revision, the expert panel reviewed survey results and discussed responses with focused attention on the phases and targets. This panel reviewed the innovation-decision process for individuals and organizations outlined in the Diffusion of Innovations [40] which provided the theoretical underpinnings for the consolidated four phases in the framework. Our process for operationalizing strategies included suggested elements created by Proctor et al. [31]. We reviewed a variety of additional theories, models, and frameworks (TMF). We then considered the description of elements for specifying individual implementation strategies [29–31, 33]. The panel created a crosswalk to outline previously developed strategy elements (i.e., title, phase, focus, definition, benefits, procedure, example, citations) [33, 50] and recommendations for specifying (i.e., name, temporality, action target, definition, outcome, action) [31] and missing specifications [31, 40, 51, 52]. We then created a template of constructs (i.e., name, phase, actor, target, definition, function, action procedure, considerations, clinician example, patient example, and citations) for specifying the implementation strategies included in the framework. Our next step was to select the TMF best matching each specifying element to work synergistically with the Diffusion of Innovations theory [40]. From Proctor et al. [31], we specified elements including leadership, a key contextual factor for implementation [53-55], and the target or “who” or “where” the implementation strategy is directed. Constructs of the Consolidated Framework for Implementation Research [51] were used and provided a crosswalk to the Diffusion of Innovations theory [40, 51], while recognizing the systems perspective needed for implementation planning. We chose the Behavior Change Wheel [52] in order to include function as a specifying element to consider the link between each strategy and the related mechanism of action.

Identify strategy domains

The expert panel recruited nurses in clinical nurse specialist and clinical nurse leader roles at our 800-bed quaternary academic medical center to inform the framework because of their vast clinical experience using it. A standing shared governance council meeting created opportunity to provide a description of the purpose and procedure, and a chance to volunteer. All 26 attendees at this meeting were recruited. Participants represented clinical areas and expertise across the health system and care of a wide variety of patient populations (e.g., ambulatory, emergency and trauma, intensive care, medical-surgical, and pediatrics). Cultural domain analysis [56, 57] was used as a similar methodology to stage 2 concept mapping described in the ERIC protocol [58] to capture additional user input. Participants were provided 85 cards, with one strategy name on each card, to sort into common categories. These cards did not include definitions of the strategies because we wanted the participants to rely on their own knowledge about each strategy. They were asked to put strategies into piles in whatever way made sense to them. The rules were to have more than one card in a pile and not put all cards in one pile. Each participant clipped each of their piles of cards together and placed all their piles in an envelope with an anonymous study identification number. A research assistant transcribed each participant’s pile sorting into a text document. Data were entered into ANTHROPAC, a freely available domain analysis software program, and checks were run to ensure data entry accuracy. We also randomly selected three respondents’ pile sorts for a full review of data entry to check for accuracy. We then analyzed the data using multidimensional scaling to produce a domain map and Johnson’s hierarchical clustering matrix to visually display the clustering [59]. We cross referenced the two-dimensional domain map with Johnson’s hierarchical clustering matrix to define strategies clustering near each other on the map and having the closest association between each strategy in the matrix. For the implementation strategies that were not proximal on the map and did not cluster quantitatively with other strategies, we noted them as outliers for further discussion.

Revise and finalize framework

The same expert panel re-convened to discuss study data and make final recommendations for the framework. First, we reviewed outliers from each of our methodologies. We discussed the three strategies survey respondents suggested adding to the framework (i.e., self-learning, gaming, and plan-do-study-act). We determined based on the literature these suggestions were not implementation strategies and so did not include them in the framework. For the outliers from the analysis of the two-dimensional domain map with Johnson’s hierarchical clustering matrix, we considered if they conceptually belonged in existing domains already identified (decisions described in more detail in the “Results” section). We also discussed implementation strategies that were identified after the survey and not included in the sorting exercise (patient input, patient decision aid, training, facilitator) and decided in which group they belonged. We reviewed the template of constructs for specifying implementation strategies. In preparation, we discussed the TMFs associated with each construct. We used these discussions to create a common understanding among the panel members, while remaining focused on keeping the implementation strategies discrete and actionable. One team member created a grid with each implementation strategy and associated constructs for the group to consider in specifying actor, function, target, and evaluation process measures for each discrete strategy. Group discussion centered on construct definitions, the form and function of implementation strategies, and resulted in a consensus for each strategy. The grid was also used to inform the review of the literature for each strategy and ongoing work to operationalize implementation strategies by updating or adding to a definition, procedure, considerations, and examples.

Results

We had 4059 requests for the original Iowa implementation framework which yielded 1578 active email addresses and 127 (8% response rate) completed the survey. Almost all (98%) were nurses, with 88% having at least a masters, and 84% from the USA (Table 1). The majority used the framework for an EBP change or student assignment (Table 1) and just 4% identified as a novice user (Table 1).
Table 1

Characteristics of survey respondents, N = 127

CharacteristicsN (%)
Discipline
 Nurse (includes nurse practitioner)125 (98.4)
 Physician1 (0.8)
 Others1 (0.8)
Education
 Doctoral70 (55.1)
 Masters50 (39.4)
 Bachelors6 (4.7)
 Others1 (0.8)
Current role
 Educator41 (32.3)
 Administrator21 (16.5)
 Clinician16 (12.6)
 Researcher12 (9.5)
 Student3 (2.4)
 Clinical nurse specialist/nurse leader20 (15.8)
 Others14 (11.0)
Organization type
 Hospital67 (52.3)
 College or university40 (31.3)
 Ambulatory clinic6 (4.7)
 Community3 (2.3)
 Long-term skilled care2 (1.6)
 Others10 (7.8)
Type of hospital
 Community31 (44.3)
 Academic medical center19 (27.1)
 Public (state or federal)12 (17.1)
 Critical access hospital2 (2.9)
 Others6 (8.6)
Location
 USA107 (84.3)
 Asia/Pacific Islands9 (7.1)
 North America (non-US)5 (3.9)
 Middle East4 (3.2)
 Africa1 (0.8)
 Europe1 (0.8)
Purpose for using the implementation modela
 Student paper or assignment38 (18.2)
 Organization EBP project38 (18.2)
 Unit/clinic EBP project34 (16.3)
 Classroom teaching30 (14.4)
 Research/grant18 (8.6)
 Publications/presentations15 (7.2)
 Magnet® submission10 (4.8)
 Others6 (2.9)
 Have not used20 (9.6)
User rated experience with model
 Expert—extensive experience, highly skilled, confident, able to hone in on solutions11(10.9)
 Proficient—skilled, experienced, confident, able to troubleshoot problems14 (13.9)
 Competent—building skill and experience but confident in use39 (38.6)
 Advanced beginner—just beginning, developing experience and confidence33 (32.7)
 Novice—lacking experience and confidence4 (4.0)

aSelect all that apply—respondents could have selected more than one response to this item

Characteristics of survey respondents, N = 127 aSelect all that apply—respondents could have selected more than one response to this item Most respondents found the original implementation framework as useful or very useful for EBP (92%), easy or very easy to use (71.9%), and intuitive or very intuitive for novice users (67%) (Fig. 3). All respondents (100%) agreed or somewhat agreed that the implementation framework is generalizable to different disciplines, settings, and populations and that the framework is flexible and adaptive to be used in conjunction with other EBP process models and frameworks. Most respondents (96.4%) agreed or somewhat agreed that the four phases accurately represented the stages of implementation. Nearly all agreed or somewhat agreed (98.2%) that the implementation framework contains a comprehensive selection of strategies. Lastly, 87.5% of respondents reported they were likely or very likely to use the implementation framework in the future.
Fig. 3

Usability evaluation

Usability evaluation Only 51 (40.2%) of survey respondents completed the section of the survey asking to select in which phase each implementation strategy belonged. Two-thirds of the 81 implementation strategies (n = 54, 66.7%) had a statistically significant agreement (p < 0.5) that the strategy has one primary phase (Table 2). Among those implementation strategies, half matched the phase on the original framework. Of the 54 strategies which respondents selected as having a statistically significant primary phase, 24 strategies (44.4%) did not have agreement between respondents and the original framework. Some of those strategies that were a mismatch between respondents and the original framework were local adaptation and simplify (p = 0.0006), make observable (0.044), and troubleshoot use (p <.0001). Respondents had no significant consensus identifying a primary phase for one-third of the strategies (n = 27, 33.3%). Some of these strategies were unit inservice (p = 0.174), unit newsletter (p = 0.932), posters or postings (p = 0.943), and unit orientation (p = 0.619) (Table 2). The expert panel reviewed the data, definition, and form and function for all strategies and identified phases to determine final placement.
Table 2

Respondents’ selections for which strategies belonged to a phase (n = 51), n (%)

Strategy number/nameCreating awareness and interestBuilding knowledge and commitmentPromoting action and adoptionPursuing integration and sustainabilityp-value
01Action plan32 (62.8)34 (66.7) §42 (82.4) §32 (62.8)0.034
02Actionable and timely data feedback19 (37.3)26 (51.0)34 (66.7) §32 (62.8)0.003
03Advisory boards
04Announcements and broadcasts42 (82.4) §25 (49.0)25 (49.0)22 (43.1)< 0.0001
--Annual report23 (45.1)16 (31.4)16 (31.4)3 (64.7) §0.0008
05Audit and feedback21 (41.2)21 (41.2)31 (60.8)35 (68.6) §0.003
06Audit key indicators21 (41.2)22 (43.1)30 (58.8) §33 (64.7)0.015
07Benchmark data27 (52.9)24 (47.1) §30 (58.8)36 (70.6)0.034
08Case studies29 (56.9)31(60.8) §31 (60.8)21 (41.2)0.056
09Celebrate local unit progress23 (45.1)25 (49.0)30 (58.8)37 (72.6) §0.004
10Change agents (e.g., change champion, core group, opinion leader, thought leader, etc.)32 (62.8)35 (68.6) §34 (66.7) §28 (54.9)0.241
11Checklist19 (37.3)22 (43.1)37 (72.6) §25 (49.0)0.0002
12Clinician input27 (52.9)35 (68.6) §32 (62.8)29 (56.9)0.102
13Cultural broker
--Competency metric for discontinuing training16 (31.4)20 (39.2)18 (35.3)24 (47.1) §0.313
--Continuing education programs20 (39.2) §36 (70.6)26 (51.0)28 (54.9)0.002
14Data collection by clinicians19 (37.3)24 (47.1)31 (60.8) §28 (54.9)0.032
15Decision algorithm
16Demonstrate workflow or decision algorithm18 (35.3)29 (56.9)34 (66.7) §30 (58.8)0.002
17Disincentives
18Disseminate credible evidence with clear implications for practice25 (49.0)22 (43.1) §28 (54.9)32 (62.8)0.10
19Distribute key evidence22 (43.1) §26 (51.0)25 (49.0)29 (56.9)0.419
20Documentation15 (29.4)17 (33.3)25 (49.0) §26 (51.0)0.014
21Education (e.g., live, virtual, or computer-based)27 (52.9)35 (68.6) §28 (54.9)22 (43.1)0.004
22Educational outreach or academic detailing23 (45.1)31 (60.8) §25 (49.0) §22 (43.1)0.116
23“Elevator speech”29 (56.9)26 (51.0)20 (39.2) §15 (29.4)0.001
--Financial incentives16 (31.4)13 (25.5)22 (43.1)25 (49.0) §0.017
24Focus groups for planning change23 (45.1)23 (45.1) §28 (54.9)22 (43.1)0.477
25Gaming
26Gap assessment/gap analysis27 (52.9)26 (51.0) §25 (49.0)18 (35.3)0.135
27Give evaluation results to colleagues18 (35.3)19 (37.3)28 (54.9) §31 (60.8)0.002
28Highlight advantages or anticipated impact27 (52.9) §27 (52.9)26 (51.0)25 (49.0)0.939
29Highlight compatibility16 (31.4) §21 (41.2)23 (45.1)18 (35.3)0.332
30Incentives21 (41.2)20 (39.2)28 (54.9) §25 (49.0)0.215
31Individual performance evaluation15 (29.4)17 (33.3)31 (60.8) §25 (49.0)0.0006
32Individualize data feedback15 (29.4)16 (31.4)29 (56.9)25 (49.0) §0.0016
33Inform organizational leaders21 (41.2)21 (41.2) §25 (49.0)24 (47.1)0.692
34Integrate practice change with other EBP protocols13 (25.5)20 (39.2) §29 (56.9)30 (58.8)0.0003
35Interprofessional discussion and troubleshooting
36Journal club27 (52.9) §32 (62.8)21 (41.2)12 (23.5)< .0001
37Knowledge broker(s)26 (51.0) §30 (58.8)22 (43.1)14 (27.5)0.001
38Learning collaborative
39Link practice change and power holder/stakeholder priorities22 (43.1)28 (54.9) §28 (54.9)25 (49.0)0.373
40Link to patient/family needs and organizational priorities23 (45.1)23 (45.1)28 (54.9) §25 (49.0)0.521
41Local adaptation and simplify14 (27.5)16 (31.4) §28 (54.9)25 (49.0)0.0006
42Make impact observable19 (37.3)21 (41.2) §26 (51.0)29 (56.9)0.044
43Match practice change with resources and equipment12 (23.5)16 (31.4) §27 (52.9)27 (52.9)0.0001
44Mobile “show on the road”23 (45.1) §20 (39.2)21 (41.2)22 (43.1)0.861
--Multidisciplinary discussion and troubleshooting16 (31.4)25 (49.0)30 (58.8) §27 (52.9)0.001
45Non-punitive discussion of results13 (25.5)15 (29.4)25 (49.0) §26 (51.0)0.001
46Patient decision aids16 (31.4)23 (45.1)28 (54.9) §22 (43.1)0.032
47Patient reminders13 (25.5)16 (31.4)26 (51.0) §21 (41.2)0.005
--Peer influence24 (47.1)24 (47.1)26 (51.0)22 (43.1) §0.738
48Personalize the messages to staff (e.g., reduces work, reduces infection exposure, etc.) based on actual improvement data27 (52.9)27 (52.9)23 (45.1)20 (39.2) §0.322
49Pocket guides19 (37.3)22 (43.1) §29 (56.9)24 (47.1)0.047
50Positive deviance
51Posters and postings/fliers23 (45.1) §25 (49.0)24 (47.1)23 (45.1)0.943
--Present in educational programs25 (49.0)30 (58.8)28 (54.9)18 (35.3) §0.008
52Project responsibility in unit or organizational committee14 (27.5)23 (45.1)27 (52.9)26 (51.0) §0.004
53Provide recognition at the point of care15 (29.4)20 (39.2)25 (49.0) §26 (51.0)0.016
54Public recognition18 (35.3)16 (31.4)22 (43.1)27 (52.9) §0.013
55Publicize new equipment18 (35.3) §16 (31.4)27 (52.9)24 (47.1)0.010
56Reminders or practice prompts16 (31.4)20 (39.2)29 (56.9) §26 (51.0)0.004
57Report into quality improvement program11 (21.6)15 (29.4)24 (47.1) §25 (49.0) §0.0007
58Report progress and updates12 (23.5)14 (27.5)25 (49.0) §28 (54.9)< .0001
59Report to senior leaders11 (21.6)12 (23.5) §22 (43.1) §27 (52.9) §0.0002
60Report within organizational infrastructure14 (27.5)18 (35.3) §23 (45.1)28 (54.9)0.005
60Resource manual or materials (i.e., electronic or hard copy)15 (29.4)24 (47.1) §28 (54.9)24 (47.1)0.011
61Resource materials and quick reference guides17 (33.3)25 (49.0)30 (58.8) §25 (49.0)0.010
62Quick reference guide
63Revise policy, procedure, or protocol11 (21.6)15 (29.4)24 (47.1)28 (54.9) §< .0001
64Revise professional roles
65Role model18 (35.3)23 (45.1)29 (56.9) §27 (52.9)0.008
66Rounding by unit and organizational leadership23 (45.1)21 (41.2)29 (56.9) §25 (49.0)0.154
67Self-learning
68Senior executives’ announcements19 (37.3) §14 (27.5)21 (41.2)22 (43.1)0.207
69Share protocol revisions with clinician that are based on feedback from clinicians, patient, or family13 (25.5)19 (37.3)27 (52.9)22 (43.1) §0.009
70Simplify
71Skill competence11 (21.6)23 (45.1)30 (58.8) §20 (39.2)< .0001
72Slogans and logos25 (49.0) §19 (37.3)21 (41.2)16 (31.4)0.138
73Social media influencer
74Sound bites26 (51.0) §16 (31.4)20 (39.2)16 (31.4)0.036
75Staff meetings25 (49.0) §24 (47.1)26 (51.0)23 (45.1)0.842
76Standing orders9 (17.7)13 (25.5)27 (52.9) §26 (51.0)< .0001
--Strategic plan14 (27.5)15 (29.4)24 (47.1)26 (51.0) §0.004
77Teamwork26 (51.0)29 (56.9) §31 (60.8)29 (56.9)0.459
78Trend results16 (31.4)18 (35.3)27 (52.9)24 (47.1) §0.025
79Troubleshoot use/application9 (17.7)21 (41.2) §28 (54.9)24 (47.1)< .0001
80Troubleshooting at the point of care/bedside8 (15.7)17 (33.3)27 (52.9) §26 (51.0)< .0001
81Try the practice change11 (21.6)17 (33.3)31 (60.8) §21 (41.2)< .0001
82Unit inservices22 (43.1) §27 (52.9)29 (56.9)23 (45.1)0.174
83Unit newsletter25 (49.0) §27 (52.9)25 (49.0)26 (51.0)0.932
84Unit orientation22 (43.1)24 (47.1)24 (47.1) §20 (39.2)0.619
--Update practice reminders18 (35.3)21 (41.2)24 (47.1)28 (54.9) §0.033
85Visit other sites

Bold represents the most frequently selected phase and statistical significance at p < 0.05 using Cochran’s Q test

§Indicates which phase the strategy was listed in the original framework

The p-value examines whether there was a significant difference in responses between the phases. If the strategy does not have a strategy #, that strategy was not used for the card pile sort

Italicized strategies indicate the strategy was not used in the survey

Respondents’ selections for which strategies belonged to a phase (n = 51), n (%) Bold represents the most frequently selected phase and statistical significance at p < 0.05 using Cochran’s Q test §Indicates which phase the strategy was listed in the original framework The p-value examines whether there was a significant difference in responses between the phases. If the strategy does not have a strategy #, that strategy was not used for the card pile sort Italicized strategies indicate the strategy was not used in the survey The template for this step in the study is shown in Table 3. In summary, the expert panel ended this step with 75 implementation strategies, thus reducing the total number of discrete strategies in the framework. Discussion led to strategies being unbundled (i.e., change agents became knowledge broker, opinion leader, change champion, etc.; posters and postings/flyers became two strategies—poster and flyer), simplified (incentive, financial incentive, and disincentive became incentive), and duplicates or redundancies being eliminated (i.e., audit, feedback, audit and feedback, and individual data feedback were revised to become audit indicators, data feedback to group, and data feedback to individual). Results of the specifying activity are available (Supplemental Table 4) and include phases into which the expert panel placed the strategies, as well as their domain, function, actor, and target.
Table 3

Implementation strategy specifications. This table provides descriptions for implementation strategy specifications or recommended specifying [31]

Implementation Strategy Specifications (Related Theory)
Name
Addresses what to call each implementation strategy to be able to identify and communicate those selected.
Phase(s) [10, 33, 40]
Addresses when to use each implementation strategy(ies).
➭ Create Awareness & Interest➭Promote Action & Adoption
➭ Build Knowledge & Commitment➭Pursue Integration & Sustained Use
Additional phases
Addresses additional options for when to use each implementation strategy(gies):
Domain(s) [60, 61]
Adds guidance for which implementation strategy(ies) to include.
▪ Marketing▪ Information▪ Learning
▪ Commitment▪ Change Agents▪ Decision Support
▪ Adaptation▪ Data▪ Organizational Infrastructure
▪ Reinforcement
Definition
Provides detail for what each implementation strategy is addressing. It is not an operational definition (see action).
Function(s) (43)
Describes why the implementation strategy may work.
▪ Education▪ Modeling▪ Coercion
▪ Enablement▪ Incentivization▪ Training
▪ Persuasion▪ Restrictions▪ Environmental restructuring
Actor (30)
Identifies who could best provide this implementation strategy.
▪ Payer▪ Community stakeholders▪ Outside consultants
▪ Clinicians▪ Administrators▪ Implementers (within organization)
▪ Intervention developers▪ Patient
Target (42)
Describes where each implementation strategy will impact.
▪ Intervention/localized protocol▪ Characteristics of individuals▪ Outer setting
▪ Process▪ Inner setting
Action procedure
Provides directions for how to do each implementation strategy to improve their effectiveness.
Considerations
Addresses additional details for how to use each implementation strategy.
Implementation strategy specifications. This table provides descriptions for implementation strategy specifications or recommended specifying [31] Among the attendees of the nurse specialists shared governance council, 26 participated and three were excluded for returning the cards as a single pile, resulting in 23 usable responses. Using the pile sorting methodology, our cultural domain analysis resulted in two visual displays of the nurse specialists group consensus regarding the categorization of implementation strategies (see Figs. 4 and 5). The expert panel used the domain map (Fig, 4) to identify domains of implementations strategies and referred to Johnson’s hierarchical clustering (Fig. 5) to help determine in which domain to include an implementation strategy, when strategies where on the border. This resulted in strategies being clustered into 10 domains. In addition, three of the strategies (i.e., skill competence, performance evaluation, and link to patient need) were outliers and did not fit meaningfully into the domains of either visual display. These were discussed individually.
Fig. 4

Domain mapping of implementation strategies identified by nurse leaders

Fig. 5

Domains of implementation strategies from Johnson’s hierarchical clustering

Domain mapping of implementation strategies identified by nurse leaders Domains of implementation strategies from Johnson’s hierarchical clustering After the ten domains were determined, the expert panel independently reviewed and suggested labels for each domain. The expert panel considered the commonalities of the strategies and how the participants would think about each strategy and put them into action. Through the consensus process, we named the domains as follows: Marketing, Information, Learning, Commitment, Change Agents, Decision Support, Adaptation, Data, Organizational Infrastructure, and Reinforcement. Next, the 75 strategies were placed vertically in the primary phase of the implementation framework (identified in step 2), while keeping them horizontally within their domains. The result was a visual cascade of implementation strategies, within the four phases by domain. For the three miscellaneous (outlier) and four added, unsorted implementation strategies, the expert panel individually reviewed the strategies and placed them in a related domain. They then met and formed a consensus regarding the domain for each strategy. Skill competence and training were placed into the Learning domain. Facilitator was placed in the Change Agent domain. Link to patient needs was placed in the Commitment domain. Patient decision aid was placed in the Decision Support domain. Patient input was placed in the Adaptation domain and performance evaluation was placed in the Organizational Infrastructure domain. Finally, we discussed the design of the framework from a user perspective. To convey the iterative nature of the implementation step and the reality that team members are in different places relative to adoption (e.g., late adopter, new hires) while the team is making forward progress, we opted for arrows going forward through phases with an option to go back to other phases reflecting midstream corrections. We finalized the primary and other useful phases for each strategy. Each strategy was placed in the primary phase with superscripts for other phases in which a strategy could be useful. Strategies were clustered within their domain, as rows, when placed within their primary phase, as columns. To make strategies with at least some empirical evidence in healthcare stand out as potentially being more effective (as determined by literature reviews and content expertise by the first author), we used a bold type face. The expert panel then reviewed the two visualizations of the cultural domain analysis, finalized the implementation strategies in each domain, and labeled the domains. Lastly, we finalized the framework display and assigned a more descriptive name—Iowa Implementation for Sustainability Framework (Fig. 6).
Fig. 6

The Iowa Implementation for Sustainability Framework (IISF)

The Iowa Implementation for Sustainability Framework (IISF)

Discussion

The Iowa Implementation for Sustainability Frameworks (IISF) was developed based on user input and designed for clinicians to make implementation actionable, while promoting its use among researchers as a clinician-developed, clinician-facing framework. The IISF was originally designed based on the Diffusion of Innovations theory [40] and continues to reflect those theoretical underpinnings, while being adapted for application in practice. An important advantage is the IISF inclusion of the 75 discrete implementation strategies which offer a variety of options when planning implementation to avoid over-reliance on education and information sharing [13, 21, 62–64]. The framework phases and domains provide guidance on when to use strategies and suggest how to bundle them by crossing domains to address the cognitive, motivational, psychomotor, social, and organizational influences. Implementation is a journey, not an event, and recognized to occur in phases over time [10, 40, 65–67]. The four phases originally adapted from the Diffusion of Innovations theory [40] include create awareness and interest, build knowledge and commitment, promote action and adoption, and finally pursue integration and sustained use. These four phases were confirmed by clinicians using the framework. Phases for implementation differ from steps in the EBP process, though these have been confused [14, 15, 68]. Basic EBP process steps include identification of a need, determining fit of the identified need within the local context, creating a team, use of best evidence to determine practice recommendation, designing the EBP change, implementation, evaluation, sustain the change, and dissemination [8]. Phases are part of the implementation step, but often in reality of a non-linear practice change, both EBP process steps and implementation phases overlap (e.g., cycling between evaluation and implementation). As implementation science develops, clarity is required in the language used to name implementation strategies and determinants used in TMF. Implementation strategies and frameworks should focus specifically on the implementation step within the EBP process and avoid mixing language relating implementation strategies (e.g., academic detailing, change agents, audit, and feedback), with project management (e.g., develop relationships, planning, organize team meetings, access funding, resource sharing agreement) and steps in the EBP process (e.g., create a clear purpose statement, create a team, identify practice recommendations, pilot/small test of change, implementation planning, evaluate). Our process for specifying also kept in mind the need to differentiate the use of effective implementation strategies from the evidence-based intervention, implementation process, and implementation outcomes [19, 27, 34, 35, 69–71]. This study and the 75 implementation strategies included in the IISF advance previous typologies by separating EBP process steps from implementation strategies. Users were asked to identify the primary phase for each implementation strategy and were able to identify a primary phase for only two-thirds of the strategies. This may reflect the implementation strategy being useful and effective across phases (e.g., relative advantage). This may also reflect that users lack understanding of strategy names or inexperience with using the full range of implementation strategies available, despite these strategies having good evidence of effectiveness (e.g., academic detailing). Other strategies tend to be commonly used across the breadth of implementation phases and may represent a lack of guidance from and underdevelopment of the related mechanism of action. These strategies may be overused, and their effectiveness limited by relying too heavily on passive information sharing (e.g., distribute credible evidence, inservice, poster). These findings highlight the need for clear guidance and expertise to know how to bundle implementation strategies for a comprehensive, yet efficient implementation plan. Grouping of strategies into domains offers an intermediary and explanatory step which may help to identify unique mechanisms of action [9, 72], and the associated implementation outcome [34, 37]. The domains identified in this framework may create a bridge to specifying implementation strategies and guiding use. These domains offer an added benefit when reporting results of systematic reviews of implementation strategies. Implementers and clinician leaders will benefit from the added explanatory details provided by the ten domains if research can evaluate and support the use of domains. Understanding how implementation strategies work and the mechanism of action facilitates efficient and effective selection of implementation strategies, identifying strategy bundles, and importantly matching strategies to local setting’s contextual needs [19, 21, 73, 74]. Strategies must be selected to address local context needs [39] that cut across learning needs beyond cognitive and psychomotor skill development to build the EBP change into the team’s workflow [16, 75] and create new practice habits that can be sustained [76, 77]. In work based on the initial Iowa implementation framework, we identified the first known compilation of specified strategies for users [33, 50]. The current study and the IISF better align the names and specification of the strategies with other TMF. In addition to the elements specified by this study, we are currently reviewing literature and operationalizing each strategy by updating or adding to a definition, procedure, considerations, and examples for a full compilation of the IISF strategies in the next edition for the Evidence-Based Practice in Action monograph. Still needed when specifying strategies is the development of methods to match local context with implementation strategy selection. Including the organizational perspective in adoption involves matching the local setting and the practice change [40]. Implementation science has focused on building organizational capacity [78-80]. To date, assessments of organizational needs and capacity building fall short of adding the final link required between the intervention developer and implementer with the clinician and patient as end users. The use of baseline evaluation of local data for implementation can provide guidance for matching implementation strategies with the local needs [8, 39]. Unlike implementation research frameworks for specifying outcome measures (e.g., RE-AIM) [31, 35, 81], the unique KABOB framework (i.e., Knowledge, Attitude, Behavior, Outcome, and Balancing Measures) for evaluation offers direct guidance from local assessment to select matching implementation strategies [39]. Research is needed to establish how best to create that critical match to local needs. Actionable guidance and procedures to increase fidelity in use and reporting are needed. The expert panel took the first steps to develop this specification of implementation strategies for application in practice [29-31]. Discrete strategies require a name and conceptual definition and must be operationalized with further detail, so that each can be executed, measured, and compared in meaningful ways [31]. The strategy names were selected to be brief and, when possible, consistent with common use in practice and the implementation literature [31, 36, 38].

Limitations

This study was needed to address updates in implementation science. The IISF is intended to be useful for clinicians across a range of expertise. Yet, survey respondents were highly educated and may not reflect novice users’ perspectives. While feedback from more experienced users may have been well informed, additional insights from novice users would strengthen the usefulness of the IISF as a resource for EBP. Another limitation is nurse leaders who completed the pile sorts and the expert consensus panel were all from the same institution; however, the strength of this approach was the depth of knowledge of the framework and it was representative of the culture of the institution. In addition, the domains created from the cultural domain analysis were not returned to the committee members for review. The COVID pandemic and surge needs for patient care took precedent. Thus, their confirmation of those findings is missing. Participants were almost exclusively nurses. While additional evaluation by other interprofessional team members is warranted, we believe a study on the perspectives of nurses working at the point of care is a strength. Nurses who provide direct care to patients are the ones who need implementation strategies in their work and thus implementation science should be informed by them. Additionally, our response rate (8%) was low; thus, those who responded may have more favorable perceptions than those who did not respond. Those who requested the original implementation guide may also not have used the guide so they may not have had an opinion. Also missing from the framework is the patient perspective [38]. The next step is co-design with engagement and empowering patient’s role in implementation [82]. To that end, additional work is needed to operationalize implementation strategies for patients to be drivers of EBP improvements and implementation strategy use. Currently, the IISF has some strategies focused on patients (i.e., patient decision aid) and the team is identifying patient-driven examples (e.g., asthma action plan); this continues to be a gap for implementation science to fill. Preliminary anecdotal feedback has been positive when the IISF has been reported in research presentation and workshops using the IISF. Early input indicates the framework is more visually appealing, less dense, and easy to interpret after minimal orientation to phases, domains, and implementation strategies [60, 61]. Further evaluation and validation are warranted. Use will be tracked through the same online automated permission request portal [83], creating an opportunity for future research.

Conclusion

This study reduces gaps between conduct of research and application in EBP critical to quality healthcare. Through a multi-step iterative process, this study evaluates and begins to validate and strengthen the previous Iowa implementation framework to become the Iowa Implementation for Sustainability Framework (IISF). The four implementation phases and 75 distinctive implementation strategies in the IISF were identified. The targets in the original Iowa implementation framework were focused on people and systems; the structure of the IISF shifted to include 10 newly identified domains that are indicative of the mechanism of action. The external validity of these domains has yet to be established. Standard use of strategy names is foundational to compare and understand what implementation strategies are being used and when they are effective, in what dose, for which topics, by whom, and in what context. Implementation and the Iowa implementation framework have evolved over three decades as a step in the Iowa Model of EBP which has primarily been used by nurses but has broad applications for any interprofessional team [8, 10, 84–86]. The IISF is now more effective as a framework because it (1) offers implementation strategy names that are discrete and actionable, while remaining distinct to the implementation step within the EBP process; and (2) provides a structure that is usable by a novice or expert and offers a typology to guide nurses, interprofessional teams, and researchers as they strive to efficiently implement and sustain evidence-based improvements in healthcare. This study builds upon an implementation framework widely used in nursing. There is a need to bridge the silos currently reflected in implementation research. Nursing has been on the forefront and that early research is largely missing from current developments, resulting in reinvention and rediscovery when the needs in healthcare have become even more pressing. We challenge public health and medicine to look at nursing research and TMF, and we challenge nursing to look at the implementation science literature when trying to select strategies for change. It is time for nursing to model this interprofessional and international work and bring the nursing perspective into presentations at international implementation and dissemination conferences and publish in interprofessional and international journals. The IISF is designed to be application oriented and lead to effective implementation planning using actionable implementation strategies. Establishing standard and descriptive names for discrete strategies is warranted to promote comparison and determine core elements of associated action procedures. Additional work is needed to determine if these domains guide bundling of implementation strategies to improve implementation outcomes—adoption, sustained use, and cost. Additional file 1: Survey: Evaluation of the Evidence-Based Implementation Model. Additional file 2: Table 4. Implementation strategies selected by expert panel on second step of study.
  71 in total

Review 1.  Evaluating the public health impact of health promotion interventions: the RE-AIM framework.

Authors:  R E Glasgow; T M Vogt; S M Boles
Journal:  Am J Public Health       Date:  1999-09       Impact factor: 9.308

2.  In search of how people change. Applications to addictive behaviors.

Authors:  J O Prochaska; C C DiClemente; J C Norcross
Journal:  Am Psychol       Date:  1992-09

3.  The effect of a translating research into practice intervention to promote use of evidence-based fall prevention interventions in hospitalized adults: A prospective pre-post implementation study in the U.S.

Authors:  Marita G Titler; Paul Conlon; Margaret A Reynolds; Robert Ripley; Alex Tsodikov; Deleise S Wilson; Mary Montie
Journal:  Appl Nurs Res       Date:  2015-12-21       Impact factor: 2.257

4.  What patients do and their impact on implementation.

Authors:  Annette Boaz; Glenn Robert; Louise Locock; Gordon Sturmey; Melanie Gager; Sofia Vougioukalou; Sue Ziebland; Jonathan Fielden
Journal:  J Health Organ Manag       Date:  2016

5.  Strategies for the successful implementation of disinfecting port protectors to reduce CLABSI in a large tertiary care teaching hospital.

Authors:  Cole Beeler; Denise Kerley; Carmen Davis; Dana Hazen; William Snyderman; Kathy Lyons; Josh Sadowski; Jennifer Sweeney; Lana Dbeibo; Kristen Kelley; Douglas H Webb
Journal:  Am J Infect Control       Date:  2019-07-16       Impact factor: 2.918

Review 6.  A compilation of strategies for implementing clinical innovations in health and mental health.

Authors:  Byron J Powell; J Curtis McMillen; Enola K Proctor; Christopher R Carpenter; Richard T Griffey; Alicia C Bunger; Joseph E Glass; Jennifer L York
Journal:  Med Care Res Rev       Date:  2011-12-26       Impact factor: 3.929

Review 7.  Meausures of organizational characteristics associated with adoption and/or implementation of innovations: A systematic review.

Authors:  Jennifer D Allen; Samuel D Towne; Annette E Maxwell; Lisa DiMartino; Bryan Leyva; Deborah J Bowen; Laura Linnan; Bryan J Weiner
Journal:  BMC Health Serv Res       Date:  2017-08-23       Impact factor: 2.655

8.  What do cost-effective health behaviour-change interventions contain? A comparison of six domains.

Authors:  Emma Beard; Robert West; Fabiana Lorencatto; Ben Gardner; Susan Michie; Lesley Owens; Lion Shahab
Journal:  PLoS One       Date:  2019-04-17       Impact factor: 3.240

9.  Enhancing the Impact of Implementation Strategies in Healthcare: A Research Agenda.

Authors:  Byron J Powell; Maria E Fernandez; Nathaniel J Williams; Gregory A Aarons; Rinad S Beidas; Cara C Lewis; Sheena M McHugh; Bryan J Weiner
Journal:  Front Public Health       Date:  2019-01-22

10.  Expert Recommendations for Implementing Change (ERIC): protocol for a mixed methods study.

Authors:  Thomas J Waltz; Byron J Powell; Matthew J Chinman; Jeffrey L Smith; Monica M Matthieu; Enola K Proctor; Laura J Damschroder; JoAnn E Kirchner
Journal:  Implement Sci       Date:  2014-03-26       Impact factor: 7.327

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.