Literature DB >> 32299461

A systematic review of empirical studies examining mechanisms of implementation in health.

Cara C Lewis1,2,3, Meredith R Boyd4, Callie Walsh-Bailey5,6, Aaron R Lyon7, Rinad Beidas8, Brian Mittman9, Gregory A Aarons10, Bryan J Weiner11, David A Chambers12.   

Abstract

BACKGROUND: Understanding the mechanisms of implementation strategies (i.e., the processes by which strategies produce desired effects) is important for research to understand why a strategy did or did not achieve its intended effect, and it is important for practice to ensure strategies are designed and selected to directly target determinants or barriers. This study is a systematic review to characterize how mechanisms are conceptualized and measured, how they are studied and evaluated, and how much evidence exists for specific mechanisms.
METHODS: We systematically searched PubMed and CINAHL Plus for implementation studies published between January 1990 and August 2018 that included the terms "mechanism," "mediator," or "moderator." Two authors independently reviewed title and abstracts and then full texts for fit with our inclusion criteria of empirical studies of implementation in health care contexts. Authors extracted data regarding general study information, methods, results, and study design and mechanisms-specific information. Authors used the Mixed Methods Appraisal Tool to assess study quality.
RESULTS: Search strategies produced 2277 articles, of which 183 were included for full text review. From these we included for data extraction 39 articles plus an additional seven articles were hand-entered from only other review of implementation mechanisms (total = 46 included articles). Most included studies employed quantitative methods (73.9%), while 10.9% were qualitative and 15.2% were mixed methods. Nine unique versions of models testing mechanisms emerged. Fifty-three percent of the studies met half or fewer of the quality indicators. The majority of studies (84.8%) only met three or fewer of the seven criteria stipulated for establishing mechanisms.
CONCLUSIONS: Researchers have undertaken a multitude of approaches to pursue mechanistic implementation research, but our review revealed substantive conceptual, methodological, and measurement issues that must be addressed in order to advance this critical research agenda. To move the field forward, there is need for greater precision to achieve conceptual clarity, attempts to generate testable hypotheses about how and why variables are related, and use of concrete behavioral indicators of proximal outcomes in the case of quantitative research and more directed inquiry in the case of qualitative research.

Entities:  

Keywords:  Causal model; Determinant; Implementation; Mechanism; Mediator; Moderator; Theory

Mesh:

Year:  2020        PMID: 32299461      PMCID: PMC7164241          DOI: 10.1186/s13012-020-00983-3

Source DB:  PubMed          Journal:  Implement Sci        ISSN: 1748-5908            Impact factor:   7.327


This is the first systematic review of implementation mechanisms across health that assesses the quality of studies and the extent to which they offer evidence in support of establishing mechanisms of implementation. We summarize nine examples of models for evaluating mechanisms. We offer conceptual, theoretical, and methodological guidance for the field to contribute to the study of implementation mechanisms.

Background

Implementation research is the scientific evaluation of strategies or methods used to support the integration of evidence-based practices or programs (EBPs) into healthcare settings to enhance the quality and effectiveness of services [1]. There is mounting evidence that multi-faceted or blended implementation strategies are necessary (i.e., a discrete strategy is insufficient) [2, 3], but we have a poor understanding of how and why these strategies work. Mechanistic research in implementation science is in an early phase of development. As of 2016, there were only nine studies included in one systematic review of implementation mediators1 specific to the field of mental health. Mediators are an intervening variable that may statistically account for the relation between an implementation strategy and outcome. We define the term mechanism as a process or event through which an implementation strategy operates to affect one or more implementation outcomes (see Table 1 for key terms and definitions used throughout this manuscript). Mechanisms offer causal pathways explaining how strategies operate to achieve desired outcomes, like changes in care delivery. Some researchers conflate moderators, mediators, and mechanisms [6], using the terms interchangeably [7]. Mediators and moderators can point toward mechanisms, but they are not all mechanisms as they typically are insufficient to explain exactly how change came about.
Table 1

Terms and definitions

TermDefinition
MechanismProcess or event through which an implementation strategy operates to affect desired implementation outcomes.
PreconditionFactor that is necessary in order for an implementation mechanism to be activated.
StrategyMethods used to promote the implementation of an evidence-based practice or program
DeterminantAlso commonly referred to as “barriers” and “facilitators,” a factor that enables or hinders the implementation strategy from eliciting the desired effect.
MediatorIntervening variable that may account for the relationship between the implementation strategy and the implementation outcome.
ModeratorFactor that increase or decrease the level of influence of an implementation strategy.
Proximal outcomeThe product of the implementation strategy that is realized because of its specific mechanism of action, the most immediate, observable outcome in the causal pathway.
Distal outcomeOutcome that the implementation processes is ultimately intended to achieve, not the most immediate outcome in the causal pathway.
Terms and definitions In addition to these linguistic inconsistencies and lack of conceptual clarity, there is little attention paid to the criteria for establishing a mechanistic relation. Originally, Bradford-Hill [8], and more recently Kazdin offers [4] at least seven criteria for establishing mechanisms of psychosocial treatments that are equally relevant to implementation strategies: strong association, specificity, consistency, experimental manipulation, timeline, gradient, plausibility, or coherence (see Table 2 for definitions). Taken together, these criteria can guide study designs for building the case for mechanisms over time. In lieu of such criteria, disparate models and approaches for investigating mechanisms are likely to exist that make synthesizing findings across studies quite challenging. Consequently, the assumption that more strategies will achieve better results is likely to remain, driving costly and imprecise approaches to implementation.
Table 2

Kazdin’s criteria for establishing a mechanism

TermDefinition
Strong associationAssociation between implementation strategy and mechanism AND between mechanism and behavior change.
SpecificityOne plausible construct accounts for behavior change.
ConsistencyReplication of observed results across studies, samples, and conditions.
Experimental manipulationDirect manipulation of implementation strategy or proposed mediator or mechanism shows impact on outcomes.
TimelineCauses and mediators temporally precede effects and outcomes.
GradientDose response relationship between mediator and outcome.
Plausibility or coherenceExplanation invokes other info and steps in a process-outcome relation that are reasonable or supported by other research.
Kazdin’s criteria for establishing a mechanism Understanding the mechanisms of implementation strategies, defined as the processes by which strategies produce desired effects [4, 8], is important for both research and practice. For research, it is important to specify and examine mechanisms of implementation strategies, especially in the case of null studies, in order to understand why a strategy did or did not achieve its intended effect. For practice, it is crucial to understand mechanisms so that strategies are designed and selected to directly target implementation determinants or barriers. In the absence of this kind of intentional, a priori matching (i.e., strategy targets determinant), it is possible that the “wrong” (or perhaps less potent) strategy will be deployed. This phenomenon of mismatched strategies and determinants was quite prevalent among the 22 tailored improvement intervention studies included in Bosch et al.’s [9] multiple case study analysis. Upon examining the timing of determinant identification and the degree to which included studies informed tailoring of the type versus the content of the strategies using determinant information, they discovered frequent determinant-strategy mismatch across levels of analysis (e.g., clinician-level strategies were used to address barriers that were at the organizational level) [9]. Perhaps what is missing is a clear articulation of implementation mechanisms to inform determinant-strategy matching. We argue that, ultimately, knowledge of mechanisms would help to create a more rational, efficient bundle of implementation strategies that fit specific contextual challenges. Via a systematic review, we sought to understand how mechanisms are conceptualized and measured, how they are studied (by characterizing the wide array of models and designs used to evaluate mechanisms) and evaluated (by applying Kazdin’s seven criteria), and how much evidence exists for specific mechanisms. In doing so, we offer a rich characterization of the current state of the evidence. In reflecting on this evidence, we provide recommendations for future research to optimize their contributions to mechanistic implementation science.

Methods

Search protocol

The databases, PubMed and CINAHL Plus, were chosen because of their extensive collection of over 32 million combined citations of medical, nursing and allied health, and life science journals, as well as inclusiveness of international publications. We searched both databases in August 2018 for empirical studies published between January 1990 and August 2018 testing candidate mechanisms of implementation strategies. This starting date was selected given that the concept of evidence-based practice/evidence-based treatment/evidence-based medicine first gained prominence in the 1990’s with the field of implementation science following in response to a growing consciousness of the research to practice gap [10, 11]. The search terms were based on input from all authors who represent a variety of methodological and content expertise related to implementation science and reviewed by a librarian; see Table 3 for all search terms. The search string consisted of three levels with terms reflecting (1) implementation science, (2) evidence-based practice (EBP), and (3) mechanism. We adopted Kazdin’s [4] definition of mechanisms, which he indicates are the basis of an effect. Due to the diversity of definitions that exist in the literature, the term “mechanism” was supplemented with the terms “mediator” and “moderator” to ensure all relevant studies were collected.
Table 3

Search strategy

Search termsExplanation
Implement* OR disseminate* OR “knowledge translation”These terms were chosen to target Implementation Science literature.
AND
“empirically supported treatment” OR “evidence-based practice” OR “evidence-based treatment” OR innovation OR guidelineThese terms were chosen to target the implementation evidence-based practices
AND
Mediate* OR moderator OR mechanism*These terms were chosen to target mechanisms explaining the implementation of evidence-based practices
NOT
Biology OR microbiologyThese terms were chosen to exclude mechanistic studies in biology and microbiology
Search strategy

Study inclusion and exclusion criteria

Studies were included if they were considered an empirical implementation study (i.e., original data collection) and statistically tested or qualitatively explored mechanisms, mediators, or moderators. We did not include dissemination studies given the likely substantive differences between strategies, mechanisms, and outcomes. Specifically, we align with the distinction made between dissemination and implementation put forth by the National Institutes of Health program announcement for Dissemination and Implementation Research in Health that describes dissemination as involving distribution of evidence to a target audience (i.e., communication of evidence) and implementation as involving use of strategies to integrate evidence into target settings (i.e., use of evidence in practice) [12]. However, the word “dissemination” was included in our search terms because of the tendency of some researchers to use “implementation” and “dissemination” interchangeably. Studies were excluded if they were not an implementation study, used the terms “mediator,” “moderator,” or “mechanism” in a different context (i.e., conflict mediator), did not involve the implementation of an EBP, or were a review, concept paper, or opinion piece rather than original research. All study designs were considered. Only studies in English were assessed. See Additional File 1 for exclusion criteria and definitions. We strategically cast a wide net and limited our exclusions so as to characterize the broad range of empirical studies of implementation mechanisms. Citations generated from the search of PubMed and CINAHL were loaded into EPPI Reviewer 4, an online software program used for conducting literature reviews [13]. Duplicate citations were identified for removal via the duplicate checking function in EPPI and via manual searching. Two independent reviewers (MRB, CWB) screened the first ten citations on title and abstract for inclusion. They then met to clarify inclusion and exclusion criteria with the authorship team, as well as add additional criteria if necessary, and clarify nuances of the inclusion/exclusion coding system (see Additional File 1 for exclusion criteria and definitions). The reviewers met once a week to compare codes and resolve discrepancies through discussion. If discrepancies could not be easily resolved through discussion among the two reviewers, the first author (CCL) made a final determination. During full text review, additional exclusion coding was applied for criteria that could not be discerned from the abstract; articles were excluded at this phase if they only mentioned the study of mechanisms in the discussion or future directions. Seven studies from the previous systematic review of implementation mechanisms [14] were added to our study for data extraction; these studies likely did not appear in our review due to differences in the search strategy in that the review undertaken by Williams hand searched published reviews of implementation strategies in mental health.

Study quality assessment

The methodological quality of included studies was assessed using the Mixed Methods Appraisal Tool (MMAT-version 2018) [15]. This tool has been utilized in over three dozen systematic reviews in the health sciences. The MMAT includes two initial screening criteria that assess for the articulation of a clear research question/objective and for the appropriateness of the data collected to address the research question. Studies must receive a “yes” in order to be included. The tool contains a subset of questions to assess for quality for each study type—qualitative, quantitative, and mixed methods. Table 4 summarizes the questions by which studies were evaluated, such as participant recruitment and relevance and quality of measures. Per the established approach to MMAT application, a series of four questions specific to each study design type are assigned a dichotomous “yes” or “no” answer. Studies receive 25 percentage points for each “yes” response. Higher percentages reflect higher quality, with 100% indicating all quality criteria were met. The MMAT was applied by the third author (CWB). The first author (CCL) checked the first 15% of included studies and, based on reaching 100% agreement on the application of the rating criteria, the primary reviewer then applied the tool independently to the remaining studies.
Table 4

MMAT

Bardosh et al. 2017 [16]Brewster et al. 2015 [17]Carrera et al. 2015 [18]Frykman et al. 2014 [19]Wiener-Ogilvie et al. 2008 [20]Atkins et al. 2008 [21]Baer et al. 2009 [22]Bonetti et al. 2005 [23]Garner et al. 2011 [24]Glisson et al. 2010 [25]Holth et al. 2011 [26]Lee et al. 2018 [27]Lochman et al. 2009 [28]Rapkin et al. 2017 [29]Rohrbach et al. 1993 [30]Seys et al. 2018 [31]Williams et al. 2014 [32]Williams et al. 2017 [33]
1. Qualitative
Data sources relevant?YYYYYN/AN/AN/AN/AN/AN/AN/AN/AN/AN/AN/AN/AN/A
Data analysis process relevant?YYYYYN/AN/AN/AN/AN/AN/AN/AN/AN/AN/AN/AN/AN/A
Findings relate to context?YYYYYN/AN/AN/AN/AN/AN/AN/AN/AN/AN/AN/AN/AN/A
Findings relate to researchers' influence?NNNYNN/AN/AN/AN/AN/AN/AN/AN/AN/AN/AN/AN/AN/A
2. Quantitative randomized
Clear description of the randomization?N/AN/AN/AN/AN/ANNYYNNYNYNYNY
Clear description of allocation or concealment?N/AN/AN/AN/AN/ANNNNYNNYNYNNN
Complete outcome data?N/AN/AN/AN/AN/AYYYYYYYYYNYNY
Low withdrawal/drop-out?N/AN/AN/AN/AN/AYYNYNNNYNYNYY
Total score (%)7575751007550505075502550755050502525
Aarons et al. 2009 [34]Becker et al. 2016 [35]Beenstock et al. 2012 [36]Beets et al. 2008 [37]Bonetti et al. 2009 [38]Chou et al. 2011 [39]Cummings et al. 2017 [40]David and Schiff 2017 [41]Edmunds et al. 2014 [42]Gnich et al. 2018 [43]Guerrero et al. 2018 [44]Huis et al. 2013 [45]Little et al. 2015 [46]Llasus et al. 2014 [47]Nelson and Steele 2007 [48]Potthoff et al. 2017 [49]Presseau et al. 2016 [50]Simmonds et al. 2012 [51]Stockdale et al. 2018 [52]Wanless et al. 2015 [53]
3. Quantitative - non-randomized
Recruitment minimizes selection bias?YNNYYYNNYYYYYNYYYYYY
Measurements appropriate?YYYYYYYYYYYYYYYYYYYY
Comparable groups or control for differences?YYYNNYNYYYYNYYYNYYYY
Complete outcome data, acceptable response rate, or acceptable follow-up rate?NNNNNNNYNNNYYNNNNYNN
Total score (%)755050505075257575757575100507550751007575
Armson et al. 2018 [54]Birken et al. 2015 [55]Kauth et al. 2010 [56]Lukas et al. 2009 [57]Panzano et al. 2012 [58]Rangachari et al. 2015 [59]Shrubsole et al. 2018 [60]
1. Qualitative
Data sources relevant?YYYYYYY
Data analysis process relevant?YYYNNYY
Findings relate to context?YYYYYYY
Findings relate to researchers' influence?NNNNNNN
2. Quantitative randomized
Clear description of the randomization?N/AN/ANN/AN/AN/AY
Clear description of allocation or concealment?N/AN/ANN/AN/AN/AY
Complete outcome data?N/AN/AYN/AN/AN/AY
Low withdrawal/drop-out?N/AN/AYN/AN/AN/AN
3. Quantitative non-randomized
Recruitment minimizes selection bias?YYN/AYNYN/A
Measurements appropriate?YYN/AYYYN/A
Comparable groups or control for differences?NNN/ANNNN/A
Complete outcome data, acceptable response rate, or acceptable follow-up rate?YNN/ANYYN/A
4. Mixed methods
Research design relevant?YYYNNYY
Integration of qualitative and quantitative data relevant?YYYYYYY
Appropriate consideration given to limitations associated with integration?YYNNNNN
Total score (%)75505025257575
MMAT

Data extraction and synthesis

Data extraction focused on several categories: study information/ background (i.e., country, setting, and sample), methods (i.e., theories that informed study, measures used, study design, analyses used, proposed mediation model), results (i.e., statistical relations between proposed variables of the mediation model tested), and criteria for establishing mechanisms (based on the seven listed in Table 2 [4];). All authors contributed to the development of data extraction categories that were applied to the full text of included studies. One reviewer (MRB) independently extracted relevant data and the other reviewer (CWB) checked the results for accuracy, with the first author (CCL) addressing any discrepancies or questions, consistent with the approach of other systematic reviews [61]. Extracted text demonstrating evidence of study meeting (or not meeting) each criterion for establishing a mechanism was further independently coded as “1” reflecting “criterion met” or “0” reflecting “criterion not met” by MRB and checked by CWB. Again, discrepancies and questions were resolved by the first author (CCL). Technically, mechanisms were considered “established” if all criteria were met. See Additional File 2 for PRISMA checklist for this study.

Results

The search of PubMed and CINAHL Plus yielded 2277 studies for title and abstract screening, of which 447 were duplicates, and 183 moved on to full-text review for eligibility. Excluded studies were most frequently eliminated due to the use of mechanism in a different context (i.e., to refer to a process, technique, or system for achieving results of something other than implementation strategies). After full article review, 39 studies were deemed suitable for inclusion in this review. Two of the included studies appeared in the only other systematic review of implementation mechanisms in mental health settings [14]. For consistency and comprehensiveness, the remaining seven studies from the previously published review were added to the current systematic review for a total of 46 studies.2 See Fig. 1 for a PRISMA Flowchart of the screening process and results.
Fig. 1

Mechanisms of Implementation Systematic Review PRISMA Flowchart

Mechanisms of Implementation Systematic Review PRISMA Flowchart

Study characteristics

Setting, sampling, and interventions

Table 5 illustrates the characteristics of the 46 included studies. Twenty-five studies (54.3%) were completed in the USA, while 21 studies were conducted in other countries (e.g., Australia, Canada, Netherlands, UK). Settings were widely variable; studies occurred in behavioral health (e.g., community mental health, residential facilities) or substance abuse facilities most frequently (21.7%), followed by hospitals (15.2%), multiple sites across a health care system (15.2%), schools (15.2%), primary care clinics (10.9%), and Veteran’s Affairs facilities (8.7%). Sampling occurred at multiple ecological levels, including patients (17.4%), providers (65.2%), and organizations (43.5%). Seventeen (40.0%) studies examined the implementation of a complex psychosocial intervention (e.g., Cognitive behavioral therapy [42, 56];, multisystemic therapy [25, 26, 58]).
Table 5

Descriptive summary

StudySettingSampleIntervention/InnovationComplex psychosocial interventionDesign
Qualitative
Bardosh et al. 2017 [16]Health care facilities, multiple countriesKey informants (researchers, Mhealth staff, clinic staff, government officials; n = 32)Mobile health applicationNQualitative, cross sectional, comparative case study, non-randomized
Brewster et al. 2015 [17]HospitalsHospitals (k = 10); hospital employees (hospital staff, n = 82; state hospital representatives n = 8)Initiative to reduce rehospitalization ratesNQualitative, descriptive, cross sectional, non-randomized
Carrera and Lambooij 2015 [18]Primary carePatients (n = 12); health care providers (n = 4)Blood pressure monitoring guidelinesNQualitative descriptive, cross sectional, non-randomized
Frykman et al. 2014 [19]Emergency departmentsDepartments (k = 2), health care providers (n = 11)Multi-professional teamwork guidelineNqualitative, longitudinal (2 assessment points, 21 months), comparative case study, non-randomized
Wiener-Ogilvie et al. 2008 [20]Primary careHealth care providers (n = 9)Asthma management guidelineNqualitative, cross sectional, comparative case study, non-randomized
Quantitative randomized
Atkins et al. 2008 [21]SchoolsTeachers (n = 127); mental health providers (n = 21)Attention Deficit Hyperactivity Disorder guidelinesYquantitative, longitudinal (5 assessment points, 2 years), randomized
Baer et al. 2009 [Substance abuse treatment facilitiesSubstance abuse treatment facilities (k = 6); Mental health providers (n = 118)Motivational InterviewingYquantitative, longitudinal (3 assessment points, 6 months), randomized
Bonetti et al. 2005 [23]Primary careHealth care providers (n = 152)Spinal X-ray referral guidelinesNquantitative, longitudinal (2 assessment points, 2 months), randomized control trial
Garner et al. 2011 [24]Substance abuse treatment facilitiesSubstance abuse treatment facilities (k = 29); mental health providers (n = 95)Adolescent Community Reinforcement Approach and Assertive Continuing CareYquantitative, longitudinal (2 assessment points, 3 years), randomized control trial
Glisson et al. 2010 [25]Juvenile courtsCounties (k = 14); patients (n = 615)Multisystemic TherapyYquantitative, longitudinal (weekly, quarterly, 4 years), randomized control trial
Holth et al. 2011 [26]Behavioral health facilitiesMental health providers (n = 21); families (youth and primary caregiver; n = 41)Multisystemic Therapy, Cognitive Behavior TherapyYquantitative, longitudinal (monthly, 17 months), block randomized control trial
Lee et al. 2018 [27]Schools, child care facilitiesOrganizations (n = 121)Nutritional guidelinesNquantitative, longitudinal (two time points; 2 studies at 6 months, 1 study at 12 months), analysis of aggregated datasets from three randomized control trials
Lochman et al. 2009 [28]SchoolsSchools (k = 57); patients (n = 531); mental health providers (n = 49)Coping Power ProgramYquantitative, longitudinal (2 assessment points, 2 years), randomized
Rapkin et al. 2017 [29]Public library systemCommunities (k = 20); community members (n = 9374)Cancer screening and prevention education programsNquantitative, randomized, stepped-wedge, longitudinal
Rohrbach et al. 1993 [30]SchoolsSchools (k = 25); administrators (n = 25); teachers (n = 60); patients (n = 1147)Adolescent Alcohol Prevention TrialYquantitative, longitudinal (3 assessment points, 2 years), randomized control trial
Seys et al. 2018 [31]HospitalsCare teams (k = 19); care team members (n = 284); patients (n = 257)Care pathway for Chronic Obstructive Pulmonary DiseaseNquantitative, longitudinal (two assessment points, 30 days), randomized
Williams et al. 2014 [32]Behavioral health facilitiesBehavioral health facilities (k = 92); administrators (n = 311)Motivational InterviewingYquantitative, longitudinal (3 assessment points, 3 months), randomized control trial
Williams et al. 2017 [33]Behavioral health facilitiesOrganizations (k = 14); clinicians (n = 475)Evidence-based practice (not specified)Evidence-based practice implemented not reportedquantitative, longitudinal, randomized (4 assessment points, 4 years)
Quantitative non-randomized
Aarons et al. 2009 [34]Behavioral health facilitiesMental health care providers (n = 174)31 child or family evidence-based practicesYaquantitative, cross-sectional, survey, non-randomized
Becker et al. 2016 [35]Substance abuse treatment facilitiesClinics (k = 15); treatment providers (n = 60)Contingency management treatmentYquantitative, longitudinal (biweekly, 12 months), non-randomized
Beenstock et al. 2012 [36]HospitalsHospitals (k = 8); health care providers (n = 364)Smoking cessation guidelineNquantitative, cross sectional, survey, non-randomized
Beets et al. 2008 [37]SchoolsTeachers (n time 1 = 171, n time 2 = 191)Positive Action ProgramYquantitative, cross sectional at two time points, survey, non-randomized
Bonetti et al. 2009 [38]Dentist officesHealth care providers (n = 133)Fissure sealant evidence-based practiceNquantitative, longitudinal, predictive cohort study (3 assessment points, 28 months), non-randomized
Chou et al. 2011 [39]Veterans AffairsHospitals (k = 132), health care providers (n = 2,438)Major depressive disorder screening guidelineNquantitative, cross sectional, survey, randomized
Cummings et al. 2017 [40]Nursing homesNursing homes (k = 7); nursing home staff (n = 333)Coaching for Impressive CareNquantitative, , non-randomized two-group crossover
David and Schiff 2017 [41]Health care system, multiple sitesHealth care providers (n = 77)Child-Parent PsychotherapyYquantitative, cross sectional, survey, non-randomized
Edmunds et al. 2014 [42]Behavioral health facilitiesMental health providers (n = 50)Cognitive Behavioral TherapyYquantitative, longitudinal, non-randomized
Gnich et al. 2018 [43]Dentist officesHealth care providers (n = 709)Fluoride varnish applicationNquantitative, longitudinal (2 assessment points, 18 months), non-randomized
Guerrero et al. 2018 [44]Behavioral health facilitiesBehavioral health facilities (k = 112), mental heal providers (n = 427)Contingency management treatment and medicationassisted treatmentYquantitative, longitudinal (2 assessment points), survey, non-randomized
Huis et al. 2013 [45]HospitalsHospitals (k = 3); departments (k = 67): health care providers (k = 2733)Hand hygiene guidelinesNquantitative, longitudinal, process evaluation of a cluster randomized controlled trial
Little et al. 2015 [46]SchoolsSchool districts (k = 183); departments (k = 22)Tobacco Use Prevention EducationNquantitative, cross sectional, survey, non-randomized
Llasus et al. 2014 [47]University nursing programsNursing students (n = 174)Evidence-based practices (not specified)Nquantitative, descriptive, correlational, cross sectional, survey, non-randomized
Nelson and Steele 2007 [48]Health care system, multiple sitesMental health providers (n = 214)Evidence-based practices (not specified)Nquantitative, cross sectional, survey, non-randomized
Potthoff et al. 2017 [49]Primary careOrganizations (k = 99); health care providers (n = 489)Type 2 diabetes management guidelineNquantitative, longitudinal (2 assessment points, 1 year), correlational, survey, non-randomized
Presseau et al. 2016 [50]Primary careFamily physicians (time 1 n = 632; time 2 n = 426)Prescription of hypertension medicationNquantitative, longitudinal (2 assessment points, approximately 8 months), 2X3 factorial
Simmonds et al. 2012 [51]Health care system, multiple sitesHealth care providers (n = 108)Lower back pain management guidelinesNquantitative, cross sectional, survey, non-randomized
Stockdale et al. 2018 [52]Veterans AffairsHealth care providers (n = 149), patients (n = 3329)Patient Centered Medical HomeNquantitative, cross sectional, survey, non-randomized
Wanless et al. 2015 [53]SchoolsSchools (k = 13); teachers (n = 1114)Responsive ClassroomYquantitative, longitudinal, non-randomized (focuses on one condition in an RCT)
Yamada et al. 2017 [62]HospitalsCare units (k = 32); nurses (n = 779); patients (n = 1,604)Instrumental and conceptual research use, evidence-based pain assessmentNquantitative, cross sectional, non-randomized
Mixed Methods
Armson et al. 2018 [54]Health care system, multiple sitesHealth care providers (n = 70)Breast cancer screening guidelineNmixed method, longitudinal, observational/ naturalist field study, non-randomized
Birken et al. 2015 [55]Health care system, multiple sitesOrganizations (k = 149); administrators (n = 223)Quality improvement initiative based on Chronic Care ModelNmixed method sequential, cross sectional, non-randomized
Kauth et al. 2010 [56]Veterans AffairsClinics (k = 21); mental health providers (n = 23)Cognitive Behavioral TherapyYmixed method, quasi-experimental, longitudinal (2 assessment points, 6 months), randomized
Lukas et al. 2009 [57]Veterans AffairsOrganizations (k = 78); health care providers, non-clinical staff (n = 3870)Advance Clinic AccessNmixed method, cross sectional, observational, non-randomized
Panzano et al. 2012 [58]Behavioral health facilitiesConsultants (n = 34); mental health providers (n = 70)Multisystemic Therapy, Dual Disorder Treatment, Ohio medication algorithms, Cluster-based Outcomes ManagementYmixed method, longitudinal, observational/ naturalist field study, non-randomized
Rangachari et al. 2015 [59]HospitalsDepartments (k = 2); health care providers (n = 101); administrators (n = 6)Central line bundle catheter insertion evidence-based practiceNprospective, longitudinal, exploratory field study, mixed-method analysis
Shrubsole et al. 2018 [60]HospitalsHospitals (k = 4); health care providers (n = 37); patients (n = 107)Aphasia management practicesNmixed method, longitudinal, cross-over, cluster randomized control trial

aMultiple EBPs, some of which were complex psychosocial interventions

Descriptive summary aMultiple EBPs, some of which were complex psychosocial interventions

Study design

Our review included six qualitative (10.9%), seven mixed methods (15.2%), and 34 quantitative studies (73.9%). The most common study design was quantitative non-randomized/observational (21 studies; 45.7%), of which 11 were cross-sectional. There were 13 (28.3%) randomized studies included in this review. Twenty-nine studies (63.0%) were longitudinal (i.e., included more than one data collection time point for the sample).

Study quality

Table 4 shows the results of the MMAT quality assessment. Scores for the included studies ranged from 25 to 100%. Six studies (13.0%) received a 25% rating based on the MMAT criteria [15], 17 studies (40.0%) received 50%, 21 studies (45.7%) received 75%, and only three studies (6.5%) scored 100%. The most frequent weaknesses were the lack of discussion on researcher influence in qualitative and mixed methods studies, lack of clear description of randomization approach utilized in the randomized quantitative studies, and subthreshold rates for acceptable response or follow-up in non-randomized quantitative studies.

Study design and evaluation of mechanisms theories, models, and frameworks

Twenty-seven (58.7%) of the studies articulated their plan to evaluate mechanisms, mediators, or moderators in their research aims or hypotheses; the remaining studies included this as a secondary analysis. Thirty-five studies (76.1%) cited a theory, framework, or model as the basis or rationale for their evaluation. The diffusion of innovations theory [63, 64] was most frequently cited, appearing in nine studies (19.6%), followed by the theory of planned behavior [65], appearing in seven studies (15.2%). The most commonly cited frameworks were the theoretical domains framework (five studies; 10.9%) [66] and Promoting Action on Research in Health Services (PARiHS) [67] (three studies; 6.5%).

Ecological levels

Four studies (8.7%) incorporated theories or frameworks that focused exclusively on a single ecological level; two focusing on leadership, one at the organizational level, and one at the systems level. There was some discordance between the theories that purportedly informed studies and the potential mechanisms of interest, as 67.4% of candidate mechanisms or mediators were at the intrapersonal level, while 30.4% were at the interpersonal level, and 21.7% at the organizational level. There were no proposed mechanisms at the systems or policy level. Although 12 studies (26.1%) examined mechanisms or mediators across multiple ecological levels, few explicitly examined multilevel relationships (e.g., multiple single-level mediation models were tested in one study).

Measurement and analysis

The vast majority of studies (38, 82.6%) utilized self-report measures as the primary means of assessing the mechanism, and 13 of these studies (28.3%) utilized focus groups and/or interviews as a primary measure, often in combination with other self-report measures such as surveys. Multiple regression constituted the most common analytic approach for assessing mediators or moderators, utilized by 25 studies (54.3%), albeit this was applied in a variety of ways. Twelve studies (26.1%) utilized hierarchical linear modeling (HLM) and six studies (13.0%) utilized structural equation modeling (SEM); see Table 6 for a complete breakdown. Studies that explicitly tested mediators employed diverse approaches including Baron and Kenny’s (N = 8, 17.4 causal steps approach [78], Preacher and Hayes’ (N = 3, 6.5%) approach to conducting bias-corrected bootstrapping to estimate the significance of a mediated effect (i.e., computing significance for the product of coefficients) [95, 126], and Sobel’s (N = 4, 8.9%) approach to estimating standard error for the product of coefficients often using structural equation modeling [79]. Only one study tested a potential moderator, citing Raudenbush’s [80, 82]. Two other studies included a potential moderator in their conceptual frameworks, but did not explicitly test moderation.
Table 6

Mechanism analysis

StudyAimsTheory, framework, modelMechanism measurementMediation testing citation
Qualitative
Bardosh et al. 2017 [16]NConsolidated framework for implementation research [68]InterviewsNone
Brewster et al. 2015 [17]YImplementation innovation framework [69]InterviewsNone
Carrera and Lambooij 2015 [18]NTechnology acceptance model [70]; Theory of planned behavior [65]; Model of personal computing utilization [71]Focus groupsNone
Frykman et al. 2014 [19]NDirection, competence, opportunity and motivation (DCOM) [72, 73]Interviews; observationsNone
Wiener-Ogilvie et al. 2008 [20]NNone reportedInterviews; focus groupsNone
Quantitative- randomized
Atkins et al. 2008 [21]YDiffusion of innovation theory [63]Interviews; self-report[74]
Baer et al. 2009 [22]YNone reportedinterviews; self-report[75]
Bonetti et al. 2005 [23]NTheory of planned behavior [65]; Social cognitive theory [76, 77]Self-report[78, 79]
Garner et al. 2011 [24]NTheory of planned behavior [65]Self-report[80, 81]
Glisson et al. 2010 [25]NNone reportedSelf-report, audiotape coding and interviews[82]
Holth et al. 2011 [26]YNone reportedInterviews; self-report[83]
Lee et al. 2018 [27]YTheoretical domains framework [84]Self-report, secondary analysis[85, 86]
Lochman et al. 2009 [28]NDiffusion of innovation theory [87]Coder ratings[88]
Rapkin et al. 2017 [29]YNone reportedSelf-report[89]
Rohrbach et al. 1993 [30]NDiffusion of innovation theory [64]Interviews; self-report; observationsNone
Seys et al. 2018 [31]YNone reportedChart review; self-report[78]
Williams et al. 2014 [32]YDiffusion of innovation theory [87]Self-report[90, 91]
Williams et al., 2017 (66)YOrganizational culture theory [32] and Theory of planned behavior [65]Self-report[92]
Quantitative- non-randomized
Aarons et al. 2009 [34]YInstitutional theory [93], Theory of planned behavior [65], Theory of perceived organizational support [94]Self-report[78]
Becker et al. 2016 [35]YDiffusion of innovation theory [64]Self-reportNone
Beenstock et al. 2012 [36]NTheoretical domains framework [66]Self-report[95]
Beets et al. 2008 [37]YTheory driven evaluation [96]; Diffusion of innovation theory [63]Self-report[97, 98]
Bonetti et al. 2009 [38]NTheory of planned behavior [65]; Social cognitive theory [99]; Operant learning theory [100]; Action planning [101]; Common sense self-regulation model [102]; Precaution adoption process model [103]; Stage theory [103, 104]Self-report; objective measure[78, 79]
Chou et al. 2011 [39]NGoal setting theory [105]; Goal commitment theory [106]Self-report[80, 107]
Cummings et al. 2017 [40]NPromoting action on research in health services (PARiHS) [67]Self-report[108]
David and Schiff 2017 [41]YDiffusion of innovation theory [87, 109]Self-report[110]
Edmunds et al. 2014 [42]YEPIS framework [111]Self-report[80, 112]
Gnich et al. 2018 [43]YTheoretical domains framework [66]Self-reportNone
Guerrero et al. 2018 [44]YTheory on middle manager s[69]Self-report[113]
Huis et al. 2013 [45]NNone reportedObservations; self-report; website visitor registration; logs; field Notes; effect evaluation; quiznone
Little et al. 2015 [46]NDiffusion of innovation theory [64]Self-report[114116]
Llasus et al. 2014 [47]NKnowledge to action conceptual framework [117]Self-report[78, 79, 95]
Nelson and Steele 2007 [48]NNone reportedSelf-reportNone
Potthoff et al. 2017 [49]YDual process model of behavior [118]Self-report[79]
Presseau et al. 2016 [50]YTheory of planned behavior [65]Self-reportNone
Simmonds et al. 2012 [51]YNone reportedSelf-report[78]
Stockdale et al. 2018 [52]YNone reportedSelf-report[119]
Wanless et al. 2015 [53]YNone reportedSelf-report, observation[110]
Yamada et al. 2017 [62]YPromoting action on research in health services (PARiHS) [120]Self-report, chart reviewNone
Mixed methods
Armson et al. 2018 [54]YTheoretical domains framework [66]Interviews; self-reportNone
Birken et al. 2015 [55]NHierarchical taxonomy of leader behavior [121]Interviews; self-report[95, 122]
Kauth et al. 2010 [56]YFixsen model [123]; Promoting action on research in health services (PARiHS) [120]Self-report; logsNone
Lukas et al. 2009 [57]YDiffusion of Innovations Theory [63, 124]Interviews[78]
Panzano et al. 2012 [58]YNone reportedSelf-report[78]
Rangachari et al. 2015 [59]NComplex systems theory [125]Infection rate; chart review; hospital records; logsNone
Shrubsole et al. 2018 [60]NTheoretical domains framework [66]Chart review; self-reportnone
Mechanism analysis

Emergent mechanism models

There was substantial variation in the models that emerged from the studies included in this review. Table 7 represents variables considered in mediating or moderating models across studies (or identified as candidate mediators, moderators, or mechanisms in the case of qualitative studies). Additional file 3 depicts the unique versions of models tested and their associated studies. We attempted to categorize variables as either (a) an independent variable (X) impacting a dependent variable; (b) a dependent variable (Y), typically the outcome of interest for a study; or (c) an intervening variable (M), a putative mediator in most cases, though three studies tested potential moderators. We further specified variables as representing a strategy, determinant, and outcome; see Table 1 for definitions.3
Table 7

Model tested

StudyIndependent variable (X)Intervening variable (M)Dependent variable (Y)
Qualitative
Bardosh et al. 2017 [16]Mobile and text follow up with patientsService organization at clinic level, clinician norms and practices, availability of local champions staff, adaptability and co-design of strategy, receptivity and capacity of local managementCulture of care
Brewster et al. 2015 [17]Patient education, follow-up phone calls to patients after discharge, discharge planning, collaboration with post-acute providersIntrinsic reward to staff --> shift in norms and attitudesReduced hospital readmissions
Carrera and Lambooij 2015 [18]None reported

Mediators: perceived usefulness, perceived ease of use, self-efficacy, attitudes,social norm

Moderator: enabling conditions

Intervention acceptability (providers and patients)
Frykman et al. 2014 [19]Senior manager and consultant-driven teamwork strategy, senior manager and staff-driven teamwork strategyDirection, communication, opportunity, motivationChange in staff behavior
Wiener-Ogilvie et al. 2008 [20]Guideline implementationPractice organization (delegation of work to nurses)Compliance with guidelines
Quantitative—randomized
Atkins et al. 2008 [21]Training and consultationKey opinion leader instrumental supportmental health professional instrumental supportTeacher self-reported used of ADHD guidelines
Baer et al. 2009 [22]Climate for organizational changePost training agency activities to support use of Motivational InterviewingFidelity to intervention (Motivational Interviewing spirit and response to question ratio)
Bonetti et al. 2005 [23]Audit and feedbackDecision difficulty, behavioral controlSimulated behavior
Garner et al. 2011 [24]Pay for performance

1. Subjective norms

2. Attitudes toward intervention

3. Perceived control

1. Therapists’ intention to achieve monthly competence

2. Therapists’ intention to achieve targeted threshold

Glisson et al. 2010 [25]Availability responsiveness and continuity (ARC) Intervention + Multisystemic Therapy quality assurance, pay for performanceFidelity to multisystemic therapyRate of change in child behavior out of home placements
Holth et al. 2011 [26]Workshop + manual, intensive quality assurance + workshop + manualAdherence to contingency management and cognitive behavioral therapy techniquesYouth cannabis use
Lee et al. 2018 [27]Implementation strategy bundles (varied across studies)Knowledge, skills, social/professional role and identity, environmental resourcesNutrition guideline implementation
Lochman et al. 2009 [28]Intensive training + feedback, basic training# of sessions attended, # of objectives completed, # of contacts with trainers, counselor engagement w/clientsClient externalizing behaviors, client social skills, client study skills, client expectancies re: aggression, consistent parenting, client assaultive acts
Rapkin et al. 2017 [29]Indicators of program activities: cumulative local programs, attendance at local programs, time since most recent local program, personal awareness of programs, cumulative outside programsMediators: awareness of free/low cost cancer screening, cancer knowledge, cancer information seeking, having health insurance, annual physical moderator: frequency of library useCancer screening attempts to quit smokingtobacco cessation
Rohrbach et al. 1993 [30]

1. Teacher training

2. Principal support intervention

1a. Teacher self-efficacy, 1b. enthusiasm, 1c. preparedness

2a. Principal encouragement, 2b. Principal beliefs about program

Quantity of program implementation
Seys et al. 2018 [31]Care pathway implementationAdherence to evidence-based recommendations, level of competence, team climate for innovation, burnout, level of organized care30-day hospital readmission
Williams et al. 2014 [32]Information packets and Motivational Interviewing webinarAttitudes towards EBPs, pressure for change, barriers to EBPs, resources, organizational climate, management supportMotivational Interviewing adoption
Williams et al. 2017 [33]Availability,Responsiveness, and Continuity (ARC) intervention implementationProficiency culture --> evidence-based practice intention, barrier reductionEBP adoption, EBP use
Quantitative—non-randomized
Aarons et al. 2009 [34]Agency type

1. organizational support for EBP --> provider attitudes towards EBP

2, 3 organizational support for EBP organizational support for EBP

1,3 provider EBP use2. provider EBP attitudes
Becker et al. 2016 [35]Training as usual, training + ongoing technical assistance, support from in-house champion, specialized training on change process, monthly conference calls and online forum to support changeOrganizational readiness to change (motivation for change, adequacy of resources, staff attributes, organizational climate),perceived intervention characteristics (relative advantage, observability, trialability, compatibility, and complexity)Adoption
Beenstock et al. 2012 [36]Main place of workPropensity to actReferral of women to smoking cessation services
Beets et al. 2008 [37]Perception of school climate

1. Beliefs about responsibility to teach program

2. beliefs about responsibility to teach program --> attitudes towards program --> curriculum delivered

1. Attitudes towards program

2. curriculum delivered to schoolwide material usage

Bonetti et al. 2009 [38]Behavioral intentionAction planningPlacing fissure sealants
Chou et al. 2011 [39]Receipt of individual performance feedback, clinician input into guideline implementation and quality improvement, clinician expectancy, clinician self-efficacyAgreement with guidelines, adherence to guidelines, improved knowledge, practice deliveryFidelity to screening patients for depression
Cummings et al. 2017 [40]Culture, feedback, leadership and resourcesManager support, coaching conversations, job satisfactionConceptual research use, persuasive research use, instrumental research use
David and Schiff 2017 [41]

Child-parent psychotherapy social network

Child-parent psychotherapy supervision

Self-efficacyNumber of child-parent psychotherapy cases, intention to use child-parent psychotherapy
Edmunds et al. 2014 [42]Time following trainingTime spent in consultationKnowledge of cognitive behavioral therapy for anxiety, attitudes towards EBPs
Gnich et al. 2018 [43]Pay-per item financial incentiveKnowledge, skills, social/professional role and identity, beliefs about consequences, motivation and goals (intention), environmental context and resources, social influences (norms), emotion, behavioral regulationFluoride varnish delivery
Guerrero et al. 2018 [44]Top manager transformational leadershipMiddle managers’ implementation leadershipEmployee attitudes towards EBPs, EBP implementation
Huis et al. 2013 [45]individual and organization targeted strategies (education, reminders, feedback), individual and organizational targeted strategies + team and leader strategySocial influence, leadership, performance feedbackHandwashing fidelity
Little et al. 2015 [46]Community priority, organizational support, program championbeliefs about effectiveness of interventions --> funding to adopt programAdoption
Llasus et al. 2014 [47]EBP knowledgeSelf confidence in one's EBP competencies (defined as readiness)EBP implementation behaviors
Nelson and Steele 2007 [48]EBP training, openness of clinical setting to EBPsPositive attitudes towards treatment research, negative attitudes towards treatment researchEBP use
Potthoff et al. 2017 [49]Action planning, coping planningHabitClinical behaviors (prescribing, advising, examining)
Presseau et al. 2016 [50]Printed informational materialsAttitudes toward prescribing, subjective norms, perceived behavioral control, intention to prescribeSelf-reported prescribing behavior
Simmonds et al. 2012 [51]Intolerance of uncertaintyTreatment orientation toward back pain

Recommendations to return to work

2. recommendations to return to usual activities,estimated risk of back pain disability

Stockdale et al. 2018 [52]Health care team communicationPatient-provider communicationPatient satisfaction with primary care provider
Wanless et al. 2015 [53]Use of responsive classroom practices, global emotional support, self-efficacy, collective responsibilityTeacher training engagementFidelity to intervention
Yamada et al. 2017 [62]Instrumental research use, conceptual research use

Organizational context:

leadership, culture, evaluation, social capital, informal interactions, formal interactions, resources, slack space, slack staff, slack time

Pain assessment, evidence-based pain procedure use, pain intensity
Mixed methods
Armson et al. 2018 [54]Implementation tools (printed education materials, informational video, decision aid)Evidence-based information in guideline, evidence-based information in screening module, discussions with peers, application of implementation tools, discussions with patients, lack of evidence about benefits, patients' screening expectations, fear of misdiagnosis, problems with having patient materials availableUse of breast cancer screening guidelines
Birken et al. 2015 [55]

1. Top manager support

2. Performance reviews

3. Human resources

Mediators:

1a. Performance reviews

1b. Human resources

1c. Training

1d. Funding

1e. Local social network involvement

Moderator:

2/3. top manager support

1, 2, 3 middle manager commitment to innovation
Kauth et al. 2010 [56]Facilitation + workshop, workshop

Job-related barriers, # of contacts with facilitator,

time spent in facilitation

% time conducting Cognitive Behavioral Therapy
Lukas et al. 2009 [57]Higher management support, group culture, hierarchical cultureTeam effectivenessExtent of implementation
Panzano et al. 2012 [58]

1. Strategic fit of intervention

2. Climate for innovation

1. Climate for innovation2. Fidelity to intervention

1. Fidelity to intervention

2. Assimilation

Rangachari et al. 2015 [59]Emails containing intervention information and unit level adherence feedback + brief weekly trainingProactive communication between nurses and physicians emergence of championsNumber of catheter days
Shrubsole et al. 2018 [60]Tailored training intervention targeting information provision

Mechanisms of Intervention 1 targeting information provision implementation):

knowledge, beliefs about consequences, social influence, beliefs about capabilities, environmental context and resources

Mechanisms of Intervention 2 targeting implementation of goal setting): beliefs about consequences, social influences, beliefs about capabilities, environmental context and resources

Information provisiongoal setting

Numbering is used to denote match variables across models; not all models tested the same sets of variables

Model tested Mediators: perceived usefulness, perceived ease of use, self-efficacy, attitudes,social norm Moderator: enabling conditions 1. Subjective norms 2. Attitudes toward intervention 3. Perceived control 1. Therapists’ intention to achieve monthly competence 2. Therapists’ intention to achieve targeted threshold 1. Teacher training 2. Principal support intervention 1a. Teacher self-efficacy, 1b. enthusiasm, 1c. preparedness 2a. Principal encouragement, 2b. Principal beliefs about program 1. organizational support for EBP --> provider attitudes towards EBP 2, 3 organizational support for EBP organizational support for EBP 1. Beliefs about responsibility to teach program 2. beliefs about responsibility to teach program --> attitudes towards program --> curriculum delivered 1. Attitudes towards program 2. curriculum delivered to schoolwide material usage Child-parent psychotherapy social network Child-parent psychotherapy supervision Recommendations to return to work 2. recommendations to return to usual activities,estimated risk of back pain disability Organizational context: leadership, culture, evaluation, social capital, informal interactions, formal interactions, resources, slack space, slack staff, slack time 1. Top manager support 2. Performance reviews 3. Human resources Mediators: 1a. Performance reviews 1b. Human resources 1c. Training 1d. Funding 1e. Local social network involvement Moderator: 2/3. top manager support Job-related barriers, # of contacts with facilitator, time spent in facilitation 1. Strategic fit of intervention 2. Climate for innovation 1. Fidelity to intervention 2. Assimilation Mechanisms of Intervention 1 targeting information provision implementation): knowledge, beliefs about consequences, social influence, beliefs about capabilities, environmental context and resources Mechanisms of Intervention 2 targeting implementation of goal setting): beliefs about consequences, social influences, beliefs about capabilities, environmental context and resources Numbering is used to denote match variables across models; not all models tested the same sets of variables

Common model types

The most common model type (29; 63.0%) was one in which X was a determinant, M was also a determinant, and Y was an implementation outcome variable (determinant ➔ determinant ➔ implementation outcome). For example, Beenstock et al. [36] tested a model in which propensity to act (determinant) was evaluated as a mediator explaining the relation between main place of work (determinant) and referral to smoking cessation services (outcome). Just less than half the studies (22; 47.8%) included an implementation strategy in their model, of which 16 (34.8%) evaluated a mediation model in which an implementation strategy was X, a determinant was the candidate M, and an implementation outcome was Y (strategy ➔ determinant ➔ implementation outcome); ten of these studies experimentally manipulated the relation between the implementation strategy and determinant. An example of this more traditional mediation model is a study by Atkins and colleagues [21] which evaluated key opinion leader support and mental health practitioner support (determinants) as potential mediators of the relation between training and consultation (strategy) and adoption of the EBP (implementation outcome). Five studies included a mediation model in which X was an implementation strategy, Y was a clinical outcome, and M was an implementation outcome (strategy ➔ implementation outcome ➔ clinical outcome) [25, 26, 28, 29, 31].

Notable exceptions to model types

While the majority of quantitative studies tested a three-variable model, there were some notable exceptions. Several studies tested multiple three variable models that held the independent variable and mediator constant but tested the relation among several dependent variables. Several studies tested multiple three variable models that held the independent variable and dependent variables constant but tested several mediators.

Qualitative studies

Five studies included in this review utilized qualitative methods to explore potential mechanisms or mediators of change, though only one explicitly stated this goal in their aims [17]. Three studies utilized a comparative case study design incorporating a combination of interviews, focus groups, observation, and document review, whereas two studies employed a cross-sectional descriptive design. Although three of the five studies reported their analytic design was informed by a theory or previously established model, only one study included an interview guide in which items were explicitly linked to theory [19]. All qualitative studies explored relations between multiple ecological levels, drawing connections between intra and interpersonal behavioral constructs and organization or system level change.

Criteria for establishing mechanisms of change

Finally, with respect to the seven criteria for establishing mechanisms of change, the plausibility/coherence (i.e., a logical explanation of how the mechanism operates that incorporates relevant research findings) was the most frequently fulfilled requirement, met by 42 studies (91.3%). Although 20 studies (43.5%), of which 18 were quantitative, provided statistical evidence of a strong association between the dependent and independent variables, only 13 (28.2%) studies experimentally manipulated an implementation strategy or the proposed mediator or mechanism. Further, there was only one study that attempted to demonstrate a dose-response relation between mediators and outcomes. Most included studies (39; 84.8%) fulfilled three or fewer criteria, and only one study fulfilled six of the seven requirements for demonstrating a mechanism of change; see Table 8.
Table 8

Kazdin criteria

AssociationSpecificityConsistencyManipulationTimelineGradientPlausibilityTotal
Qualitative
Bardosh et al. 2017 [16]00100012
Brewster et al. 2015 [17]00100012
Carrera and Lambooij 2015 [18]00000011
Frykman et al. 2014 [19]00001012
Wiener-Ogilvie et al. 2008 [20]00100012
Quantitative—randomized
Atkins et al. 2008 [21]00110013
Baer et al. 2009 [22]00101013
Bonetti et al. 2005 [23]11101015
Garner et al. 2011 [24]01010013
Glisson et al. 2010 [25]00011013
Holth et al. 2011 [26]10111014
Lee et al. 2018 [27]00000011
Lochman et al. 2009 [28]01011003
Rapkin et al. 2017 [29]10001114
Rohrbach et al. 1993 [30]00011013
Seys et al. 2018 [31]11011015
Williams et al. 2014 [32]01011014
Williams et al. 2017 [33]11111016
Quantitative—non-randomized
Aarons et al. 2009 [34]10100014
Becker et al. 2016 [35]00011013
Beenstock et al. 2012 [36]10000001
Beets et al. 2008 [37]10100013
Bonetti et al. 2009 [38]10100013
Chou et al. 2011 [39]10000012
Cummings et al., 2017 [40]00100013
David and Schiff 2017 [41]10100013
Edmunds et al. 2014 [42]00001001
Gnich et al. 2018 [43]00101013
Guerrero et al. 2018 [44]10100012
Huis et al. 2013 [45]00011013
Little et al. 2015 [46]10000012
Llasus et al. 2014 [47]10100013
Nelson and Steele 2007 [48]10000012
Potthoff et al. 2017 [49]10000012
Presseau et al. 2016 [50]00100012
Simmonds et al. 2012 [51]10000012
Stockdale et al. 2018 [52]10000012
Wanless et al. 2015 [53]00001012
Mixed methods
Armson et al. 2018 [54]00100012
Birken et al. 2015 [55]00000011
Kauth et al. 2010 [56]00011013
Lukas et al. 2009 [57]11000013
Panzano et al. 2012 [58]10000012
Rangachari et al. 2015 [59]00001001
Shrubsole et al. 2018 [60]00011013

Studies that only tested mediation relationships are not included in this table

Kazdin criteria Studies that only tested mediation relationships are not included in this table

Discussion

Observations regarding mechanistic research in implementation science

Mechanism-focused implementation research is in an early phase of development, with only 46 studies identified in our systematic review across health disciplines broadly. Consistent with the field of implementation science, no single discipline is driving the conduct of mechanistic research, and a diverse array of methods (quantitative, qualitative, mixed methods) and designs (e.g., cross-sectional survey, longitudinal non-randomized, longitudinal randomized, etc.) have been used to examine mechanisms. Just over one-third of studies (N = 16; 34.8%) evaluated a mediation model with the implementation strategy as the independent variable, determinant as a putative mediator, and implementation outcome as the dependent variable. Although this was the most commonly reported model, we would expect a much higher proportion of studies testing mechanisms of implementation strategies given the ultimate goal of precise selection of strategies targeting key mechanisms of change. Studies sometimes evaluated models in which the determinant was the independent variable, another determinant was the putative mediator, and an implementation outcome was the dependent variable (N = 11; 23.9%). These models suggest an interest in understanding the cascading effect of changes in context on key outcomes, but without manipulating or evaluating an implementation strategy as the driver of observed change. Less common (only 5, 10.9%) were more complex models in which multiple mediators and outcomes and different levels of analyses were tested (e.g., [37, 39]), despite that this level of complexity is likely to characterize the reality of typical implementation contexts. Although there were several quantitative studies that did observe significant relations pointing toward a mediator, none met all criteria for establishing a mechanism. Less than one-third of the studies experimentally manipulated the strategy-mechanism linkage. As the field progresses, we anticipate many more tests of this nature, which will allow us to discern how strategies exert their effect on outcomes of interest. However, implementation science will continue to be challenged by the costly nature of the type of experimental studies that would be needed to establish this type of evidence. Fortunately, methodological innovations that capitalize on recently funded implementation trials to engage in multilevel mediation modeling hold promise for the next iteration of mechanistic implementation research [14, 127] As this work unfolds, a number of scenarios are possible. For example, it is likely the case that multiple strategies can target the same mechanism; that a single strategy can target multiple mechanisms; and that mechanisms across multiple levels of analysis must be engaged for a given strategy to influence an outcome of interest. Accordingly, we expect great variability in model testing will continue and that more narrowly focused efforts will remain important contributions so long as shared conceptualization of mechanisms and related variables is embraced, articulated, and rigorously tested. As with other fields, we observed great variability in the degree to which mechanisms (and related variables of interest) were appropriately specified, operationalized, and measured. This misspecification coupled with the overall lack of high-quality studies (only three met 100% of the quality criteria), and the diversity in study methods, strategies tested, and mediating or moderating variables under consideration, we were unable to synthesize the findings across studies to point toward promising mechanisms.

The need for greater conceptual clarity and methodological advancements

Despite the important advances that the studies included in this review represent, there are clear conceptual and methodological issues that need to be addressed to allow future research to more systematically establish mechanisms. Table 1 offers a list of key terms and definitions for the field to consider. We suggest the term “mechanism” be used to reflect a process or event through which an implementation strategy operates to affect desired implementation outcomes. Consistent with existing criteria [4], mechanisms can only be confidently established via carefully designed (i.e., longitudinal; experimentally manipulated) empirical studies demonstrating a strong association, and ideally a dose-response relation, between an intervening variable and outcome (e.g., via qualitative data or mediation or moderator analyses) that are supported by very specific theoretical propositions observed consistently across multiple studies. We found the term “mediator” to be most frequently used in this systematic review, which can point toward a mechanism, but without consideration of these full criteria, detection of a mediator reflects a missed opportunity to contribute more meaningfully to the mechanisms literature. Interestingly, the nearly half of studies (43.5%) treated a variable that many would conceptualize as a “determinant” as the independent variable in at least one proposed or tested mediation pathway. Presumably, if researchers are exploring the impact of a determinant on another determinant and then on an outcome, there must be a strategy (or action) that caused the change in the initial determinant. Or, it is possible that researchers are simply interested in the natural associations among these determinants to identify promising points of leverage. This is a prime example where the variable or overlapping use of concepts (i.e., calling all factors of interest “determinants”) becomes particularly problematic and undermines the capacity of the field to accumulate knowledge across studies in the service of establishing mechanisms. We contend that it is important to differentiate among concepts to use more meaningful terms like preconditions, putative mechanisms, proximal and distal outcomes, all of which were under-specified in the majority of the included studies. Several authors from our team have articulated an approach to building causal pathway diagrams [128] that clarifies that preconditions are necessary factors for a mechanism to be activated and proximal outcomes are the immediate result of a strategy that is realized only because the specific mechanism was activated. We conceptualize distal outcomes as the eight implementation outcomes articulated by Proctor and colleagues [129]. Disentangling these concepts can help characterize why strategies fail to exert an impact on an outcome of interest. Examples of each follow in the section below.

Conceptual and methodological recommendations for future research

Hypothesis generation

With greater precision among these concepts, the field can also generate and test more specific hypotheses about how and why key variables are related. This begins with laying out mechanistic research questions (e.g., How does a network intervention, like a learning collaborative, influence provider attitudes?) and generating theory-driven hypotheses. For instance, a testable hypothesis may be that learning collaboratives [strategy] operate through sharing [mechanism] of positive experiences with a new practice to influence provider attitudes [outcome]. As another example, clinical decision support [strategy] may act through helping the provider to remember [mechanism] to administer a screener [proximal outcome] and flagging this practice before an encounter may not allow the mechanism to be activated [precondition]. Finally, organizational strategy development [strategy] may have an effect because it means prioritizing competing demands [mechanism] to generate a positive implementation climate [proximal outcome]. Research questions that allow for specific mechanism-focused hypotheses have the potential to expedite the rate at which effective implementation strategies are identified.

Implementation theory

Ultimately, theory is necessary to drive hypotheses, explain implementation processes, and effectively inform implementation practice by providing guidance about when and in what contexts specific implementation strategies should or should not be used. Implementation theories can offer mechanisms that extend across levels of analysis (e.g., intrapersonal, interpersonal, organizational, community, macro policy [130]). However, there is a preponderance of frameworks and process models, with few theories in existence. Given that implementation is a process of behavior change at its core, in lieu of implementation-specific theories, many researchers draw upon classic theories from psychology, decision science, and organizational literatures, for instance. Because of this, the majority of the identified studies explored intrapersonal-level mechanisms, driven by their testing of social psychological theories such as the theory of planned behavior [65] and social cognitive theory [76, 77, 99]. Nine studies cited the diffusion of innovations [63, 64] as a theory guiding their mechanism investigation, which does extend beyond intrapersonal to emphasize interpersonal, and to some degree community level mechanisms, although we did not see this materialize in the included study analyses [63–65, 76, 77]. Moving forward, developing and testing theory is critical for advancing the study of implementation mechanisms because theories (implicitly or explicitly) tend to identify putative mechanisms instead of immutable determinants.

Measurement

Inadequate measurement has the potential to undermine our ability to advance this area of research. Our coding indicated that mechanisms were assessed almost exclusively via self-report (questionnaire, interview, focus group) suggesting that researchers conceptualize the diverse array of mechanisms to be latent constructs and not directly observable. This may indeed be appropriate, given that mechanisms are typically processes like learning and reflecting that occur within an individual and it is their proximal outcomes that are directly observable (e.g., knowledge acquisition, confidence, perceived control). However, conceptual, theoretical, and empirical work is needed to (a) articulate the theorized mechanisms for the 70+ strategies and proximal outcomes [128], (b) identify measures of implementation mechanisms and evaluate their psychometric evidence base [131] and pragmatic qualities [132], and (c) attempt to identify and rate or develop objective measures of proximal outcomes for use in real-time experimental manipulations of mechanism-outcome pairings.

Quantitative analytic approaches

The multilevel interrelations of factors implicated in an implementation process also call for sophisticated quantitative and qualitative methods to uncover mechanisms. With respect to quantitative methods, it was surprising that the Baron and Kenny [78] approach to mediation testing remains most prevalent despite that most studies are statistically underpowered to use this approach, and the other most common approach (i.e., the Sobel test [79]) relies on an assumption that the sampling distribution of the mediation effect is normal [14, 133], neither of which were reported on in any of the 12 included studies that used these methods. Williams [14] suggests the product of coefficients approach [134, 135] is more appropriate for mediation analysis because it is a highly general approach to both single and multi-level mediation models that minimizes type I error rates, maximizes statistical power, and enhances accuracy of confidence intervals [14]. The application of moderated mediation models and mediated moderator models will allow for a nuanced understanding of the complex interrelations among factors implicated in an implementation process.

Qualitative analytic approaches

Because this was the first review of implementation mechanisms across health disciplines, we believed it was important to be inclusive with respect to methods employed. Qualitative studies are important to advancing research on implementation mechanisms in part because they offer a data collection method in lieu of having an established measure to assess mechanisms quantitatively. Qualitative research is important for informing measure development work, but also for theory development given the richness of the data that can be gleaned. Qualitative inquiry can be more directive by developing hypotheses and generating interview guides to directly test mechanisms. Diagramming and tracing causal linkages can be informed by qualitative inquiry in a structured way that is explicit with regard to how the data informs our understanding of mechanisms. This kind of directed qualitative research is called for in the United Kingdom’s MRC Guidance for Process Evaluation [136]. We encourage researchers internationally to adopt this approach as it would importantly advance us beyond the descriptive studies that currently dominate the field.

Limitations

There are several limitations to this study. First, we took an efficient approach to coding for study quality when applying the MMAT. Although it was a strength that we evaluated study quality, the majority of studies were assessed only by one research specialist. Second, we may have overlooked relevant process evaluations conducted in the UK where MRC Guidance stipulates inclusion of mechanisms that may have been described using terms not included in our search string. Third, although we identified several realist reviews, we did not include them in our systematic review because they conceptualize mechanisms differently than how they are treated in this review [137]. That is, realist synthesis posits that interventions are theories and that they imply specific mechanisms of action instead of separating mechanisms from the implementation strategies/interventions themselves [138]. Thus, including the realist operationalization would have further confused an already disharmonized literature with respect to mechanisms terminology but ultimately synthesizing findings from realist reviews with standard implementation mechanism evaluations will be important. Fourth, our characterization of the models tested in the identified studies may not reflect those intended by researchers given our attempt to offer conceptual consistency across studies, although we did reach out to corresponding authors for whom we wished to seek clarification on their study. Finally, because of the diversity of study designs and methods, and the inconsistent use of relevant terms, we are unable to synthesize across the studies and report on any robustly established mechanisms.

Conclusion

This study represents the first systematic review of implementation mechanisms in health. Our inclusive approach yielded 46 qualitative, quantitative, and mixed methods studies, none of which met all seven criteria (i.e., strong association, specificity, consistency, experimental manipulation, timeline, gradient, plausibility or coherence) that are deemed critical for empirically establishing mechanisms. We found nine unique versions of models that attempted to uncover mechanisms, with only six exploring mediators of implementation strategies. The results of this review indicated inconsistent use of relevant terms (e.g., mechanisms, determinants) for which we offer guidance to achieve precision and encourage greater specificity in articulating research questions and hypotheses that allow for careful testing of causal relations among variables of interest. Implementation science will benefit from both quantitative and qualitative research that is more explicit in their attempt to uncover mechanisms. In doing so, our research will allow us to test the idea that more is better and move toward parsimony both for standardized and tailored approaches to implementation. Additional file 1: Figure S1. Inclusion and Exclusion Criteria and Definitions. Additional file 2. PRISMA 2009 Checklist. Additional file 3. Emergent Mechanism Models.
  82 in total

1.  A general approach to causal mediation analysis.

Authors:  Kosuke Imai; Luke Keele; Dustin Tingley
Journal:  Psychol Methods       Date:  2010-12

2.  Multilevel Modeling of Individual and Group Level Mediated Effects.

Authors:  J L Krull; D P MacKinnon
Journal:  Multivariate Behav Res       Date:  2001-04-01       Impact factor: 5.923

Review 3.  Mediators and mechanisms of change in psychotherapy research.

Authors:  Alan E Kazdin
Journal:  Annu Rev Clin Psychol       Date:  2007       Impact factor: 18.561

4.  Team effectiveness and organizational context in the implementation of a clinical innovation.

Authors:  Carol Vandeusen Lukas; David C Mohr; Mark Meterko
Journal:  Qual Manag Health Care       Date:  2009 Jan-Mar       Impact factor: 0.926

5.  Asymptotic and resampling strategies for assessing and comparing indirect effects in multiple mediator models.

Authors:  Kristopher J Preacher; Andrew F Hayes
Journal:  Behav Res Methods       Date:  2008-08

6.  Engagement in Training as a Mechanism to Understanding Fidelity of Implementation of the Responsive Classroom Approach.

Authors:  Shannon B Wanless; Sara E Rimm-Kaufman; Tashia Abry; Ross A Larsen; Christine L Patton
Journal:  Prev Sci       Date:  2015-11

Review 7.  The precaution adoption process.

Authors:  N D Weinstein
Journal:  Health Psychol       Date:  1988       Impact factor: 4.267

8.  Teamwork and Adherence to Recommendations Explain the Effect of a Care Pathway on Reduced 30-day Readmission for Patients with a COPD Exacerbation.

Authors:  Deborah Seys; Luk Bruyneel; Walter Sermeus; Cathy Lodewijckx; Marc Decramer; Svin Deneckere; Massimiliano Panella; Kris Vanhaecht
Journal:  COPD       Date:  2018-02-20       Impact factor: 2.409

9.  Self-Efficacy as a Mediator in Bottom-Up Dissemination of a Research-Supported Intervention for Young, Traumatized Children and Their Families.

Authors:  Paula David; Miriam Schiff
Journal:  J Evid Inf Soc Work       Date:  2017-03-23

10.  Explaining the effects of two different strategies for promoting hand hygiene in hospital nurses: a process evaluation alongside a cluster randomised controlled trial.

Authors:  Anita Huis; Gerda Holleman; Theo van Achterberg; Richard Grol; Lisette Schoonhoven; Marlies Hulscher
Journal:  Implement Sci       Date:  2013-04-08       Impact factor: 7.327

View more
  55 in total

1.  Achieving oncology mental health providers' usage of an empirically supported treatment: Lessons learned.

Authors:  Barbara L Andersen; Caroline S Dorfman; Claire C Conley
Journal:  Psychooncology       Date:  2021-05       Impact factor: 3.894

2.  Understanding key implementation determinants for a school-based universal prevention intervention: a qualitative study.

Authors:  Andria B Eisman; Sarah Kiperman; Laney A Rupp; Amy M Kilbourne; Lawrence A Palinkas
Journal:  Transl Behav Med       Date:  2022-03-17       Impact factor: 3.046

3.  Replicating dissemination and identifying mechanisms of implementation of an empirically supported treatment.

Authors:  Stephen B Lo; Claire C Conley; Brittany M Brothers; Marlena M Ryba; Georita F Frierson; Rebecca A Shelby; Lisa M Thornton; Kristen M Carpenter; Barbara L Andersen
Journal:  Health Psychol       Date:  2021-07       Impact factor: 5.556

4.  Sustainment of Integrated Care in Addiction Treatment Settings: Primary Outcomes From a Cluster-Randomized Controlled Trial.

Authors:  Helene Chokron Garneau; Mehret T Assefa; Booil Jo; James H Ford; Lisa Saldana; Mark P McGovern
Journal:  Psychiatr Serv       Date:  2021-08-04       Impact factor: 3.084

Review 5.  A mapping review of NIDA-funded implementation research studies on treatments for opioid and/or stimulant use disorders.

Authors:  Hannah Cheng; Hélène Chokron Garneau; Mina Yuan; Mark P McGovern
Journal:  Drug Alcohol Depend       Date:  2021-05-21       Impact factor: 4.852

6.  Integration of Improvement and Implementation Science in Practice-Based Research Networks: a Longitudinal, Comparative Case Study.

Authors:  Melinda M Davis; Rose Gunn; Erin Kenzie; Caitlin Dickinson; Cullen Conway; Alex Chau; LeAnn Michaels; Steven Brantley; Devon K Check; Nancy Elder
Journal:  J Gen Intern Med       Date:  2021-04-14       Impact factor: 6.473

7.  Cross-cultural adaption and psychometric investigation of the German version of the Evidence Based Practice Attitude Scale (EBPAS-36D).

Authors:  Katharina Szota; Jonathan F B Thielemann; Hanna Christiansen; Marte Rye; Gregory A Aarons; Antonia Barke
Journal:  Health Res Policy Syst       Date:  2021-06-02

8.  Context and mechanisms that enable implementation of specialist palliative care Needs Rounds in care homes: results from a qualitative interview study.

Authors:  Jane Koerner; Nikki Johnston; Juliane Samara; Wai-Man Liu; Michael Chapman; Liz Forbat
Journal:  BMC Palliat Care       Date:  2021-07-22       Impact factor: 3.234

9.  Strengthening methods for tracking adaptations and modifications to implementation strategies.

Authors:  Amber D Haley; Byron J Powell; Callie Walsh-Bailey; Molly Krancari; Inga Gruß; Christopher M Shea; Arwen Bunce; Miguel Marino; Leah Frerichs; Kristen Hassmiller Lich; Rachel Gold
Journal:  BMC Med Res Methodol       Date:  2021-06-26       Impact factor: 4.615

Review 10.  Ten years of implementation outcome research: a scoping review protocol.

Authors:  Rebecca Lengnick-Hall; Enola K Proctor; Alicia C Bunger; Donald R Gerke
Journal:  BMJ Open       Date:  2021-06-18       Impact factor: 2.692

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.