Literature DB >> 33990462

Conceptualising interventions to enhance spread in complex systems: a multisite comprehensive medication review case study.

Laura Lennox1,2, Julie E Reed3,4, Susan Barber1,2, Neil Stillman1, Sophie Spitters1, Emily Ward5, Vanessa Marvin5.   

Abstract

BACKGROUND: Advancing the description and conceptualisation of interventions in complex systems is necessary to support spread, evaluation, attribution and reproducibility. Improvement teams can provide unique insight into how interventions are operationalised in practice. Capturing this 'insider knowledge' has the potential to enhance intervention descriptions.
OBJECTIVES: This exploratory study investigated the spread of a comprehensive medication review (CMR) intervention to (1) describe the work required from the improvement team perspective, (2) identify what stays the same and what changes between the different sites and why, and (3) critically appraise the 'hard core' and 'soft periphery' (HC/SP) construct as a way of conceptualising interventions.
DESIGN: A prospective case study of a CMR initiative across five sites. Data collection included: observations, document analysis and semistructured interviews. A facilitated workshop triangulated findings and measured perceived effort invested in activities. A qualitative database was developed to conduct thematic analysis.
RESULTS: Sites identified 16 intervention components. All were considered essential due to their interdependency. The function of components remained the same, but adaptations were made between and within sites. Components were categorised under four 'spheres of operation': Accessibility of evidence base; Process of enactment; Dependent processes and Dependent sociocultural issues. Participants reported most effort was invested on 'dependent sociocultural issues'. None of the existing HC/SP definitions fit well with the empirical data, with inconsistent classifications of components as HC or SP.
CONCLUSIONS: This study advances the conceptualisation of interventions by explicitly considering how evidence-based practices are operationalised in complex systems. We propose a new conceptualisation of 'interventions-in-systems' which describes intervention components in relation to their: proximity to the evidence base; component interdependence; component function; component adaptation and effort. © Author(s) (or their employer(s)) 2022. Re-use permitted under CC BY-NC. No commercial re-use. See rights and permissions. Published by BMJ.

Entities:  

Keywords:  breakthrough groups; clinical practice guidelines; collaborative; complexity; healthcare quality improvement; implementation science

Mesh:

Year:  2021        PMID: 33990462      PMCID: PMC8685660          DOI: 10.1136/bmjqs-2020-012367

Source DB:  PubMed          Journal:  BMJ Qual Saf        ISSN: 2044-5415            Impact factor:   7.035


Introduction

The reproduction of successful interventions in new settings is essential for maximising the benefits of treatments and innovations for patients and healthcare systems.1 2 However, it is recognised that spreading interventions is a significant challenge for healthcare teams and managers.2–6 Many promising initiatives fail to spread, remaining isolated within the local context in which they were created.1 7 This results in missed opportunities to improve patient care at scale and can lead to significant variation in care processes and outcomes across healthcare organisations.8 9 The difficulty in spreading interventions is often attributed to a lack of consideration for the complexity and variation among different healthcare settings.5 10–12 Healthcare organisations are increasingly recognised as complex systems (ie, systems that dynamically change overtime, where activities are interdependent and inter-related, and where system actors have agency and are self-organising).1 13–15 Therefore, the role of local variation and context must be understood in order to unpick the processes that lead to spread and optimise intervention functioning in new settings.16 This includes recognition of diverse work processes, organisational structures, people, and cultures which may support or hinder the intervention’s desired effects.17 Several theories, models and frameworks have been designed to conceptualise the spread of healthcare innovations.4 18–27 There is no consensus on optimal ways to spread an intervention, but the importance of defining an intervention and establishing a description of components of an initiative believed to be essential to its effectiveness is widely accepted as an important step.4 19 22 27 However, defining complex interventions is not straightforward. The limited ability of those outside of the system (eg, academics and policymakers) to fully design interventions in advance is recognised.28 29 People in the system have unique ‘insider knowledge’ about how things really work which can help identify facilitators and overcome barriers to success.30 Such knowledge is often tacitly held, and no one person has a comprehensive view of how a system works, nor how an intervention will operate in practice.22 30 Defining interventions therefore requires a complex process of intervention testing, clarification and improvement over time.22 31 32 Quality improvement (QI) approaches tend to focus on building capacity and competencies for people inside the system to understand and intervene in their local settings to achieve improvement goals.33 34 While this builds capacity for improvement within individual sites, historically there has been less emphasis on describing interventions to share learning between improvement sites,35–37 and as such this ‘insider knowledge’ is rarely made explicit or captured in easily shareable forms.30 38 There are limited examples of methodological approaches to describe and conceptualise intervention components from an insider perspective.28–30 Prospectively studying and working in partnership with improvement initiatives could improve our ability to capture the richness of intervention spread and make explicit the tacit knowledge that emerges during this process.39 This study aimed to co-produce a detailed description of an evidence-based intervention and capture how it was spread. The co-production of knowledge was achieved through a partnership: improvement team members brought expert clinical, practice and QI knowledge needed to understand and intervene in their local system; researchers brought expertise in observing, capturing and making sense of what happens, and finding patterns in the busy world of frontline practice.39 The focus of this research was to develop a more in-depth understanding of the intervention, namely, to describe how the intervention was operationalised, and the work required to do this. In particular, we were interested in understanding the extent to which a common or shared view of this work was experienced between sites, and what differences or adaptations emerged, and why. To conceptualise the intervention, we examined the popular implementation constructs of ‘hard core’ and ‘soft periphery’ (HC/SP). The innovation diffusion model by Denis et al suggests that each intervention contains a theoretical ‘hard core’ that is ‘well defined and fixed’ and a ‘soft periphery’ that is ‘less clear and more flexible to manipulation by the adopting system’.19 40 In practice, this means aiming for fidelity to an intervention’s HC components, while the SP components could be adapted to fit new settings. We felt that the HC/SP concept could provide a useful lens to conceptualise the intervention and address the question of ‘what stayed the same, and what changed?’ between sites. However, an exploration of the literature revealed that while HC/SP is commonly used, (including within many popular theoretical frameworks41–43), there is no consistent definition, with several authors proposing various definitions.4 19 44 For example, in contrast to Denis et al’s definition above, Greenhalgh et al define core components as ‘the irreducible elements of the innovation itself’ and peripheral components as ‘the organisational structures and systems required for the full implementation of the innovation’.4 The Medical Research Council (MRC) uses ‘core components’ interchangeably with ‘active ingredients’, defined as ‘essential to achieving good outcomes for consumers at an implementation site’. The peripheral components can then be defined as those that can be adapted without impacting outcomes.44 A small number of studies have attempted to define the HC/SP of complex interventions.45–49 However, these studies have included only high-level descriptions of the HC/SP made in advance or post-hoc and were not based on structured empirical study. With little work done to assess these theoretical constructs or provide guidance on their practical application, we recognised an opportunity to critically appraise the HC/SP construct within our empirical case study.50–53 Our research therefore had three objectives: To describe the work required to operationalise evidence-based practice from the perspective of improvement team members. To identify what stayed the same and what changed between the different sites and why. To critically appraise the ‘HC’ and ‘SP’ construct as a way of conceptualising interventions in complex systems.

Methods

Study setting

This study took place under the National Institute for Health Research Collaboration for Leadership in Applied Health Research and Care Northwest London. The purpose of the programme was to investigate how to support frontline care teams to translate research evidence into practice (eg, evidence-based medicine and clinical guidelines recommendations).45–47 The programme advocated and supported a specific implementation process using an ‘iterative systematic approach to improvement’ which included (QI) tools and methods (eg, model for improvement, plan–do–study–act (PDSA) cycles, process mapping), along with iterative evaluation, to support the implementation process.54–57 The approach also included the development of Action Effect Diagrams for each initiative to build programme theory for evaluation of impact.58–60

Study design

A longitudinal case study approach was taken to prospectively examine an initiative that aimed to spread a comprehensive medication review (CMR) intervention.61 62 For the purposes of this work, it will be referred to as the ‘CMR initiative’.

The case

This study reports on the spread of the CMR initiative between September 2014 and August 2018. The initiative was developed in response to evidence that with increasing polypharmacy there is a greater risk of medication-related harm and poorer patient outcomes.17 40 63 To optimise medication use and reduce harm, the National Institute for Health and Care Excellence (NICE) guidelines recommend conducting CMR for people with an indication for a review.63 A CMR is a medication review done systematically with adequate information about the patient and with the patient’s involvement, often using structured criteria for detecting potentially inappropriate prescribing. NICE recommends that ‘organisations should determine locally the most appropriate health professional to carry out a structured medication review, based on their knowledge and skills’.63 This often includes the prescribing physician, a pharmacist and/or another health professional who is part of a multidisciplinary team.63–65 The initiative began at site A, an acute hospital, and was led by a team with significant medicines optimisation expertise and experience in QI. To support a structured CMR, the initiative adapted evidence-based criteria from the Screening Tool of Older People’s Prescriptions (STOPP)65 and produced an updated tool ‘Screening Tool for Older People’s Inappropriate Treatments’.66 67 CMRs using the STOPP criteria were implemented for anyone aged 70 years or above, who presented acutely and were taking medicines regularly. Following results demonstrating improved patient safety and care at the initial site, the initiative gained interest from other sites across northwest London.67 68 From March 2015, four additional acute hospital sites (sites B–E) sought to implement this initiative, with varying levels of prior QI experience. All sites reported an improvement in the number of structured medication reviews being conducted (demonstrated by Statistical Process Control rule breaks), and described a culture shift in how clinicians planned and discussed medication reviews as part of their comprehensive assessments.69 70 The evaluation of outcomes across sites will be published separately.

Data collection methods

Data were collected in two phases. Observation: non-participant observation of the teams took place (24 hours) at facilitated workshops (about the intervention and its operationalisation) and routine meetings that discussed plans, progress and emergent learning. Observations were recorded in a field notebook. Documentary analysis: 53 initiative documents were collected and analysed (meeting minutes, presentations, initiative review reports) in addition to QI tool outputs (action effect (programme theory) diagrams, PDSA cycles and a structured sustainability tool) to capture initiative components and activities.57 58 71 Semistructured interviews: key informant sampling was used to recruit interviewees from all sites.72 73 Participants were selected based on their knowledge, expertise and role within the initiative.74 Eight interviews were carried out face-to-face with one interviewer (LL) and elicited information on intervention components, progress and adaptations. An interview guide was used, and all interviews were recorded and professionally transcribed (interviewee details are found in online supplemental file 1). Facilitated workshop: a facilitated workshop (n=10) was conducted to triangulate interpretation of intervention components; assess perceptions of the importance of components and establish the proportion of effort invested in activities. Purposive sampling was used to invite individuals who had played a substantial role within the initiative across the five sites (online supplemental file 1). A discussion was conducted (facilitated by SB) to present the intervention components and remove or adapt any components based on participant views. An activity was then conducted to explore the perceived effort expended across each of the components where each participant was provided with 100 counter chips to distribute between each of the components (represented by printed headings and pots to place the counters). The focus group was audio recorded and professionally transcribed, and field notes were taken.

Data analysis

A qualitative database was developed using NVivo V.10 to conduct thematic analysis.75–77 Interview transcripts, documents and observational fieldnotes were imported into the database. Following familiarisation with each source and modifying (eg, correcting transcription errors) material as necessary, a preliminary coding structure was inductively developed, establishing thematic groupings to understand the spread of the intervention. The coding structure and themes were iteratively developed and refined as further data (ie, the workshop transcript and fieldnotes) were added.78 The data were eventually summarised to highlight overarching ‘spheres of operation’, intervention component functions and adaptations. The ratings from the counter chip ‘effort’ exercise were input into an Excel database and descriptive analysis performed. Findings were then used to examine possible HC/SP configurations of the CMR initiative based on three existing definitions.

Results

Insights from the improvement team members provided a rich picture of how the CMR initiative evolved over time within and across the five sites. Teams addressed many system issues that extended beyond the immediate conduct of the CMR that were required to support the successful operationalisation of the evidence base. Results are presented in five sections: (I) Intervention components; (II) Interdependence of components; (III) Initiative evolution and adaptation; (IV) Effort to operationalise the intervention; (V) Critical appraisal of the ‘HC’ and ‘SP’.

I. Intervention components

In our analysis, the evidence base (NICE recommendation to conduct a CMR with structured criteria for identifying inappropriate medications, here the STOPP criteria49) provided a focal point for understanding the work that was required to operationalise it into practice. Translating this evidence base into practice was the focus of the improvement teams’ work, and this remained consistent across all sites. Analysis revealed a total of 16 components which had essential functions in the operationalisation of the CMR initiative. The components were organised under four ‘spheres of operation’ that represent what needed to be in place in the system (eg, the necessary procedures, processes and behaviours) to support the intervention in practice: Accessibility of evidence base: the need to interpret the evidence base and make the evidence easily available for clinicians to use in daily clinical practice within their specific context. Processes of enactment: the clinical activities involved in enacting the evidence-based intervention in local practice, that is, ensuring the right patients are identified and receive the intervention, and that results from the intervention are acted appropriately on. Dependent processes: the organisational processes and infrastructures that directly or indirectly support the clinical activities of the intervention. Dependent sociocultural issues: the beliefs, common practices and/or behaviour that are required to support effective operationalisation of the intervention. The team was required to take action within all spheres to ensure the successful operationalisation of the intervention. table 1 presents the four ‘spheres of operation’ with their respective intervention components. It highlights each component’s function, adaptations made and effort expended on each.
Table 1

Spheres of operation and intervention components of the CMR initiative

Intervention components(% perceived effort)Component functionsAdaptations
Accessibility of evidence base (16.7)
 Summaries for clinicians(3.7)To provide condensed and easily accessible information from STOPP relevant to local patient populationsWritten summaries of STOPP evidence and potentially inappropriate prescriptions list66 remained consistent across sites however they were adapted over time to reflect updated evidence base92
 Portable aide-memoires(6.1)To remind and prompt clinicians (particularly junior doctors/pharmacists) to look at high-risk medication and symptom presentationsPocket-sized laminated cards were developed and this remained consistent across sites but adapted to reflect updated evidence base.92 Each site devised local processes for obtaining and distributing cards to staff
 Education materials(6.9)To increase clinician (doctor, pharmacist) knowledge, skills and confidence in conducting a comprehensive medication reviewAdapted over time to reflect updated evidence base and initiative learning.64 67 93 94 Adapted between sites to fit with existing materials and courses, teaching style preferences and target audience
Process of enactment (24.6)
 Patient identification(7.4)To identify and target individuals and groups who may be at increased risk, prioritising available resourcesAdapted in response to patient demographics/acuity and local resource availability and constraints. Adapted to existing local processes to flag patients for comprehensive medication review
 Taking individual patient medication history(2.5)To screen patients admitted to hospital, identifying an accurate medication list, and any indications of potential adverse drug reactionsAdaptations between sites in the method for obtaining the medication histories, for example, if clinicians could/could not access electronic primary care records that included regular medications
 Conducting a CMR(9.3)To perform a structured critical examination of all current medication with the aim of reaching an agreement with the patient about optimising treatment and minimising medicines-related problems.95 Ideally a CMR is conducted as part of a multidisciplinary team ward roundAdapted between sites to account for ward round consistency and availability of staff. Adaptation of who would conduct the review, when and how this would take place (eg, whether led by consultant geriatrician, junior doctor, pharmacists or conducted jointly) based on the interests and availability of expertise within each site
 Communicate and act on review(5.4)To act on and inform what, if any, changes are going to be made to the patient’s medicines regime. This requires good communication processes between hospital staff and patient/carer, as well as healthcare professionals from other settingsAll sites recorded the review in the discharge care summaries but adaptations were made on how this was documented (eg, varying electronic systems), who was responsible for documenting, and the processes for internal communication (eg, to other care professionals)
Dependent processes (20.9)
 Staff availability and timing of rotas(5.9)To ensure the availability of appropriately qualified and trained staff to conduct the CMRAdaptations made between sites based on patterns and structures of ward rounds (eg, to account for consultant availability and preferences, and trained pharmacy staff availability). Adaptations also addressed staffing-related issues including turnover, planned and unplanned leave, and weekend staffing
 Medicines reconciliation process(2.2)To provide clarity about medicines taken by the patient on admissionEach site had existing processes for taking patients’ medication histories. Adaptations between sites to increase reliability and availability of information for CMR (eg, definitions were developed for different levels of medicines reconciliation and work conducted to implement and improve care standards)
 Documentation of review decisions(3.7)To have a clear and structured template for documenting CMR which ensures that the correct review process is followed. This ensures clear documentation is available to other care professionals, to inform discharge summaries, and to support data collection and analysisAdapted between sites to reflect existing systems and current structure of records (paper based or electronic) as embedded into standard documentation practices. In sites C–E, electronic patient record systems changed over time requiring further adaptations
 Handover processes(1.2)To establish clear processes and practices for handover of patient care from one clinical setting to another to ensure CMR is acted on. This requires good documentation, particularly as patients are discharged from hospital back to the community or other care settingsAdaptations made to respond to established processes and practices at each site. Adaptations used to improve local documentation processes including the accuracy and clarity of the information being shared
 Documentation infrastructure(7.9)To ensure documentation infrastructure is set up, and the authorities and permissions required to make changes, and the technological knowledge and skills required to enact them are in placeAdaptations to patient record structure depended on local differences including which individuals/boards were required to approve changes, the perceived priority compared with other organisational changes, the time and effort required to build technical understanding and relationships with those responsible for enacting the changes as well as the extent of use of electronic systems and the functionality of different systems in use
Dependent sociocultural issues (37.8)
 Collaboration and communication between doctors and pharmacists(5.2)To ensure staff engage in a multidisciplinary approach to deliver the intervention effectively. It may be necessary to improve relationships, increasing levels of trust and value of different professional opinionsEach site invested time and energy building relationships between doctors and pharmacists to facilitate collaborative working. This was adapted in each site depending on the historical relationships and practices between doctors and pharmacists, and the way trust and relationships grew and evolved over time
 Staff awareness and competencies(10.1)To increase the awareness and competence of doctors and pharmacists on the evidence base, including the need to conduct CMR, and how to conduct one well, was necessary to ensure enough trained and capable staff were available to conduct reviewsAdaptations between sites made based on scope of the initiative (whole hospital vs specific wards), existing curriculum requirements, training schedules and the need to develop new education material based on recognised needs across sites
 Role expectations(4.7)To set expectations about roles to ensure staff are aware of their role in medicines review and outline the need for skilled staff to perform the reviews, and the time required to do soAdapted to reflect site processes, organisation, and staff availability and training. Adaptations made between sites on who would conduct and communicate the CMR. Role expectations varied across sites with particular groups holding different responsibilities for the CMR (eg, diverse staff combinations of doctors, junior doctors, pharmacists and nurses)
 Review of decision-making processes(4.2)To clarify who should conduct the review, at what point in the patient journey, and how the review would be used to inform future decisions about patient careAdaptations were required for each site depending on which patient group was targeted, the staff involved in leading the review (eg, pharmacist or doctor), whether it occurred as part of multidisciplinary team ward round or not, and other factors, for example, length of stay of the patient in hospital
 Junior doctor and junior pharmacist empowerment(13.8)To support junior doctors and pharmacists in making appropriate decisions regarding potentially inappropriate prescriptions and empowerment to challenge senior decision-makersAdaptations made to training programmes to support junior staff empowerment. Adaptations between sites used to determine when and how medication risks should be communicated depending on their level of severity to help empower junior staff to take actions (eg, one site invested in a local consensus building exercise)

CMR, comprehensive medication review; STOPP, Screening Tool of Older People’s Prescriptions.

Spheres of operation and intervention components of the CMR initiative CMR, comprehensive medication review; STOPP, Screening Tool of Older People’s Prescriptions.

II. Interdependence of components

The cumulative and interdependent roles of the components emerged as critical to successful operationalisation of the intervention. Although the 16 components are represented as distinct activities, the participants described their interactions. For example, portable aide-memoires (Accessibility of the evidence base) were distributed to junior staff to help them identify patients who would benefit from CMR (Process of enactment). Conducting a CMR was dependent on the quality of medical history available for individual patients (Process of enactment). This was dependent on the quality of the processes and practices in place for medicine reconciliation (Dependent processes), which in turn was influenced by the availability of trained staff to perform medicines reconciliation (Dependent processes) and the role expectations of these staff members (Dependent sociocultural issues). Interdependencies such as these provided insight into how wider systems, processes and practices needed to be considered to support effective and routine application of the evidence base in practice. While some of these interdependencies were anticipated by staff, others emerged throughout the initiative (see section III).

III. Intervention evolution and adaptation

Spread sites developed a shared aim to ‘sustainably improve our delivery of consistent, high quality medicines optimisation to achieve better patient experience and outcomes in northwest London’ (Document_AED). However, participants acknowledged the need for adaptations across sites, to ‘Design your intervention so that it fits with your local practices and processes’ (I9_Pharmacist_Site B). While there was commonality in the components deployed across all sites, adaptations were observed including adaptations between sites, within sites and over time in response to local needs, emerging evidence and site learning (table 1, column 3).

Adaptations made between sites in response to local resources, needs and infrastructure

Each site made adaptations to the scope of the initiative. At site A, the initiative was implemented across the whole hospital, but other sites delivered the initiative on acute care of the elderly wards (sites B, C, E), and an acute assessment unit (site D). The need to adapt scope at sites B–E was attributable to resource availability. ‘(Site B) can’t spread further than the geriatric wards…(Site B) can’t spread because there isn’t the resource. So it comes down to the same thing for every project: we need staff and for the staff we need funding’ (I8_Clinical lead_Sites A and B). Sites decided to target the most vulnerable groups to maximise potential impact from the intervention. Decisions to limit scope caused adaptations to be made between sites including how patients were identified for CMR; which staff were involved with conducting CMR in each setting and which staff received training. Site digital infrastructure also played a key role in adaptation required across the sites. For example, an initial CMR template was produced by site A to guide clinicians on how, when and why to carry out a medication review. Each site invested significant time in navigating local infrastructure and information governance processes to introduce the CMR templates in their settings. For sites C–E, this was further complicated by pilots and upgrades to new digital infrastructure.

Adaptations made within sites to respond to local variation

The process of delivering the initiative was dependent on resources and preferences of staff. For example, it was agreed by all sites that, in principle, the CMR should be conducted as part of a multidisciplinary team ward round. ‘We’ve basically decided that the only way you were going to get this done in a meaningful way was to have a multidisciplinary review and have everyone going round at the same time…that really helped to build some relationships that weren’t there before’ (I1_Project manager_Site A). However, adaptations were required on a daily and weekly basis within sites in response to staff resource and preferences. This was influenced by many factors including the availability of qualified pharmacists on individual sites, staff shortages, parental leave and competing care priorities. ‘The project was ongoing so we had huge variability in our pharmacy staff during the project period’ (Facilitated workshop_20170803).

Adaptations over time in response to emerging evidence and site learning

All sites made adaptations in response to changing evidence and emergent learning. A major impact for all sites was the update of the STOPP guidance in 2015. The new guidance reflected an updated systematic review of problematic prescriptions.65 This required an update of initiative materials including prompt cards and educational materials. As well as adapting to new evidence, local learning was a driving force for further adaptation and improvement. For example, the teams had not anticipated the low levels of confidence that junior doctors had in conducting a CMR. They found that although junior doctors received training in how to prescribe medicines, no training was provided on how to deprescribe (stop) medications. This insight led site A to develop a new curriculum to increase competence in this area.65 ‘To improve doctor competency and confidence in deprescribing inappropriate medicines a new module on deprescribing…has been incorporated into junior doctor training’ (Document _Annual Report).

IV. Effort to operationalise the intervention

The effort expended to operationalise the intervention components varied across spheres of operation (figure 1 and table 1 (%)). Participants reported that most effort had been dedicated to components categorised within the sphere of ‘Dependent sociocultural issues’ (37.8%). Within this sphere, the specific components, Junior doctor and pharmacist empowerment (13.8%); Staff awareness and competencies (10.1%) were noted as requiring significant effort to ensure continued delivery of the initiative. Components Conducting a comprehensive medication review (CMR) (9.3%) and Documentation infrastructure (7.9%) were also judged to require a large amount of effort, due to the time required to conduct CMRs routinely in addition to normal practice. Participants reported that the least effort was invested in the sphere of ‘Accessibility of evidence base’ (16.7%). This was viewed as requiring less effort as components involved evidence consolidation tasks (eg, Summaries for clinicians (3.7%), Portable aide-memoires (6.1%)) which were completed at designated times (eg, beginning of the initiative and when the evidence was updated), while the other tasks were seen as ongoing activities.
Figure 1

Perception of effort expended on intervention components of the quality improvement initiative by site participants.

Perception of effort expended on intervention components of the quality improvement initiative by site participants.

V. Critical appraisal of the ‘HC and ‘SP’

To identify and reflect on possible HC/SP configurations for the CMR initiative, three definitions were examined (table 2).4 19 44 In critically assessing the HC/SP construct, we found the different construct definitions challenging to interpret and apply to the CMR case study. Due to the ambiguity and lack of specificity of the given definitions, the intervention components were often unable to exclusively meet the requirements for either HC or SP. Therefore, results show contradictory configurations of the HC/SP for the initiative.
Table 2

Applying hard core/soft periphery (HC/SP) definitions to the CMR case study

PerspectiveApplication to empirical findings
HCSP
Denis et al 19 Components that are well defined and fixed CMR/STOPP evidence base is well defined but all components (including the evidence base) were adapted, for example, nothing was fixed Components that are less clear and more flexible to manipulation by the adopting system All interventions’ components partially met this definition as all components were adapted to some extent
Greenhalgh et al 4 Irreducible elements of the innovation itself All CMR components potentially considered HC, due to participants considering all to be irreducible elements due to their interdependent nature Organisational structures and systems required for the full implementation of the innovation Two spheres of the CMR initiative—‘dependent processes’ and ‘dependent sociocultural issues’—may be considered SP as they represent structures and systems supporting implementation—but also reflected ‘irreducible elements’ of the intervention and therefore could be viewed as both HC and SP
Medical Research Council(Craig et al)44 ‘Active ingredients’ or components essential to achieving good outcomes for consumers at an implementation site All components could be perceived as HC as all were perceived as being required to operationalise the CMR/STOPP evidence base and achieve positive outcomes for patients Components of an intervention which can be adapted without impacting outcomes Inconclusive. All components were adapted in response to local resources/needs/evidence, and yet these adaptations were viewed as essential to successful implementation and achieving outcomes

CMR, comprehensive medication review; STOPP, Screening Tool of Older People’s Prescriptions.

Applying hard core/soft periphery (HC/SP) definitions to the CMR case study CMR, comprehensive medication review; STOPP, Screening Tool of Older People’s Prescriptions. Denis et al’s19 definition of HC components as ‘fixed’ meant that none of the CMR intervention components were considered core as none were completely fixed: even the evidence base informing the intervention was updated during the course of the initiative. Craig et al’s44 definition also implies that the ‘active’ core components are not adapted, and the components that are adapted (periphery) are not important to outcomes. In our case study all components were adapted, and yet all were deemed necessary to achieving the desired outcomes. Conversely, Greenhalgh et al’s4 definition fit more closely with our empirical findings, as it does not associate the HC or SP with adaptation and similarly recognised the importance of supporting structures and process components within intervention description. However, some ambiguity around this definition remains as it draws a distinction between the core ‘irreducible elements of the intervention’ and the peripheral ‘supporting systems and structures’, which makes the implicit assumption that systems and structures are not essential components of an intervention. Our study demonstrated that participants considered all components to be ‘irreducible elements’, including those within the organisational structure and processes, due to their interdependent nature. This made the differentiation between HC and SP ambiguous in this case.

Discussion

The ability to accurately describe interventions and understand how evidence can be translated into routine practice is critical to reproduce improvements in different settings.22 27 30 This study provides rich insight into the work conducted from the perspective of improvement teams to appraise and advance the description and conceptualisation of interventions in complex systems.

An insider’s view: interventions-in-systems

Analysing the improvement initiative from the perspective of the improvement teams revealed a rich understanding of ‘interventions-in-systems’, recognising that the CMR was not an isolated activity, but dependent on its interaction with established processes and practices. Team actions were not limited by any predefined notion of what the intervention ‘should’ entail, and instead they actively sought to ensure the system was operating in a way that supported the intervention. These findings resonate with previous research which suggests interventions can be seen as ‘events in systems’ with importance placed on understanding how the intervention interacts with and disrupts the system it is deployed in.51 Findings also demonstrate that evidence operationalisation is not a one-off or time-bounded activity that disrupts the system, but instead an ongoing effort to address and maintain system performance aligning to the idea of ‘intervening in systems’.54

Consistency of function, adaptation of form

The improvement teams described intervention components and functions that were common across sites. Due to their interdependency, all component functions were seen as critical, and from this perspective a shared view of how to operationalise the CMR evidence base in practice was held between the five sites. However, while all components were perceived as necessary, all were adapted. This finding resonates with what has previously been defined as the ‘function and form’ of interventions: the ‘function’ represents a high-level statement of what an intervention is aiming to achieve (eg, increase competency), and the ‘form’ represents the specific form in which that intervention is delivered (eg, an educational workshop). Hawe et al 51 observe d that when working in complex systems, the ‘function’ is likely to be what can be standardised, whereas the ‘form’ will be likely to be adapted to meet the different needs of local systems. Our findings support this proposal: the component ‘functions’ were common across all sites, whereas the component ‘form’ always required adaptation.

Advancing the conceptualisation of interventions in complex systems

Based on our analysis, we conclude that the HC/SP construct is not well suited to conceptualising interventions in complex systems. We found the notion of core and periphery and the consideration of adaptation helpful but not conflation of the two. Previous HC/SP definitions tend to adopt a more reductionist view, focusing attention on the HC as the ‘fixed’, ‘essential’ or ‘active’ part of an intervention, and as such potentially trivialise the SP as ‘adaptable’, ‘supporting’, ‘inactive’ or ‘implementation activities’. This distinction risks minimising the attention paid to adaptable components as they are seen as non-active elements of the intervention. Our case and many others have shown that adaptations to interventions are not only common but often necessary for success.17 36 43 79–86 This suggests that the previous HC/SP perspectives have inappropriately conflated the concepts of effectiveness and adaptability by assuming if components can be adapted, they are not essential to the effectiveness of the intervention.44 These findings suggest the need to advance conceptualisation of interventions to more explicitly consider how an intervention is operationalised in complex systems. Drawing on insights from the improvement teams, we propose an updated intervention conceptualisation that builds on the HC/SP construct and proposes a more nuanced approach to intervention conceptualisation. The revised conceptualisation is summarised in box 1 and figure 2.
Figure 2

CMR intervention components by ’spheres of operation’. Each sphere represents the proximity of each component to the evidence base. Y-axis: hard core and soft periphery configurations based on the Medical Research Council (MRC) definition (green line), Greenhalgh et al definition (blue line) and Denis et al definition (red line). A solid line represents intervention components meeting the definition of hard core. A dotted line represents intervention components that could be classified as hard core or soft periphery. No line represents intervention components meeting the definition for soft periphery. CMR, comprehensive medication review; STOPP, Screening Tool of Older People’s Prescriptions.

Proximity to the evidence base: the ‘core’ and ‘periphery’ of an intervention can be conceptualised in relation to the proximity of the intervention components to a focal point of the initiative (eg, evidence base). This can be represented as ‘spheres of operation’ close to the core (Accessibility of evidence base, Process of enactment) as well as those more peripheral (Dependent processes and Dependent sociocultural issues). All contribute to successful outcomes. Component interdependence: the activities of the ‘core’ and the ‘periphery’ are interconnected, and understanding and articulating the interdependence of intervention components is critical for successful implementation. All components including those in the periphery are considered important to effective implementation and successful outcomes. Component function: identifying the ‘function’ of intervention components provides clarity on what each component is intending to achieve. Each function should describe how the component supports the operationalisation of evidence in routine practice. This function should be irreducible and generalisable across diverse settings. Collectively the component functions represent the requirements to operationalise evidence into practice. Component adaptation: assessing the degree and nature of adaptation is important in understanding how intervention components are effectively applied in local settings. Adaptation can take place between sites, within sites and over time in both the core and peripheral elements based on local resources, needs, constraints and infrastructure, as well as emerging evidence and site learning. Effort: significant effort is required to operationalise and adapt intervention components. There is a need to acknowledge the often neglected and undervalued labour required to enact change in practice including solving local problems, adapting interventions and motivating staff to engage in new ways of working. This effort is crucial to the success of intervention spread, and only by being explicit about the nature and amount of this effort can we manage expectations for the planning, conduct and evaluation of change efforts. CMR intervention components by ’spheres of operation’. Each sphere represents the proximity of each component to the evidence base. Y-axis: hard core and soft periphery configurations based on the Medical Research Council (MRC) definition (green line), Greenhalgh et al definition (blue line) and Denis et al definition (red line). A solid line represents intervention components meeting the definition of hard core. A dotted line represents intervention components that could be classified as hard core or soft periphery. No line represents intervention components meeting the definition for soft periphery. CMR, comprehensive medication review; STOPP, Screening Tool of Older People’s Prescriptions. Box 1 presents the revised conceptualisation of ‘interventions-in-systems’ and describes intervention components in relation to their: proximity to the evidence base; component interdependence; component function; component adaptation and effort. This conceptualisation maintains the notion of ‘core’ and ‘periphery’ as defined by the proximity of the intervention components to the evidence base. For example, ‘Accessibility of evidence base’ is directly related to the CMR and STOPP criteria, communicating the evidence to increase practitioner understanding and application. Whereas ‘Dependent sociocultural issues’ (the most distal category) had an indirect link to the evidence base yet were required to effectively deliver the CMR. The revised conceptualisation replaces the simplistic notion of ‘hard’ or ‘soft’ by recognising the interdependence and potential adaptation of all components, and the importance of defining component functions as the irreducible and generalisable aspects of an intervention (which collectively should describe the requirements to operationalise the evidence base into practice). We also emphasise the need to understand the effort required to adapt and embed intervention components in each spread site. Figure 2 presents this conceptualisation visually, picturing the evidence base as the core of the CMR intervention, surrounded by ‘spheres of operation’ and associated intervention components based on their proximity to the evidence base. Figure 2 also visually represents the challenge of applying the existing HC/SP definitions due to the ambiguity of each definition in categorising the ‘core’ or ‘peripheral’ components as either ‘hard’ or ‘soft’ (represented by solid and dotted lines). The inconsistency between the three definitions is represented by separate red, green and blue lines.

Strengths, limitations and future work

This study makes an empirical contribution by taking a systems view and drawing on ‘insider’ knowledge to provide a richer account of how interventions are spread across multiple sites. The findings from this work are complementary to building programme theory for evaluation of impact as they enable more detail description of intervention components and mechanisms of action that can support evaluation design. However, this study also has several limitations. A known weakness of prospective case study design is limited external validity. This means we cannot ascertain whether our findings are applicable and generalisable to other contexts and initiatives.62 87 For example, components might not fit within our ‘sphere’ classification, or different sites may have different views on which components are irreducible. We also recognise the sites we were working with had significant clinical expertise, QI experience and support, and dedicated time and resource for the initiative, which will have influenced the study findings. This is, however, inherent to the iterative process of theory building; future studies are needed to test and refine proposed theoretical contributions. The impact of ‘groupthink’ may have also been a limitation to our findings.88 Although there was strong consensus on the intervention components and functions proposed, groupthink may have led to fewer dissentions or identification of alternative options. Through the triangulation of multiple data sources, we aimed to mitigate this limitation. Future research is needed to assess how intervention adaptations, based on component function, impact intervention effectiveness. While intervention fidelity has been associated with positive outcomes,89 in line with others, our study suggests this might not be feasible or desirable in practice.51 The components’ functions described by this study provide a strong foundation to design and conduct more analytically rigorous studies to explore the attribution of adaptations to the overall success of the initiative. Indeed, such work would not be possible without this prior work to prepare a detailed understanding of the intervention components and their interactions.

Implications for research and practice

The approach taken in this research can provide guidance to support practitioners and researchers to study and describe interventions, to support spread and reproducibility of improvement initiatives. We propose a generic template for conceptualising interventions based on the methodology used in this study (online supplemental file 2). From a practical perspective, this template can support implementation teams to list intervention components, describe the component functions and the rationale for adaptation. This will help new sites avoid narrow artificially bounded intervention descriptions.90 91 From a research perspective, this will provide an opportunity to systematically build knowledge, test theories, and increase the scale, sophistication, rigour of evaluation approaches as interventions are spread to a greater number of sites.

Conclusion

The study makes theoretical and practical contributions towards the comprehensive conceptualisation of healthcare interventions. We demonstrate the value of co-producing knowledge of interventions between improvement teams and researchers through prospective study. This ‘insider’ knowledge revealed that interventions cannot be viewed as isolated activities, and their effective operation can only be achieved through understanding their interaction with established processes and practices. We conclude that to understand and describe interventions, we need to conceptualise them as ‘interventions-in-systems’. Adopting this view identifies the range of actions required to ensure interventions are ‘fit for purpose’ and that the surrounding systems are capable of supporting them. In applying HC/SP constructs in the context of the multisite CMR initiative, we encountered challenges in the binary conceptualisation of intervention components as either HC or SP. Therefore, we propose a more nuanced approach to intervention conceptualisation where components are described in respect to their: proximity to the evidence base (whether component parts are closer or further away from the core evidence base); component interdependence (how the components interact with and influence each other); component function (the purpose of each component which should be generalisable across sites); component adaptation (the degree and nature of adaptation in local settings) and effort (the amount of effort required to enact change in practice). Collectively, such a description should represent the work needed to operationalise evidence into practice in diverse settings, facilitating knowledge-sharing between sites, and reproducibility of successful improvement initiatives. Further work is needed to test and refine our proposed theoretical contributions to assess and strengthen their practical relevance.
  53 in total

1.  Designing quality improvement initiatives: the action effect method, a structured approach to identifying and articulating programme theory.

Authors:  Julie E Reed; Christopher McNicholas; Thomas Woodcock; Laurel Issen; Derek Bell
Journal:  BMJ Qual Saf       Date:  2014-10-15       Impact factor: 7.035

2.  Bridging the gap between prevention research and practice: the interactive systems framework for dissemination and implementation.

Authors:  Abraham Wandersman; Jennifer Duffy; Paul Flaspohler; Rita Noonan; Keri Lubell; Lindsey Stillman; Morris Blachman; Richard Dunville; Janet Saul
Journal:  Am J Community Psychol       Date:  2008-06

3.  The qualitative content analysis process.

Authors:  Satu Elo; Helvi Kyngäs
Journal:  J Adv Nurs       Date:  2008-04       Impact factor: 3.187

4.  Prevention of HIV/AIDS among injecting drug users in Russia: opportunities and barriers to scaling-up of harm reduction programmes.

Authors:  Elena Tkatchenko-Schmidt; Adrian Renton; Ruzanna Gevorgyan; Ludmila Davydenko; Rifat Atun
Journal:  Health Policy       Date:  2007-09-04       Impact factor: 2.980

5.  Twenty-Five Years of Groupthink Theory and Research: Lessons from the Evaluation of a Theory.

Authors: 
Journal:  Organ Behav Hum Decis Process       Date:  1998-02

6.  Understanding uptake of continuous quality improvement in Indigenous primary health care: lessons from a multi-site case study of the Audit and Best Practice for Chronic Disease project.

Authors:  Karen L Gardner; Michelle Dowden; Samantha Togni; Ross Bailie
Journal:  Implement Sci       Date:  2010-03-13       Impact factor: 7.327

7.  Explaining high and low performers in complex intervention trials: a new model based on diffusion of innovations theory.

Authors:  Heather McMullen; Chris Griffiths; Werner Leber; Trisha Greenhalgh
Journal:  Trials       Date:  2015-05-31       Impact factor: 2.279

8.  When complexity science meets implementation science: a theoretical and empirical analysis of systems change.

Authors:  Jeffrey Braithwaite; Kate Churruca; Janet C Long; Louise A Ellis; Jessica Herkes
Journal:  BMC Med       Date:  2018-04-30       Impact factor: 8.775

9.  Innovation sustainability in challenging health-care contexts: embedding clinically led change in routine practice.

Authors:  Graham P Martin; Simon Weaver; Graeme Currie; Rachael Finn; Ruth McDonald
Journal:  Health Serv Manage Res       Date:  2012-11

10.  Deprescribing medicines in the acute setting to reduce the risk of falls.

Authors:  Vanessa Marvin; Emily Ward; Alan J Poots; Katie Heard; Arvind Rajagopalan; Barry Jubraj
Journal:  Eur J Hosp Pharm       Date:  2016-08-19
View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.