Literature DB >> 30442094

Improving the normalization of complex interventions: part 2 - validation of the NoMAD instrument for assessing implementation work based on normalization process theory (NPT).

Tracy L Finch1, Melissa Girling2, Carl R May3, Frances S Mair4, Elizabeth Murray5, Shaun Treweek6, Elaine McColl2, Ian Nicholas Steen2, Clare Cook7, Christopher R Vernazza8, Nicola Mackintosh9, Samridh Sharma8, Gaery Barbery10, Jimmy Steele8, Tim Rapley11.   

Abstract

INTRODUCTION: Successful implementation and embedding of new health care practices relies on co-ordinated, collective behaviour of individuals working within the constraints of health care settings. Normalization Process Theory (NPT) provides a theory of implementation that emphasises collective action in explaining, and shaping, the embedding of new practices. To extend the practical utility of NPT for improving implementation success, an instrument (NoMAD) was developed and validated.
METHODS: Descriptive analysis and psychometric testing of an instrument developed by the authors, through an iterative process that included item generation, consensus methods, item appraisal, and cognitive testing. A 46 item questionnaire was tested in 6 sites implementing health related interventions, using paper and online completion. Participants were staff directly involved in working with the interventions. Descriptive analysis and consensus methods were used to remove redundancy, reducing the final tool to 23 items. Data were subject to confirmatory factor analysis which sought to confirm the theoretical structure within the sample.
RESULTS: We obtained 831 completed questionnaires, an average response rate of 39% (range: 22-77%). Full completion of items was 50% (n = 413). The confirmatory factor analysis showed the model achieved acceptable fit (CFI = 0.95, TLI = 0.93, RMSEA = 0.08, SRMR = 0.03). Construct validity of the four theoretical constructs of NPT was supported, and internal consistency (Cronbach's alpha) were as follows: Coherence (4 items, α = 0.71); Collective Action (7 items, α = 0.78); Cognitive Participation (4 items, α = 0.81); Reflexive Monitoring (5 items, α = 0.65). The normalisation scale overall, was highly reliable (20 items, α = 0.89).
CONCLUSIONS: The NoMAD instrument has good face validity, construct validity and internal consistency, for assessing staff perceptions of factors relevant to embedding interventions that change their work practices. Uses in evaluating and guiding implementation are proposed.

Entities:  

Keywords:  Complex interventions; Implementation process; Instrument development; NPT; NoMAD; Normalization process theory; Questionnaire

Mesh:

Year:  2018        PMID: 30442094      PMCID: PMC6238372          DOI: 10.1186/s12874-018-0591-x

Source DB:  PubMed          Journal:  BMC Med Res Methodol        ISSN: 1471-2288            Impact factor:   4.615


Introduction

Understanding implementation processes is key to ensuring that complex interventions in healthcare are taken up in practice and thus maximize intended benefits for service provision and (ultimately) care to patients [1]. Normalization Process Theory (NPT) [2] provides a framework for understanding how a new intervention becomes part of normal practice. This study aimed to develop and validate an adaptable survey instrument derived from NPT, to be used to improve the implementation of complex healthcare interventions within organisational settings. We know already that innovations in healthcare are themselves complex [3], and changing practice is difficult. There is a vast literature on implementation in this context and of the need for improvement in how change is implemented [4, 5], however the gap between research evidence and practice remains wide [6, 7]. It is now recognised that improved clinical and health outcomes for healthcare interventions are associated with successful implementation outcomes [1, 8]. Advancements in the measurement of implementation activity have been made recently [9], particularly in relation to the concept of organisational readiness [10-14]. Cook and colleagues [15, 16] have developed a set of measures based on Greenhalgh et al’s model of determinants of diffusion of service innovations [4]. A measurement model based on the Consolidated Framework for Implementation Research (CFIR, [17]) has also been offered, by Liang and colleagues [18]. Hodge and colleagues [19] have recently developed and tested an implementation sustainability measure, and work is also progressing to test a measure of implementation climate [13, 20]. Although these measures show some commonality in the representation of a range of factors affecting implementation success, they differ in their methods of development, and in the extent to which they are based on theoretical underpinnings [21]. Although these advances are promising, improvements in the development and testing of valid and reliable measures for assessing implementation process and outcomes [22] are called for and offer the potential to make important advances in implementation science [21]. This study extends Normalization Process Theory (NPT) [2, 23] towards improving implementation outcomes in the healthcare setting through the use of practical tools to aid implementation ‘work’. NPT provides a framework for understanding how a new intervention becomes part of normal practice, by emphasising the ways in which work must be reconfigured both individually and collectively by multiple stakeholders involved in the work of implementation. The potential for NPT to have far-reaching impact on academic and applied activity to improve the development of complex interventions that are well-placed to become effectively normalized in practice is evident, as demonstrated by the increasing volume of published research that has utilized NPT as a framework for evaluation studies. However, achieving this impact also requires more sophisticated (but simply administered) assessment measures to be developed, tested, and made available to user groups. To date, few studies have developed quantitative approaches to using NPT (May et al., Review of NPT studies, in preparation). The 16-statement interactive ‘toolkit’ developed in our previous work [24] (available at www.normalizationprocess.org) is presented as a tool to guide service planners, implementers and evaluators in thinking through their implementation problems. However, it is not developed as a research instrument, and nor was it validated for purposes of measuring aspects of implementation activity over time and across settings, as is often the objective of structured assessment in implementation research. The objectives of NoMAD were thus to: (1) develop a structured instrument for assessing activity related to the normalisation of complex interventions based on NPT, and (2) undertake initial psychometric testing of the instrument in terms of reliability and validity, across a sample of staff involved in the implementation of a range of interventions in different settings. We have reported the development methods and results in detail previously [25]. In this paper, we present the methods and results of the validation phase of the study, and the final NoMAD survey instrument. Guidance on application and adaptation of the instrument for different research and practice activities in provided.

Methods

A mixed methods approach was undertaken to develop, test and refine the NoMAD instrument using an iterative process. As the development methods are described in detail elsewhere [25], these will be briefly summarised here before detailing the methods used in the main survey validation phase of the project. First, the theoretical foundations for this work are described.

Theoretical underpinning

NPT [2] is concerned with the generative processes that underpin three core problems: implementation (bringing a practice or practices into action); embedding (when a practice or practices may be routinely incorporated in everyday work of individuals and groups); and integration (when a practice or practices are reproduced and sustained in the social matrices of an organization or institution). There are four generative processes and associated investments (see Table 1).
Table 1

NPT Construct definitions

ConstructDefinition
CoherenceSense-making that promotes or inhibits the coherence of a practice to its users. These processes are energized by investments of meaning made by participants
Cognitive participationParticipation that promotes or inhibits users’ enrolment and legitimisation of a practice. These processes are energized by investments of commitment made by participants.
Collective ActionActivity that promotes or inhibits the enacting of a practice by its users. These processes are energized by investments of effort made by participants.
Reflexive monitoringPractices that promote or inhibit users’ comprehension of the effects of a practice. These processes are energized by investments in appraisal made my participants.
NPT Construct definitions

Item generation and instrument development

Instrument development work focused primarily on generating and testing potential items to reflect each of the four constructs of NPT (coherence, cognitive participation, collective action and reflexive monitoring). An iterative process of instrument development was undertaken using the following methods: theoretical elaboration, item generation and item reduction (team workshops); item appraisal (QAS-99); cognitive testing with complex intervention teams; theory re-validation with NPT experts; and pilot testing of instrument [25]. An overview of the process is presented in Additional file 1. The NPT construct items utilised a 5 point scale of agreement for response (strongly agree; agree; neither agree nor disagree; disagree; strongly disagree) (Option A response). A set of ‘not relevant’ response options (not relevant to my role; not relevant at this stage; not relevant to [name of intervention]), termed ‘Option B responses’, were included following analysis of cognitive interview data to reflect reasons why participants may be unable to provide a genuine response on the Option A response scale on some items. This was important for validation of the NoMAD tool, in understanding how participants respond to the items. Three general ‘normalisation’ assessment items were also developed, through a combination of review of existing instruments, workshops, and consensus methods within the research team: When you use [intervention], how familiar does it feel? Do you feel that [intervention] is currently a normal part of your work? Do you feel that [intervention] will become a normal part of your work? These normalisation assessment items were rated on an 11 point scale (0–10, with appropriate descriptive anchors at 0, 5 and 10). The items were added to the NPT construct items to comprise 46 items for a version of NoMAD that was tested in the main validation study.

NoMAD validation study

Study participants

We sought to recruit at least 300 participants for reliability analysis [26]. A maximum variation approach to sampling of implementation projects was undertaken, seeking diversity in terms of the kinds of interventions being implemented, implementation timelines, and the professionals involved in implementation activity. Inclusion of a site in the project was conditional on sufficient numbers of staff participants to merit a survey approach (minimum of n = 20), and access to participants through an appropriate key contact within the site, who could support instrument administration. We sought a minimum of six implementation projects for inclusion.

Data collection

Participants in individual sites were identified and accessed through key contacts employed in the sites and invited to participate via email. At each site, instruments were administered either electronically (via SurveyMonkey Inc) or on paper, as advised by site contacts to allow the best chance of maximizing response rates [27]. As most data collection was conducted anonymously and using online methods, written consent was not deemed necessary for this study. As such, consent to participate was provided by individuals on choosing to complete the survey, as explained in the ethical committee approved participant information sheet explaining this procedure for consent, which accompanied all participation invitations. These procedures were approved by the Newcastle University Ethics Committee (Reference number 00555/2012; approval granted 1/09/2012). Site contacts worked with us to adapt the instrument appropriately to their intervention/setting, and issued invitations and reminders of behalf of the NoMAD team. At each site, at least one reminder was issued to all invited participants, within 2 weeks of the initial invitation.

Instrument refinement

Exploratory analysis was conducted on the full dataset to inform the retention of items for psychometric testing. Firstly, patterns of item responding were explored through descriptive statistics and frequencies, including checks for floor/ceiling effects, and through correlations amongst the full set of items. These descriptive analyses were combined with consensus methods within the research team, to agree the items retained for the final NoMAD instrument. Item retention was approached by considering each of the 16 sub-constructs as a set, aiming to retain at least one item per sub-construct. A summary table was produced to include descriptive data for each item in terms of: Relative strength of correlations with each other item within the sub-construct Whether the item correlated more highly with other sub-constructs within the main construct, than with items within its own sub-construct Whether there were notable correlations between the item and items outside the main construct The strength of correlation with the 3 general normalisation assessment items, and The level of ‘not relevant’ (option B) responding for the item Table 2 provides an example for the construct of ‘coherence’. Two steps were undertaken towards team consensus regarding retention of items for the final NoMAD survey. Firstly, four team members (TF, TR, MG & CM) independently made item retention selections on the basis of the summary data, with brief justifications for selection and a rating of the difficulty of the decision (0–5, from ‘easy’ to ‘difficult’). Secondly, these judgements were collated and distributed to wider team members for discussion in a full team consensus meeting (to include also EM, FM, ST), where consensus was achieved on the final set of items to be retained. The decision difficulty ratings were used to prioritise the consensus discussions, maximising the input from the wider team. The full set of retained items categorised by the 16 NPT sub-construct domains [2] is presented in Table 3.
Table 2

Item retention decision-making process - Example for Coherence

Sub-constructItemStrength of correlation with global items (high/mod/low)Correlation with other s-c items?Higher correlation with items in other s-c?Noteworthy correlation with other construct items?Level of option B responding? %
Differentiation1I can distinguish the [intervention] from usual ways of workingBest for ‘will become’ (0.35), v low for other items.Correlation of .70 with each other.NoNo8.9
Differentiation2I can see how the [intervention] differs from usual ways of workingSimilar to diff1 (0.29 highest).NoNo8.1
Communal spec1Staff in this organisation have a shared understanding of the purpose of the [intervention]Approx 0.30 across all 3.Moderate (0.59)NoIndivSpec1 (0.50)4.2
Communal spec2Staff in this organisation have shared expectations about the likelihood of the success of the [intervention]Low. Highest is 0.26 for ‘will become’.NoActivation2 (0.51). Correlations with appraisal items in RefMon (0.46–.50).5.0
Individual spec1I understand what tasks the [intervention] requires of meMod (0.40–0.47)Moderate (0.67)NoInternalization items (0.60 approx). CommSpec1 (0.50). Workability items (0.48&0.50). Legitimation items in CP (0.45 & 0.46)3.3
Individual spec2I understand how the [intervention] affects the nature of my own workMod (0.38–0.41)NoSimilar as for IndSpec1.3.1
Internalization1I can see the potential value of the [intervention] for my workVaries across 3 items. Mod (0.48) for ‘will become’, 0.34 for ‘is normal’, 0.29 for familiarity.High (0.83)NoIndSpec2, 0.61. Legit1, 0.61. Activ1, 0.65. Mod with appraisal items (RefMon), approx. 0.50–0.58 (IndApprais3).2.3
Internalization2I can see the worth of the [intervention] for meSimilar for items 1&2 (0.31 & 0.35). Mod (0.49) for ‘will become’.As for internalisation12.5
Table 3

Retained items by NPT sub-construct domains

ConstructSub-ConstructItems
Coherence Differentiation I can see how the [intervention] differs from usual ways of working
Communal specification Staff in this organisation have a shared understanding of the purpose of this [intervention]
Individual specification I understand how the [intervention] affects the nature of my own work
Internalization I can see the potential value of the [intervention] for my work
Cognitive Participation Initiation There are key people who drive the [intervention] forward and get others involved
Legitimation I believe that participating in the [intervention] is a legitimate part of my role
Enrolment I’m open to working with colleagues in new ways to use the [intervention]
Activation I will continue to support the [intervention]
Collective Action Interactional workability I can easily integrate the [intervention] into my existing work
Relational integration The [intervention] disrupts working relationships
Relational integration I have confidence in other people’s ability to use the [intervention]
Skill set workability Work is assigned to those with skills appropriate to the [intervention]
Skill set workability Sufficient training is provided to enable staff to use the [intervention]
Contextual Integration Sufficient resources are available to support the [intervention]
Contextual integration Management adequately support the [intervention]
Reflexive Monitoring Systemisation I am aware of reports about the effects of the [intervention]
Communal appraisal The staff agree that the [intervention] is worthwhile
Individual appraisal I value the effects the [intervention] has had on my work
Reconfiguration Feedback about the [intervention] can be used to improve it in the future
Reconfiguration I can modify how I work with the [intervention]
Item retention decision-making process - Example for Coherence Retained items by NPT sub-construct domains

Psychometric analysis

Psychometric analysis was conducted on the pooled dataset. Construct validity was explored through examination of the bivariate correlations between all possible pairs of construct items. We would expect items from the same construct to be more highly correlated with each other than with items measuring different constructs. Internal consistency was assessed using Cronbach’s alpha. A value of > = 0.7 is usually taken as indicative of adequate internal consistency. Data were subject to confirmatory factor analysis (CFA) which sought to investigate if the theoretically derived model approximated to the data. Since variables were judged to be unidimensional, parcelling was undertaken in order to maximise reliability [28] communality [29] and the value of the fit statistics [30] and to transform the ordinal data into a closer approximation to continuous data [31]. To create the parcels, the item-to-construct balance parcelling methodology was used [32]. Model fit was assessed by consulting the Tucker Lewis Index (TLI), Comparative Fit Index (CFI), Root Mean Square Error of Approximation (RMSEA) and the Standardised Root Mean Square Residual (SRMR). The primary aim was to achieve fit across all relevant indices. Specifically, we ideally sought a TLI and CFI of at least 0.90 and ideally 0.95 [33], an RMSEA below 0.07 [34] and an SRMR below 0.08. However in line with Hu and Bentler’s [35] guidelines a model acceptably approximates the data if a TLI or CFI of around 0.95 or an RMSEA up to around to 0.06 is observed alongside an SRMR up to around 0.08 [35]. The confirmatory factor analysis was undertaken within Mplus.

Ethics

Ethical approval for this study was granted by the University of Newcastle Ethics Committee (Reference number 00555/2012; approval granted 1/09/2012).

Results

Response

Six implementation projects contributed data for instrument validation, representing a variety of professional roles in relation to the interventions that were being implemented (for example, clinical, administrative, managerial, and other professionals in non-health contexts). Across these interventions, a total of 831 surveys were submitted. An overall response rate cannot be calculated as the denominator cannot be determined for one site (S6). Excluding S6, the response rate is 35% (495/1423). A breakdown of response data by site is provided in Table 4. Out of 831, 522 participants (63%) responded to one or more of the 43 NoMAD construct items, with 413 participants (50% of total sample) responding to all items. Excluding ‘Option B’ (‘not relevant’) responses, a total of 248 participants provided a likert scale (5 pt) response for all 43 construct items. Response rates were variable across items. Non-response at the individual item level ranged from 0.6% (n = 3) to 12% (n = 61). Response rates to option A (item deemed relevant for a likert rating) ranged from 75% (n = 389) to 97% (n = 508) across items.
Table 4

Response rates and item completion

DatasetMode of administrationInvitedRespondedRRTotal N/ InvitedNumber of items completedTotal completing 1–43 items
1–42 itemsAll 43 items
S1: Digital HealthElectronic + paper2316729%163753
S2: Smoking cessationPaper1002121%31821
S3: Patient self-managementElectronic4009123%265177
S4: Oral Health Risk AssessementPaper with personal approachElectronic (5th years)29722977%26194220
S5: System level ITElectronic3958722%145367
S6: Sports programmeElectronic – not targetedUnknown336Unknown246084
Total/ Overall>1423a83135%a109413522

aExcluding S6

Response rates and item completion aExcluding S6

Participant characteristics

Information about participants’ professional roles is provided in Tables 5 and 6.
Table 5

Description of study participants’ roles (N = 522) (% (n)

Site 1: Digital health record (N = 53)Site 2: Smoking cessation (N = 21)Site 3: Patient self-management tool (N = 77)Site 5: Technology implementation (N = 67)Site 6 (N = 84)
Professional RoleProfessional RoleProfessional RoleProfessional RoleOrganisation level
 Health visitor76 (40) Community Midwife67 (14) GP35 (27) Consultant24 (16) National Sporting organisation6 (5)
 Nursery nurse6 (3) Hospital midwife24 (5) Hospital Doctor5 (4) Trainee Doctor5 (3) State Sports Organisation (SSO)14 (12)
 Team leader2 (1) Stop smoking advisor5 (1) Consultant4 (3) Nurse Band 519 (13) League11 (9)
 Student health visitor9 (5) Stop smoking clerical/admin0 (0) Occupational Therapist1 (1) Nurse Band 622 (15) Club61 (51)
 Family nurse partnership8 (4) (missing)5 (1) Physiotherapist21 (16) Nurse Band 715 (10) Other8 (7)
 (missing)0 (0) Practice Nurse17 (13) Nurse Band 82 (1)
 Hospital Nurse9 (7) Admin clerical - secretary2 (1) Role in Organisation
 School Nurse1 (1) Manager5 (3) Administrator (Paid)20 (17)
 Dietician4 (3) Physiotherapist2 (1) Administrator (Volunteer)39 (33)
 Pharmacist3 (2) Dietician3 (2) Coach31 (26)
 missing0 (0) Pharmacist2 (1) Other10 (8)
 Other2 (1)
Main role in relation to the interventionMain role in relation to the interventionGeographic region (role not elicited)Main role in relation to the interventionMain role in relation to the intervention
 Champion13 (7) I am involved in managing or overseeing [name] intervention19 (4) Region 155 (42) I am involved in managing, overseeing or being a clinician28 (19) I’m involved in implementing [sports programme]57 (48)
 Promoter87 (46) I am involved in delivering the [name] intervention62 (13) Region 234 (26) I will be involved in working with the system of [name]69 (46) I’m involved in managing/overseeing [programme]43 (36)
 missing19 (4) Region 312 (9) Missing3 (2)
Years worked in the TrustYears worked in the TrustYears worked in the TrustYears worked in the Trust
 < 1 year15 (8) < 1 year10 (2) < 1 year0 (0) < 1 year10 (7)
 1–2 years15 (8) 1–2 years0 1–2 years1 (1) 1–2 years15 (10)
 3–5 years9 (5) 3–5 years5 (1) 3–5 years7 (5) 3–5 years19 (13)
 6–10 years21 (11) 6–10 years5 (1) 6–10 years10 (8) 6–10 years30 (20)
 11–15 years21 (11) 11–15 years24 (5) 11–15 years14 (11) 11–15 years15 (10)
 >  15 years19 (10) > 15 years52 (11) > 15 years68 (52) >  15 years10 (7)
 missing0 missing5 (1) missing0
Table 6

Site 4 Oral health risk assessment (Students and clinicians) descriptives

Site 4: Oral health risk assessmentStudent dentists (N = 189)Site 4: Oral health risk assessmentClinicians (N = 31)
Year of studyProfessional Role
 3rd Year40 (77) Clinical Fellow13 (4)
 4th Year41 (77) Senior lecturer7 (2)
 5th Year20 (37) Professor10 (3)
 Associate Clinical Lecturer55 (17)
 Clinical Trainer7 (2)
 StR/SpR7 (2)
 Missing3 (1)
Main role in relation to the interventionMain role in relation to the intervention
 I oversee others delivering [risk assessment] scoresYes: 84 (26)No: 3 (1)DA: 13 (4)
 I directly deliver the [risk assessment] scores to patientsYes: 94 (177)No: 6 (12) I directly deliver [risk assessment] scores to patientsYes: 42 (13)No: 42 (13)DA: 16 (5)
 I use and deliver [risk assessment] scores in another settingYes: 6 (12No: 74 (139)DA: 20 (38) I use and deliver [risk assessment] scores in another settingYes: 23 (7)No: 58 (18)DA: 19 (6)
Years worked in the [Dental Hospital]
 < 1 year10 (3)
 1–2 years10 (3)
 3–5 years19 (6)
 6–10 years16 (5)
 11–15 years13 (4)
 >  15 years32 (10)
Description of study participants’ roles (N = 522) (% (n) Site 4 Oral health risk assessment (Students and clinicians) descriptives

Relationships amongst NPT constructs

The NoMAD items were developed to represent pre-defined theoretical constructs. On this basis, we expected items within the four theoretical constructs (coherence, cognitive participation, collective action, and reflexive monitoring) to be more strongly related to each other, than to items in other constructs. Inspection of bivariate correlation matrices generally confirmed the expected pattern of relationships [available from authors on request]. Bivariate correlations between the NPT construct measures are shown in Table 7 revealing a moderate level of correlation for summated scores within the construct domains. This suggests that Coherence and Cognitive Participation are the most highly correlated (r = .68), and Reflexive Monitoring and Collective Action (r = .49) are the two constructs that are least correlated.
Table 7

Correlations between construct measures

CoherenceCognitive ParticipationCollective Action
Coherence1
Cognitive Participation.68 (n = 512)1
Collective Action.55 (n = 454).54 (n = 456)1
Reflexive Monitoring.60 (n = 427).59 (n = 428).49 (n = 423)

Pearson Correlation, all sig. (2 tailed) < .000

Correlations between construct measures Pearson Correlation, all sig. (2 tailed) < .000 Correlations between the construct measures scores and the overall normalisation score, with the general assessment items are shown in Table 8. Correlations were low to moderate. Of the three general assessment items, the construct measures appear to relate most strongly to perceptions that a new intervention will become a normal part of work.
Table 8

Correlations between construct measures and general assessment items

How familiar?Normal part of your work?Will it become normal?
Coherence.35 (n = 448).43 (n = 447).54 (n = 443)
Cognitive Participation.25 (n = 431).30 (n = 430).47 (n = 426)
Collective Action.41 (n = 409).48 (n = 408).45 (n = 404)
Reflexive Monitoring.26 (n = 388).28 (n = 387).40 (n = 383)
Normalisation score.41 (n = 417).48 (n = 416).58 (n = 412)

Pearson Correlation, all sig. (2 tailed) < .0

Correlations between construct measures and general assessment items Pearson Correlation, all sig. (2 tailed) < .0

Factor structure

The NoMad items were entered into a CFA to replicate the theoretical model. This item level CFA was run using the Weighted Least Squares Means and Variances (WLSMV) estimator which was considered appropriate given that the data were technically ordinal [36, 37]. The model showed reasonable fit (CFI = 0.91, TLI = 0.90, RMSEA = 0.09). Based on the factor loadings, the items were parcelled using the item-to-construct balance technique (see Table 9 for parcel composition). The CFA was then rerun using the ML estimator since the parcelled data more closely approximated continuous data. The resultant model showed a level of fit (CFI = 0.93, TLI = 0.89, RMSEA = 0.11, SRMR = 0.05) just short of acceptable [35]. The modification indices were therefore consulted and on this basis two correlated errors were modelled. The resultant model achieved acceptable fit (CFI = 0.95, TLI = 0.93, RMSEA = 0.08, SRMR = 0.03). Table 9 presents the standardized factor loadings.
Table 9

Confirmatory Factor Analysis (CFA) Parcel composition & Standardised indicator loadings

ParcelItemF1F2F3F4
Coherence parcel 1I can see the potential value of the [intervention] for my work I can see how the [intervention] differs from usual ways of working.74
Coherence parcel 2Staff in this organisation have a shared understanding of the purpose of this [intervention]I understand how the [intervention] affects the nature of my own work.67
Cognitive participation parcel 1There are key people who drive the [intervention] forward and get others involvedI will continue to support the [intervention].84
Cognitive participation parcel 2I believe that participating in the [intervention] is a legitimate part of my roleI’m open to working with colleagues in new ways to use the [intervention].86
Collective action parcel 1The [intervention] disrupts working relationshipsI can easily integrate the [intervention] into my existing work.67
Collective action parcel 2Work is assigned to those with skills appropriate to the [intervention]Sufficient resources are available to support the [intervention].57
Collective action parcel 3I have confidence in other people’s ability to use the [intervention]Sufficient training is provided to enable staff to use the [intervention]Management adequately support the [intervention].66
Reflexive monitoring parcel 1I am aware of reports about the effects of the [intervention]I value the effects the [intervention] has had on my work.55
Reflexive monitoring parcel 2The staff agree that the [intervention] is worthwhileFeedback about the [intervention] can be used to improve it in the futureI can modify how I work with the [intervention].70
Confirmatory Factor Analysis (CFA) Parcel composition & Standardised indicator loadings

Internal consistency of NPT construct subscales

Cronbach’s alpha was calculated for each of the four NPT construct groupings. Coherence consists of four items (α = 0.71); cognitive participation includes four items (α = 0.81); collective action comprises seven items (α = 0.78); and reflexive monitoring contains five items (α = 0.65). The normalisation scale overall (comprising items across all four constructs), was highly reliable (20 items, α = 0.89). Further information about item-total statistics is available from the authors.

Discussion

This paper presents NoMAD as a theoretically derived instrument for assessing implementation processes from the perspective of individuals involved in implementation activity. This further advances emerging work on the measurement of implementation processes [20–22, 36–40], by offering a theory-based measurement tool underpinned by Normalization Process Theory (NPT). To our knowledge, NoMAD represents the first systematic development and validation of a structured assessment tool based on the theoretical constructs of NPT [41]. NPT proposes that the embedding of new practices requires participants involved in the process to engage in work across four construct domains of coherence, cognitive participation, collective action and reflexive monitoring. The Confirmatory Factor Analysis (CFA) of the items we retained using descriptive data and consensus methods, supported this proposed theoretical structure. Tests of internal consistency supported the use of these items either as an overall measure of ‘normalisation’ (20 items, α = 0.89), or as four construct measures (ranging from α = 0.65–0.81). Except for reflexive monitoring, all constructs achieved the desired threshold of ≥0.07. Together with the moderate correlations between the four construct measures, the data supports the proposition that these are related but conceptually distinct domains within the theory. NoMAD has good face validity, and construct validity. NoMAD compares favourably against reviews of other implementation and research utilization measures, which generally rate the psychometric properties of instruments as poor to modest [22, 40], and show few (if any) measures to meet the full range of psychometric criteria being assessed. As a structured instrument, NoMAD offers scope to better understand the theoretical mechanisms of NPT by exploring, statistically, the relative importance of the NPT constructs in achieving sustained practice changes. NPT does not currently ascribe relative weightings to the importance of different construct domains for achieving the normalisation of a new practice, and the growing body of qualitative research framed by NPT [42] would suggest that the importance of the construct domains will vary according to the unique combination of intervention, context and human factors involved in the target practice change. A study by Jacobs and colleagues [20] to test a measure of Implementation Climate [13] found that although a common factor structure was supported in different organisational settings that was consistent with the theoretical specification (subscales on whether use of an innovation is rewarded, supported or expected within their organisation), the relevant strength of factor loadings of individual items varied across the settings. In their study, the settings differed in the extent to which participants worked primarily independently versus collaboratively, in ways that made sense in relation to the relative contexts. They concluded that indeed ‘context matters’ in measuring implementation climate, and argued that CFA models can still advance theory and knowledge of implementation even if they cannot be reasonably expected to fit other studies or contexts [20]. As authors of NoMAD, we wish to see further validation of the instrument in different settings, but caution that interpretation of results of different applications of NoMAD will need to be informed by understanding of important contextual features of the study context, often derived from qualitative work. We suggest therefore that NoMAD should be viewed as a ‘pragmatic measure’ of implementation [43] and encourage users to apply it flexibly to their implementation research and practice needs. Consistent with Glasgow and Riley’s [43] call for ‘pragmatic measures’ that balance psychometric considerations against the requirements of implementation in real life contexts, we believe that developing NoMAD with equal emphasis on (theoretical) content and face validity, and respondent usability (through cognitive interviews) has resulted in a measure that meets their required criteria of being important to stakeholders, low burden to respondents, and actionable (in that it can point to problems that can be further investigated or addressed in practice). Further, NoMAD also meets the recommended criteria of being broadly applicable (to a range of settings), unlikely to cause harm, related to theory, and psychometrically strong. A further required criterion of ‘sensitivity to change’ (often termed ‘responsiveness’ [39]) will be assessed in further validation work where multiple time point measures can be taken and compared. A review of implementation measures for community and public health settings found that all measures tested for responsiveness (7/51 measures reviewed), showed a minimum of moderate effect size (criterion of 0.5), suggesting the capacity of such measures to be responsive to change. In practical terms then, we offer NoMAD as an adaptable ‘bank of items’ that may be used flexibly by researchers or implementers [Additional file 2]. We anticipate that some will wish to use NoMAD as a complete instrument as presented here, with minimal adaptation besides adjustments for appropriate contextualisation of the survey to the target practice change. For these purposes, the validation data provided in this paper is important for the scientific endeavour of their work, as will other authors’ validation studies be important for further development of our work with NoMAD and with NPT. However in other applications of NoMAD, it may be that only certain items (or sets of items) may be relevant for use, and/or at particular stages of an implementation process. For example some items may be less useful in the very early stages of implementation, when no one has actually worked with the intervention or the intervention is not fully developed. The wording of the items may need to be altered, for example, for more anticipatory assessments. The extent to which the validation data reported in this paper can be expected to apply to future uses of NoMAD will therefore vary with the level of adaption made to the items, and will be a matter of the users’ own judgement. The NPT website (http://www.normalizationprocess.org) provides advice for how NoMAD may be used and adapted, but this is merely guidance. Consistent with this position, it is not possible to prescribe a formulaic process for scoring or combining items, or for the interpretation of the results that NoMAD generates, for all settings. In its simplest, most descriptive form, the underlying assumption of NoMAD is that more positive ratings by respondents of the implementation processes represented in each of the items are suggestive of higher potential for the practice to normalise. However this remains open to further assessment, and interpretation of results from the items used will always need to be undertaken in context. For example, whether an intervention ‘differs from usual ways of working’ may be of benefit to the implementation process in some contexts, but more problematic in others. For this reason, we present NoMAD as four sets of construct items, with reliability and validity data, and do not offer specific instructions for scoring or creating construct measures. Where assessments at the level of the construct are merited, items within the construct may be averaged to create ‘scores’ that may be compared amongst constructs, or between groups, or sites, if appropriate to the objective of the investigation. Further validation of NoMAD is required to assess how the NoMAD constructs relate to measures that are both theoretically similar (convergent validity) and conceptually distinct (discriminant validity). Appropriate measures for these purposes were not available at the time NoMAD was being developed, but a diverse range of implementation success indicators are now rapidly emerging and subject to critical review [39]. Existing implementation measures vary greatly but include tools that emphasise behavioural dimensions (eg. Theoretical Domains Framework) [44-46], organisational readiness [14, 47], implementation climate [20], research utilization in practice [39], and more focused constructs such as ‘implementation leadership’ [48]. It is likely that other measures more closely aligned with the focus of NPT will follow. Although defining and measuring concepts relating to implementation outcomes, or ‘success’ still remains complex [22, 49] due to the variety of theories and frameworks, studies that incorporate testing of sets of multiple measures will enhance our understanding of implementation process and outcomes. NoMAD should also be tested for criterion-related validity, to assess whether the construct measures are associated with implementation outcomes as proposed by NPT. These assessments can be undertaken in well designed longitudinal studies, that include robust implementation outcome measures [22]. We are currently undertaking such work in a range of contexts, including the ImpleMentAll study (http://www.implementall.eu/), which will use measures including NoMAD items,1 to assess the outcome of tailored implementation interventions in various settings across eight countries. This study provides support for NoMAD as an adaptable set of construct measures based on NPT. Key strengths include the underlying theoretical basis, and the iterative approach to item development and retention that prioritised construct validity, and usability from the perspective of a range of professionals involved in implementing complex interventions, in healthcare and other settings (education and sport). Current limitations however include validation work that is not yet undertaken, in terms of test-retest reliability, and convergent, discriminant and criterion-related validity. NoMAD is also limited to the frame of reference that NPT itself offers – a focus on the work that people undertake when implementing changes in practice, from the perspective of those involved in this work. Like any theory, or derivative instrument, its application must be appropriate to its frame of focus. NPT asks us to observe and understand social action, at an individual and collective level. As a self-report measure of individual’s perceptions of these processes, a fuller understanding of the embedding of a practice in any given setting is likely to require a combination of approaches that include quantitative and qualitative investigations. However, where larger scale implementation studies aim to compare implementation progress across multiple sites involved in an implementation project, and/or activity over time, NoMAD offers an appropriate tool to assess this. In some settings, NoMAD may be used as a diagnostic instrument, enabling data collection for local adaptation and improvement.

Conclusion

NoMAD provides a tool based on Normalization Process Theory (NPT), for the structured assessment of the work of implementation from the perspective of staff involved in implementing changes in practice. The key result of this study is the NoMAD instrument, along with validation data concerning statistical properties and other information that can be used to guide the application of the measures across different settings and for different purposes (designing, monitoring, and for trouble-shooting interventions and their implementation). As such, NoMAD should be viewed as a further tool in the NPT ‘toolkit’ (http://www.normalizationprocess.org), alongside the interactive 16 item tool [24] that was designed specifically for academic and non-academic users of NPT to think through their implementation problems, more as a ‘sensitizing’ device than as a validated tool for measuring implementation process and outcomes. We anticipate that although NoMAD is itself an important product of the study, the generation (and dissemination) of experience and knowledge in the application of these measures to real problems of implementation of complex health interventions in diverse settings will be the key to improving the design and implementation of interventions that are ultimately intended to benefit recipients of health care services. Overview of instrument development process. Provides process map of methods and data collected to develop and refine the NoMAD instrument. (DOCX 119 kb) Full NoMAD survey with adaptation guidance. Provides a copy of the full NoMAD survey instrument with guidance for adaptation. (DOCX 187 kb)
  34 in total

1.  Pragmatic measures: what they are and why we need them.

Authors:  Russell E Glasgow; William T Riley
Journal:  Am J Prev Med       Date:  2013-08       Impact factor: 5.043

2.  Outcomes for implementation research: conceptual distinctions, measurement challenges, and research agenda.

Authors:  Enola Proctor; Hiie Silmere; Ramesh Raghavan; Peter Hovmand; Greg Aarons; Alicia Bunger; Richard Griffey; Melissa Hensley
Journal:  Adm Policy Ment Health       Date:  2011-03

3.  Fostering implementation of health services research findings into practice: a consolidated framework for advancing implementation science.

Authors:  Laura J Damschroder; David C Aron; Rosalind E Keith; Susan R Kirsh; Jeffery A Alexander; Julie C Lowery
Journal:  Implement Sci       Date:  2009-08-07       Impact factor: 7.327

Review 4.  Psychometric properties of implementation measures for public health and community settings and mapping of constructs against the Consolidated Framework for Implementation Research: a systematic review.

Authors:  Tara Clinton-McHarg; Sze Lin Yoong; Flora Tzelepis; Tim Regan; Alison Fielding; Eliza Skelton; Melanie Kingsland; Jia Ying Ooi; Luke Wolfenden
Journal:  Implement Sci       Date:  2016-11-08       Impact factor: 7.327

5.  A Quantitative Test of an Implementation Framework in 38 VA Residential PTSD Programs.

Authors:  Joan M Cook; Stephanie Dinnen; Richard Thompson; Josef Ruzek; James C Coyne; Paula P Schnurr
Journal:  Adm Policy Ment Health       Date:  2015-07

Review 6.  A systematic review of the psychometric properties of self-report research utilization measures used in healthcare.

Authors:  Janet E Squires; Carole A Estabrooks; Hannah M O'Rourke; Petter Gustavsson; Christine V Newburn-Cook; Lars Wallin
Journal:  Implement Sci       Date:  2011-07-27       Impact factor: 7.327

7.  The meaning and measurement of implementation climate.

Authors:  Bryan J Weiner; Charles M Belden; Dawn M Bergmire; Matthew Johnston
Journal:  Implement Sci       Date:  2011-07-22       Impact factor: 7.327

8.  Organizational readiness for implementing change: a psychometric assessment of a new measure.

Authors:  Christopher M Shea; Sara R Jacobs; Denise A Esserman; Kerry Bruce; Bryan J Weiner
Journal:  Implement Sci       Date:  2014-01-10       Impact factor: 7.327

9.  Instrumentation issues in implementation science.

Authors:  Ruben G Martinez; Cara C Lewis; Bryan J Weiner
Journal:  Implement Sci       Date:  2014-09-04       Impact factor: 7.327

Review 10.  Achieving change in primary care--causes of the evidence to practice gap: systematic reviews of reviews.

Authors:  Rosa Lau; Fiona Stevenson; Bie Nio Ong; Krysia Dziedzic; Shaun Treweek; Sandra Eldridge; Hazel Everitt; Anne Kennedy; Nadeem Qureshi; Anne Rogers; Richard Peacock; Elizabeth Murray
Journal:  Implement Sci       Date:  2016-03-22       Impact factor: 7.327

View more
  44 in total

1.  Evaluation of Normalization After Implementation of the Digital Dutch Obstetric Telephone Triage System: Mixed Methods Study With a Questionnaire Survey and Focus Group Discussion.

Authors:  Bernice Engeltjes; Ageeth Rosman; Fedde Scheele; Christiaan Vis; Eveline Wouters
Journal:  JMIR Form Res       Date:  2022-06-17

2.  The implementation and validation of the NoMAD during a complex primary care intervention.

Authors:  Larkin Lamarche; Rebecca E Clark; Fiona Parascandalo; Dee Mangin
Journal:  BMC Med Res Methodol       Date:  2022-06-19       Impact factor: 4.612

3.  Increasing Children's physical Activity by Policy (CAP) in preschools within the Stockholm region: study protocol for a pragmatic cluster-randomized controlled trial.

Authors:  C Chen; V H Ahlqvist; P Henriksson; J H Migueles; F Christiansen; M R Galanti; D Berglind
Journal:  Trials       Date:  2022-07-19       Impact factor: 2.728

4.  Protocol for a type 2 hybrid effectiveness-implementation study expanding, implementing and evaluating electronic health record-integrated patient-reported symptom monitoring in a multisite cancer centre.

Authors:  Sofia F Garcia; Justin D Smith; Michael Kallen; Kimberly A Webster; Madison Lyleroehr; Sheetal Kircher; Michael Bass; David Cella; Frank J Penedo
Journal:  BMJ Open       Date:  2022-05-03       Impact factor: 3.006

5.  Stratified versus usual care for the management of primary care patients with sciatica: the SCOPiC RCT.

Authors:  Nadine E Foster; Kika Konstantinou; Martyn Lewis; Reuben Ogollah; Benjamin Saunders; Jesse Kigozi; Sue Jowett; Bernadette Bartlam; Majid Artus; Jonathan C Hill; Gemma Hughes; Christian D Mallen; Elaine M Hay; Danielle A van der Windt; Michelle Robinson; Kate M Dunn
Journal:  Health Technol Assess       Date:  2020-10       Impact factor: 4.014

6.  COVID-NURSE: evaluation of a fundamental nursing care protocol compared with care as usual on experience of care for noninvasively ventilated patients in hospital with the SARS-CoV-2 virus-protocol for a cluster randomised controlled trial.

Authors:  David A Richards; Holly Vr Sugg; Emma Cockcroft; Joanne Cooper; Susanne Cruickshank; Faye Doris; Claire Hulme; Phillipa Logan; Heather Iles-Smith; G J Melendez-Torres; Anne Marie Rafferty; Nigel Reed; Anne-Marie Russell; Maggie Shepherd; Sally J Singh; Jo Thompson Coon; Susannah Tooze; Stephen Wootton; Rebecca Abbott; Alison Bethel; Siobhan Creanor; Lynne Quinn; Harry Tripp; Fiona C Warren; Rebecca Whear; Jessica Bollen; Harriet A Hunt; Merryn Kent; Leila Morgan; Naomi Morley; Lidia Romanczuk
Journal:  BMJ Open       Date:  2021-05-26       Impact factor: 2.692

7.  Patients' and clinicians' perspectives on a 'fast-track' pathway for patients with sciatica in primary care: qualitative findings from the SCOPiC stratified care trial.

Authors:  Benjamin Saunders; Kika Konstantinou; Majid Artus; Nadine E Foster; Bernadette Bartlam
Journal:  BMC Musculoskelet Disord       Date:  2020-07-17       Impact factor: 2.362

8.  Normalization of a conversation tool to promote shared decision making about anticoagulation in patients with atrial fibrillation within a practical randomized trial of its effectiveness: a cross-sectional study.

Authors:  Gabriela Spencer-Bonilla; Anjali Thota; Paige Organick; Oscar J Ponce; Marleen Kunneman; Rachel Giblon; Megan E Branda; Angela L Sivly; Emma Behnken; Carl R May; Victor M Montori
Journal:  Trials       Date:  2020-05-12       Impact factor: 2.279

9.  Adaptation of the World Health Organization Electronic Mental Health Gap Action Programme Intervention Guide App for Mobile Devices in Nepal and Nigeria: Protocol for a Feasibility Cluster Randomized Controlled Trial.

Authors:  Tatiana Taylor Salisbury; Brandon A Kohrt; Ioannis Bakolis; Mark Jd Jordans; Louise Hull; Nagendra P Luitel; Paul McCrone; Nick Sevdalis; Pooja Pokhrel; Kenneth Carswell; Akin Ojagbemi; Eric P Green; Neerja Chowdhary; Lola Kola; Heidi Lempp; Tarun Dua; Maria Milenova; Oye Gureje; Graham Thornicroft
Journal:  JMIR Res Protoc       Date:  2021-06-15

Review 10.  Combining Realist approaches and Normalization Process Theory to understand implementation: a systematic review.

Authors:  Sonia Michelle Dalkin; Rebecca J L Hardwick; Catherine A Haighton; Tracy L Finch
Journal:  Implement Sci Commun       Date:  2021-06-26
View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.