| Literature DB >> 26101403 |
Damian Roland1,2.
Abstract
Extensive resources are expended attempting to change clinical practice; however, determining the effects of these interventions can be challenging. Traditionally, frameworks to examine the impact of educational interventions have been hierarchical in their approach. In this article, existing frameworks to examine medical education initiatives are reviewed and a novel '7Is framework' discussed. This framework contains seven linearly sequenced domains: interaction, interface, instruction, ideation, integration, implementation, and improvement. The 7Is framework enables the conceptualization of the various effects of an intervention, promoting the development of a set of valid and specific outcome measures, ultimately leading to more robust evaluation.Entities:
Keywords: Health expenditures; Health resources; Medical education; Outcome assessment
Year: 2015 PMID: 26101403 PMCID: PMC4536351 DOI: 10.3352/jeehp.2015.12.35
Source DB: PubMed Journal: J Educ Eval Health Prof ISSN: 1975-5937
The original domains of Kirkpatrick
| Level | Domain | Detail |
|---|---|---|
| 1 | Reaction | How well did the participants like the training? |
| 2 | Learning | What facts and knowledge were gained from the training? |
| 3 | Behaviour | Was the learning from the training utilised in the workplace? |
| 4 | Results | Did the training produce the overall intended benefits to the organisation? |
Fig. 1.The bridge of levels from reaction to results. This shows the process of understanding as the ‘bridge’ between the original intervention and the overall outcome [7]. Diagram was drawn by Damian Roland based on licence-free clip art.
Modification of Kirkpatrick’s domains by Barr et al.
| Level | Kirkpatrick's domain | Barr's modification |
|---|---|---|
| 1 | Reaction | No change |
| 2 | Learning | 2a: Modification of attitudes/perceptions |
| 2b: Acquisition of knowledge/skills | ||
| 3 | Behaviour | No change |
| 4 | Results | 4a: Change in organisational practice |
| 4b: Benefits to patients/clients |
Adapted from Barr et al. Evaluations of interprofessional education: a United Kingdom review for health and social care [Internet]. London: Centre for the Advancement of Interprofessional Education; 2000 [cited 2015 Apr 24]. Available from: http://caipe.org.uk/silo/files/evaluations-of-interprofessional-education.pdf [9].
Moore’s expanded outcomes framework
| Original (expanded) CME framework | Miller's framework | Description | Source of data |
|---|---|---|---|
| Participation (level 1) | The number of physicians and others who participated in the CME activity | Attendance records | |
| Satisfaction (level 2) | The degree to which the expectations of the participants about the setting and delivery of the CME activity were met | Questionnaires completed by attendees after a CME activity | |
| Learning (declarative knowledge level 3a) | Knows | The degree to which participants state what the CME activity intended them to know | Objective: pre- and post-tests of knowledge |
| Subjective: self-report of knowledge gain | |||
| Learning (procedural knowledge level 3b) | Knows how | The degree to which participants state how to do what the CME activity intended them to know how to do | Objective: pre- and post-tests of knowledge |
| Subjective: self-report of knowledge gain | |||
| Learning (competence level 4) | Shows how | The degree to which participants show in an educational setting how to do what the CME activity intended them to be able to do | Objective: observation in educational setting |
| Subjective: self-report of competence; intention to change | |||
| Performance (level 5) | Does | The degree to which participants do what the CME activity intended them to be able to do in their practices | Objective: observation of performance in patient care setting; patient charts; administrative databases |
| Subjective: self-report of performance | |||
| Patient health (level 6) | The degree to which the health status of patients improves due to changes in the practice behaviour of participants | Objective: health status measures recorded in patient charts or administrative databases | |
| Subjective: patient self-report of health status | |||
| Community health (level 7) | The degree to which the health status of a community of patients | Objective: epidemiological data and reports changes due to changes in the practice behaviour of participants | |
| Subjective: community self-report |
Reproduced from Moore et al. J Contin Educ Health Prof. 2009;29:1-15, with permission of Wiley [13].
CME, continuing medical education.
Hakkennes’ domains of evaluation
| Domains | Categories |
|---|---|
| Patient | Measurements of actual change in health status of the patient, i.e., pain, depression, mortality, and quality of life (A1) |
| Surrogate measures of A1, i.e., patient compliance, length of stay, and patient attitudes (A2) | |
| Health practitioner | Measurements of actual change in health practice, i.e., compliance with guidelines, changes in prescribing rates (B1) |
| Surrogate measures of B1, such as health practitioner knowledge and attitudes (B2) | |
| Organisational or process level | Measurements of change in the health system (i.e., waiting lists), change in policy, costs, and usability and/or extent of the intervention (C) |
Data from Hakkennes S, Green S. Implement Sci. 2006;1:29 [17].
Fig. 2.The range of epistemological approaches to evaluation. The researcher assesses the complexity of the outcomes and applies an appropriate framework dependent on the processes that may occur from intervention to outcome [19]. Diagram was drawn by Damian Roland.
Fig. 3.The hierarchal nature of the Kirkpatrick evaluation framework where level-four outcomes are of greater importance than those of level one [21]. Diagram was drawn by Damian Roland.
Fig. 4.A theoretical schema for evaluating outcomes of practice-changing interventions – The 7Is framework. Diagram was drawn by Damian Roland.
Description of the 7Is framework domain headings
| 7I Domain | Summary |
|---|---|
| Interaction | The degree to which participants engage with and are satisfied with the instruction |
| Interface | The degree to which participants are able to access the instruction |
| Instruction | The details of the intervention itself |
| Ideation | The perception of improvement following the instruction |
| Integration | The change, in both knowledge and behaviours, as a result of the instruction |
| Implementation | Whether change across individuals i.e., departments or organisations following the instruction has been demonstrated |
| Improvement | Whether the instruction has resulted in improvements in patient care and experience |
Steps to validate the original domains of the 7Is framework
| Domain | Study area |
|---|---|
| Interaction | A review of randomized control trials in medical education specifically looking at the concept of'intention to learn' would further validate this domain. A before and after study to demonstrate effectiveness by ensuring post-learning testing was undertaken by those not completing the intervention should be performed. Although an enforced post-learning element would introduce a level of bias, differences in the outcomes would suggest interaction analysis must be a fundamental part of evaluation. |
| Interface | The development of software, especially in light of e-learning studies, to examine the precise nature of how participants are able to, or are blocked from, accessing all modalities of a teaching package, would allow richer data in this domain to be examined. |
| Instruction | The development of taxonomy of medical education, and practice changing intervention, studies to allow valid comparisons between studies via the 7Is Framework. |
| Ideation | Further qualitative research into exploring junior doctors' understanding of competence and confidence and safety is required. It would be beneficial to repeat the meta-planning exercise on a different clinical issue (i.e., not in the field of pediatrics). If individual discriminatory concepts making up each of the terms could be validated, this would allow the creation of a questionnaire to assess and measure initial ideation. This would then allow a more detailed exploration of the proposed matrix linking the terms together and assess its practical use in a patient safety context. |
| Integration (knowledge) | If it is a case of premature ventricular contraction for example, a selection of purposefully designed disease for determining gold standard quality should be collated. A qualitative study in conjunction with this benchmarking exercise should take place to capture participants' decision making processes. This process would aim to improve the assessment of disease but may also guide future telemedicine studies to create minimum quality standards. |
| Integration (behaviour) | An observational study comparing case note review with observed interaction with patients would further validate the Rolma matrix. |
| Implementation and improvement | The results from the given study can inform effect sizes and a power calculation needed for a randomized control trial of the intervention. This would allow for an understanding of the relationship between implementation and improvement to be described. |