| Literature DB >> 29447410 |
Gareth Parry1,2, Astou Coly3, Don Goldmann1,2,4, Alexander K Rowe5, Vijay Chattu6, Deneil Logiudice7, Mihajlo Rabrenovic8,9, Bejoy Nambiar10,11.
Abstract
A lack of clear guidance for funders, evaluators and improvers on what to include in evaluation proposals can lead to evaluation designs that do not answer the questions stakeholders want to know. These evaluation designs may not match the iterative nature of improvement and may be imposed onto an initiative in a way that is impractical from the perspective of improvers and the communities with whom they work. Consequently, the results of evaluations are often controversial, and attribution remains poorly understood. Improvement initiatives are iterative, adaptive and context-specific. Evaluation approaches and designs must align with these features, specifically in their ability to consider complexity, to evolve as the initiative adapts over time and to understand the interaction with local context. Improvement initiatives often identify broadly defined change concepts and provide tools for care teams to tailor these in more detail to local conditions. Correspondingly, recommendations for evaluation are best provided as broad guidance, to be tailored to the specifics of the initiative. In this paper, we provide practical guidance and recommendations that funders and evaluators can use when developing an evaluation plan for improvement initiatives that seeks to: identify the questions stakeholders want to address; develop the initial program theory of the initiative; identify high-priority areas to measure progress over time; describe the context the initiative will be applied within; and identify experimental or observational designs that will address attribution.Entities:
Mesh:
Year: 2018 PMID: 29447410 PMCID: PMC5909656 DOI: 10.1093/intqhc/mzy021
Source DB: PubMed Journal: Int J Qual Health Care ISSN: 1353-4505 Impact factor: 2.038
Figure 1.Example of a Driver Diagram summarizing 'What' changes the initiative predicts will lead to the improvement goal.
Figure 2Example of a logic model Illustrating 'How' activities of the improvement initiative will facilitate local testing of the changes.
Examples of evaluation questions by improvement phase
| The What | The Context | The How |
|---|---|---|
| Innovation phase: Model development typically takes place in a small number of settings, and evaluation questions should focus largely on the What, for example: | ||
What is the overall impact of the model on health care quality and patient outcomes? Which elements of the model had the greatest impact on patient outcomes? | ||
| Testing phase: Testing phase, the aim is to identify where a model works or can be amended to work. Hence, although refining The What will occur, developing The How and The Context will also be important. Example evaluation questions include: | ||
What is the overall impact of the overall model on health care quality and patient outcomes? Which elements of the model had the greatest impact on patient outcomes? | To what extent can all the changes be implemented? What are barriers and facilitators to implementing the changes locally? What are the barriers and facilitators to undertaking the improvement activities as planned? | To what extent can all the changes be implemented? What are barriers and facilitators to implementing the changes locally? What are the barriers and facilitators to undertaking the improvement activities as planned? |
| Spread and scale-up phase | ||
What is the overall impact of the overall model on health care quality and patient outcomes? | To what extent did the impact of the model vary across settings? To what extent did the implementation vary from the model vary across settings? What contextual factors are associated with the implementation of the model? | To what extent can all the changes be implemented? What are barriers and facilitators to implementing the changes? What are the barriers and facilitators to undertaking the activities as planned? |
Core evaluation designs for assessing attribution
| Basic features | When to use |
|---|---|
| Factorial design | |
| Two or more interventions and a control group are compared. Participants (patients or sites) are randomized to each intervention independently | To compare two or more models of a multifaceted model in one or two settings, where the context is well-understood (e.g. comparing three improvement initiatives: one with coaching only, one with learning sessions only, and one with both coaching and learning sessions). |
| Stepped-wedge design | |
| Participants are assigned to an intervention or control group for a defined period. After this initial study period, control participants transfer into the intervention group. Can be randomized or non-randomized | To take advantage of the delayed implementation of an intervention in a representative sample of settings, and explore their use as comparator sites; or when an intervention is thought to be beneficial and/or when it is impossible or impractical to deliver the intervention to everyone at the same time |
| Controlled before and after study (CBA) | |
| Data are collected before and after the implementation of an intervention, both in one or more groups that receive the intervention and in a control group that is similar to the intervention group but that does not receive the intervention | This non-randomized design can be used to establish a counterfactual when random assignment is not ethical, possible or practical |
| Interrupted time series study (ITS) | |
| Data are collected at multiple time points before and after an intervention to determine whether the intervention has had an effect significantly greater than what would have been predicted by extending the baseline trend into the follow-up period | Where comparators are not available, a counterfactual can be estimated by the baseline trend (i.e. a historical control). The validity of this design can be strengthened by having a separate control group. Random allocation may be used to allocate to the intervention). |
| Cluster randomized controlled trials | |
| Randomization allocation of subjects or sites to an intervention or control group can be introduced if it is possible and appropriate. The internal validity of the above evaluation designs are stronger if study clusters such as health facilities are randomly assigned to intervention or control groups. Validity can be further strengthened by matching cluster groups on attributes associated with study outcomesa | |
aE.g. in a two-armed study, create pairs of health facilities with similar attributes, then randomly assign one facility per pair to the intervention group, with the remaining facility being a control.
Design considerations by improvement phase
| Innovation phase: |
When developing a multifaceted model in a small number of settings, with a well-understood context, approaches such a factorial design can be explored [ |
| Testing phase: |
Explore the possibility and potential implications of delaying implementation of the initiative in a representative sample of settings, and explore their use as comparator sites [ Consider whether the model can be disaggregated and assessed using a stepped-wedge design [ Where comparators are not available, a counterfactual may be available through use of longitudinal baseline data from participating sites, prior to participation in the initiative Where comparators are not available, use longitudinal data to assess for association in the take up of processes or activities with changes in outcomes In the testing phase, an approach to understanding where a model can be attributed to have worked or amended to work is important, and an exploration of the activity, process and outcome data across sites, and across specific contexts should be undertaken |
| Spread and scale-up phase: |
Explore the possibility and potential implications of delaying implementation of the initiative in a representative sample of settings, and explore their use as comparator sites When aiming to develop an approach for how a model can be implemented, consider whether the implementation model can be disaggregated and assessed using a stepped-wedge design Where comparators are not available a counterfactual may be available through use of longitudinal baseline data from participating sites, prior to participation in the initiative Where comparators are not available, compare within settings, activity and process data |
| For all improvement phases: |
If an experimental approach is not appropriate, the evaluation plan must offer an option for describing how attribution of outcomes can be assessed through triangulation of numerous analytical and research tasks |