Literature DB >> 35003436

Five ways to get a grip on the shortcomings of logic models in program evaluation.

Betty Onyura1,2, Hollie Mullins1, Deena M Hamza3.   

Abstract

Logic models are perhaps the most widely used tools in program evaluation work. They provide reasonably straightforward, visual illustrations of plausible links between program activities and outcomes. Consequently, they are employed frequently in stakeholder engagement, communication, and evaluation project planning. However, their relative simplicity comes with multiple drawbacks that can compromise the integrity of evaluation studies. In this Black Ice article, we outline key considerations and provide practical strategies that can help those engaged in evaluation work to identify and mitigate some limitations of logic models.
© 2021 Onyura, Mullins, Hamza; licensee Synergies Partners.

Entities:  

Year:  2021        PMID: 35003436      PMCID: PMC8740248          DOI: 10.36834/cmej.71966

Source DB:  PubMed          Journal:  Can Med Educ J        ISSN: 1923-1202


For decades, logic models have been a quintessential tool for program evaluation. They provide relatively simple, diagrammatic representations of the ‘how’ and ‘why’ that underlies how specified interventions function.[1],[2] Often, they illustrate the strategic intentions behind interventions, which highlight the perceived linkages between planned activities, and anticipated outcomes[3]. Logic models are frequently used to guide communication with program stakeholders[1],[4] as well as to identify target areas for monitoring and evaluation.[5] However, given several limitations, some question whether they are an appropriate tool for current evaluation intitiaves.[6],[7]

Black Ice

As commonly used, logic models focus on succinct, visual synthesis of program components, and linearly depict program functioning. Correspondingly they can lead stakeholders to: Ignore the complexity that underlies social and educational interventions[6]. This can manifest, for example, through inattention to variable contextual factors (e.g., resourcing, political pressures) as well as to cultural and value differences that may compound to influence intervention outcomes. Assume that a given program is a legitimate solution to an identified problem; this can constrain thinking about external influences on the problem or limit exploration of adaptations to enhance contextual compatibility.[2],[8] For example, a serious oversight might involve an intervention to enhance youth self-esteem that does not acknowledge or attend to underlying systemic oppression of racialized groups. Neglect to identify the undesirable outcomes that interventions can inadvertently precipitate.[9] Even well-intended interventions can result in problematic effects.[10] Though unintended these effects are not necessarily unpredictable.[11],[12] Rather, adverse outcomes may be anticipated by purposeful attention to research evidence or experiential expertise.[11],[13] Examples of such effects include evidence that cultural competency training can in fact trigger expressions of racism toward cultural minority groups[14] or that global health educational innovations can increase trainees’ willingness to perform skills outside their scope of training.[15] Overall, these limitations make some evaluators hesitant to use logic models.[6] There are, however, useful strategies that can minimize these limitations. Here, we draw attention to ways to get a grip on these drawbacks of logic models while reaping the benefits they offer.

1. Articulate negative logic

Proactive work to identify how interventions may lead to undesirable outcomes or harm during logic modelling has been described as negative or dark logic modelling.[9] Supplementing a conventional logic model with one that outlines the potential downsides of an innovation (e.g., loss of autonomy and identity; challenging one’s self-efficacy) has several advantages. It allows for harm mitigation planning, while informing evaluation designs such that they proactively monitor undesirable and adverse outcomes. Further, it promotes sharing of balanced information about both risks and benefits associated with innovations.

2. Perform a logic analysis

A logic analysis provides a deeper level of scrutiny of select claims that are presented in logic models.[16],[17] A direct logic analysis is a formative process that examines whether the program design and implementation is consistent with available evidence about the critical conditions required for achieving desired effects.[16] A reverse logic analysis is a summative process that contrasts program characteristics with alternative models of realizing similar outcomes. Evidence for logic analyses can be found by synthesizing research evidence or soliciting experiential expertise.[18] Whereas both forms of logic analysis offer valuable checks of the credibility and trustworthiness of the claims posited in logic models,[19] they require substantial effort[16]. Thus, we recommend judicious selection of specific sub-sets of the logic model for logic analysis such as areas of recurrent implementation difficulties, key program principles, and/or values.[16]

3. Complement traditional logic models with more sophisticated designs

There have been numerous evolutions in evaluators’ approaches to visually illustrating the ‘hows and whys’ of intervention functioning. Thus, traditional logic models can be meaningfully complemented with more sophisticated approaches. For example, nested logic models incorporate multiple logic models that are linked through shared activities or outcomes and thus ‘nested’ into one another.[20] They compartmentalize the structural complexity of multi-program systems: the highest level model of a nested logic model appears concise, but is constituted by multiple, aligned ‘sub-logic models.’[2],[20] Action models are another contemporary option that are more expansive than traditional logic models. In addition to articulating pathways via which a program may influence specified outcomes, action models outline the prescriptive assumptions about how those pathways may be activated.[21] Specifically, they include details about implementation protocols, required characteristics of target participants and implementing organizations, essential capacities of front line implementers, as well as critical norms or resources in the program’s ecological context (i.e., culture, norms, and resources of the greater system, such as an academic institution or health services unit).[21] Action models can be optimal tools for guiding evaluations focused on implementation processes. Recent scholarship outlines a typology of logic models that includes a relatively advanced type four logic model. Type four logic models both 1) outline the change mechanisms of interventions (in lieu of listing discrete program activities and resources), and 2) illustrate the contextual factors upon which target outcomes depend.[6] The result is a more flexible type of logic model that allows for variation in activities and outcomes across contexts.[6]

4. Be clear about the purpose of the logic model

The rationale for using logic models should inform decisions about how to employ the technique.[6] For example, if the primary goal is to engage stakeholders and build consensus about program activities the simpler, conventional model may be optimal. The aim here is to capture stakeholder attention by being both aesthetically appealing and visually efficient[1] (i.e., maximizing accurate interpretation while minimizing cognitive effort[22]). Where evaluation of implementation is being prioritized, action models may be optimal, due to their expansive details on the prescriptions for program functioning. More complex models (e.g., Type 4 models) are not ideal for stakeholder engagement or consensus building [6]. Instead, they are instrumental for guiding reflections about how to improve the intervention’s use in organizational context.[6] INotably, they can be a valuable toolkit for evaluation-leads, who wish to improve or optimize program monitoring and evaluation designs.

5. Cyclically review and update logic models

Typical logic models often represent perceived rather than verified relationships among program activities and outcomes. Further, illustrated relationships may change due to evolving contexts and outcomes as programs are implemented.[7] As such, logic models should be treated as dynamic rather than static, with an expectation that they will evolve over time.[7] We recommend cyclical efforts to revisit and revise logic models as contexts changes (e.g., COVID-19 pandemic) and as evaluation data become available.

Conclusion

Logic models offer a pragmatic approach to synthesizing information about interventions that could otherwise be challenging to describe. Clarity about why a logic model is needed is critical to deciding how to go about creating one. The five strategies outlined above can help educators and evaluators get a grip on the limitations of logic models and maximize their utility.
  14 in total

1.  Why so many "rigorous" evaluations fail to identify unintended consequences of development programs: How mixed methods can contribute.

Authors:  Michael Bamberger; Michele Tarsilla; Sharlene Hesse-Biber
Journal:  Eval Program Plann       Date:  2016-01-07

2.  Do we really care about unintended outcomes? An analysis of evaluation theory and practice.

Authors:  Sumera Jabeen
Journal:  Eval Program Plann       Date:  2016-01-02

3.  Adverse effects of public health interventions: a conceptual framework.

Authors:  Theo Lorenc; Kathryn Oliver
Journal:  J Epidemiol Community Health       Date:  2013-12-13       Impact factor: 3.710

4.  Using Logic Analysis to Evaluate Knowledge Transfer Initiatives: The Case of the Research Collective on the Organization of Primary Care Services.

Authors:  Astrid Brousselle; Damien Contandriopoulos; Marc Lemire
Journal:  Evaluation (Lond)       Date:  2009-04

5.  Interfacing theories of program with theories of evaluation for advancing evaluation practice: Reductionism, systems thinking, and pragmatic synthesis.

Authors:  Huey T Chen
Journal:  Eval Program Plann       Date:  2016-06-07

6.  When 'unintended effects' reveal hidden intentions: Implications of 'mutual benefit' discourses for evaluating development cooperation.

Authors:  Niels Keijzer; Erik Lundsgaarde
Journal:  Eval Program Plann       Date:  2017-09-18

7.  Cultural minority students' experiences with intercultural competency in medical education.

Authors:  Hannah Leyerzapf; Tineke Abma
Journal:  Med Educ       Date:  2017-05       Impact factor: 6.251

8.  'Dark logic': theorising the harmful consequences of public health interventions.

Authors:  Chris Bonell; Farah Jamal; G J Melendez-Torres; Steven Cummins
Journal:  J Epidemiol Community Health       Date:  2014-11-17       Impact factor: 3.710

9.  Advancing complexity science in healthcare research: the logic of logic models.

Authors:  Thomas Mills; Rebecca Lawton; Laura Sheard
Journal:  BMC Med Res Methodol       Date:  2019-03-12       Impact factor: 4.615

10.  Building a framework for the evaluation of knowledge translation for the Canadian Network for Observational Drug Effect Studies.

Authors:  Ingrid S Sketris; Nancy Carter; Robyn L Traynor; Dorian Watts; Kim Kelly
Journal:  Pharmacoepidemiol Drug Saf       Date:  2019-02-20       Impact factor: 2.890

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.