| Literature DB >> 31324251 |
Alison Laycock1,2, Jodie Bailie3, Veronica Matthews3, Ross Bailie3.
Abstract
BACKGROUND: Developmental evaluation is a growing area of evaluation practice, advocated for informing the adaptive development of change initiatives in complex social environments. The utilisation focus, complexity perspective and systems thinking of developmental evaluation suggest suitability for evaluating knowledge translation initiatives in primary healthcare. However, there are few examples in the literature to guide its use in these contexts and in Indigenous settings. In this paper, we reflect on our experience of using developmental evaluation to implement a large-scale knowledge translation research project in Australian Aboriginal and Torres Strait Islander primary healthcare. Drawing on principles of knowledge translation and key features of developmental evaluation, we debate the key benefits and challenges of applying this approach to engage diverse stakeholders in using aggregated quality improvement data to identify and address persistent gaps in care delivery. DISCUSSION: The developmental evaluation enabled the team to respond to stakeholder feedback and apply learning in real-time to successfully refine theory-informed research and engagement processes, tailor the presentation of findings to stakeholders and context, and support the project's dissemination and knowledge co-production aim. It thereby contributed to the production of robust, useable research findings for informing policy and system change. The use of developmental evaluation appeared to positively influence stakeholders' use of the project reports and their responses to the findings. Challenges included managing a high volume of evaluation data and multiple evaluation purposes, balancing facilitative sense-making processes and change with task-focused project management, and lack of experience in using this evaluation approach. Use of an embedded evaluator with facilitation skills and background knowledge of the project helped to overcome these challenges, as did similarities observed between features of developmental evaluation and continuous quality improvement.Entities:
Keywords: Developmental evaluation; Indigenous; co-production; dissemination; knowledge translation; participatory research; quality improvement; stakeholder engagement
Year: 2019 PMID: 31324251 PMCID: PMC6642555 DOI: 10.1186/s12961-019-0474-6
Source DB: PubMed Journal: Health Res Policy Syst ISSN: 1478-4505
Fig. 1Systematically applying developmental evaluation in interactive dissemination cycles
Examples of evaluation feedback, team decisions and adaptations
| Evaluation findings | Team decisions and adaptations |
|---|---|
| Stakeholders in different roles had different information needs; some required summarised findings, others required detailed research reports | Report structure was adjusted to: 1-page key messages; 3-page executive summary; 25-page full report |
| The reports needed to be accessible and useful for a wide audience, including non-researchers; some stakeholders required more guidance to understand and use the reports | Report content was adjusted (e.g. to add an explanation of theory, to explain data trends, to add diagrams, to add a section on ‘how to use this report’) |
| Data tables were difficult for some stakeholders to understand and interpret | Presentation of health indicator and service delivery data was changed from table format to box-and-whisker-plot graphsa |
| Some stakeholders did not participate in the surveys because they lacked confidence in their data analysis skills | A link to an audio-visual resource was added to support the text explanation of how to interpret box-and-whisker-plot graphsa |
| Use of ‘academic-style’ language was a barrier to engaging with the reports | Plain language summaries were developed to accompany all subsequent reports |
| Some stakeholders did not participate in surveys because they perceived them to target those in other roles (e.g. policy officers perceiving the surveys targeted clinicians) | Statements on the advantages of participation by different professional groups were added to report summaries and emails |
| Some stakeholders found the surveys too long and/or considered the questions too repetitious | Survey questions were reduced in number across phase surveys; they were refined and reduced several times as the ESP project progressed |
| Many stakeholders who were motivated to participate had competing work demands and were time poor | Survey times were extended; email reminders were sent to encourage input |
| Those who participated in multiple ESP phases and cycles were committing considerable time; ‘survey fatigue’ was identified as a risk | Two project phases (one identifying barriers/enablers and one suggesting improvement strategies) were merged to reduce the number of phase surveys and reports in each ESP cycle |
| ESP emails were easily overlooked by key stakeholders due to high volumes of emails received | Coloured banners, photos and graphics were added to emails for more visual impact |
| ESP final reports were perceived to be large, partly because aggregated and trend data were included as appendices | Separate data supplements were published; they accompanied the ESP final reports |
| Expert input to data analysis and effective use of networks were important for stakeholder engagement | An expert/lead clinician in each area of care was asked to assist with data analysis, co-author ESP reports and disseminate the reports through their professional networks |
| CQI facilitators were vital for promoting the ESP project and facilitating local engagement | The team encouraged project communications and report dissemination through CQI practitioners/leaders in the CQI network |
| Indigenous stakeholders were mainly participating through group (rather than individual) survey responses | Group input was encouraged through project communications, in recognition of the critical importance of Indigenous input |
| A resource was needed to encourage and support group discussion and interpretation | A group facilitation guide was developed and promoted; links to the guide were embedded in report summaries and emails |
| One-page overviews of key findings in each area of care were suggested as a way of encouraging stakeholders to engage with the findings and access the ESP final reports | One-page overviews of key findings were produced and distributed |
| Key messages for action were needed to promote the use of ESP findings in each area of care | Key messages for action were developed from ESP findings; they were included in ESP reports and published as plain language summaries |
| Findings needed to be presented in a variety of formats to suit different work needs and learning styles | Findings were published online in all developed formats – reports, summaries, PowerPoint presentations, journal articles |
| ESP findings needed to be widely and easily accessible in the longer term; stakeholders intended to use the reports to resource future work tasks and information needs | ESP findings were published on research institution websites and in open web-based repositories for Indigenous health and policy publications |
a Box-and-whisker-plot graphs display the distribution of data based on the five-number summary: minimum, first quartile, median, third quartile, and maximum
CQI Continuous quality improvement, ESP Engaging Stakeholders in Identifying Priority Evidence–Practice Gaps and Strategies for Improvement in Primary Health Care