Literature DB >> 31363715

Challenges and opportunities in evaluating programmes incorporating human-centred design: lessons learnt from the evaluation of Adolescents 360.

Aoife M Doyle1, Emma Mulhern2, James Rosen3, Gabrielle Appleford2, Christina Atchison1, Christian Bottomley1, James R Hargreaves1, Michelle Weinberger3.   

Abstract

Adolescents 360 (A360) is a four-year initiative (2016-2020) to increase 15-19-year-old girls' use of modern contraception in Nigeria, Ethiopia and Tanzania. The innovative A360 approach is led by human-centred design (HCD), combined with social marketing, developmental neuroscience, public health, sociocultural anthropology and youth engagement 'lenses', and aims to create context-specific, youth-driven solutions that respond to the needs of adolescent girls. The A360 external evaluation includes a process evaluation, quasi-experimental outcome evaluation, and a cost-effectiveness study. We reflect on evaluation opportunities and challenges associated with measuring the application and impact of this novel HCD-led design approach. For the process evaluation, participant observations were key to capturing the depth of the fast-paced, highly-iterative HCD process, and to understand decision-making within the design process. The evaluation team had to be flexible and align closely with the work plan of the implementers. The HCD process meant that key information such as intervention components, settings, and eligible populations were unclear and changed over outcome evaluation and cost-effectiveness protocol development. This resulted in a more time-consuming and resource-intensive study design process. As much time and resources went into the creation of a new design approach, separating one-off "creation" costs versus those costs associated with actually implementing the programme was challenging. Opportunities included the potential to inform programmatic decision-making in real-time to ensure that interventions adequately met the contextualized needs in targeted areas. Robust evaluation of interventions designed using HCD, a promising and increasingly popular approach, is warranted yet challenging. Future HCD-based initiatives should consider a phased evaluation, focusing initially on programme theory refinement and process evaluation, and then, when the intervention program details are clearer, following with outcome evaluation and cost-effectiveness analysis. A phased approach would delay the availability of evaluation findings but would allow for a more appropriate and tailored evaluation design. Copyright:
© 2019 Doyle AM et al.

Entities:  

Keywords:  Adolescents; Africa; Contraception; Design Thinking; Evaluation; Human-Centered Design; User-centered Design

Year:  2019        PMID: 31363715      PMCID: PMC6635668          DOI: 10.12688/gatesopenres.12998.2

Source DB:  PubMed          Journal:  Gates Open Res        ISSN: 2572-4754


Rationale

In the field of public health, human-centred design (HCD), also known as design thinking, is increasingly being used during intervention development, and sometimes implementation, to design innovative solutions to complex problems [1]. HCD is a creative approach to programme design that focuses on building empathy with the population of interest, and generating and iteratively testing many potential solutions to complex problems. The approach has a high tolerance for both ambiguity and failure [1– 3]. HCD shares some characteristics with other methods used to design programmes such as traditional socio-behavioural research but tends to be more ‘nimble/flexible/iterative’ and less protocol driven [4]. Intervention evaluation is an essential component of public health research and programming, and detailed guidance on evaluation approaches and methodologies exist [5, 6]. However, the iterative and flexible nature of HCD potentially poses some unique challenges for evaluation [1, 3, 7] and examples of evaluations of HCD public health interventions have so far been limited [7, 8]. There are some parallels between evaluations of HCD interventions and evaluations of interventions incorporating adaptive management or quality improvement where the intervention can also change over time [9– 14]. Also, programmatic interventions, as opposed to researcher-led interventions, often involve an element of flexibility in intervention implementation, as implementers strive to develop more context appropriate interventions, and respond to programmatic targets. In such interventions, the location and timing of intervention implementation is guided primarily by programme targets or local government priorities, and less so by a need to maximise the quality of evidence of impact [15, 16]. Despite some overlap with other evaluation scenarios, we have identified some unique challenges to the evaluation of HCD-led interventions.

Purpose of the letter

In this letter we discuss the evaluation opportunities and challenges identified during the ongoing A360 evaluation and make recommendations for future evaluations of programmes designed using HCD.

Adolescents 360

The intervention that is under evaluation is Adolescents 360 (A360), a multi-country initiative which aims to increase the voluntary use of modern contraceptives among adolescent girls aged 15–19 years. The A360 initiative is led by Population Services International (PSI) in collaboration with the Society for Family Health (SFH) in Nigeria, design partner IDEO.org, and the Centre for the Developing Adolescent at the University of California, collectively referred to here as the ‘implementers’. The implementers were all involved in the design process. PSI and SFH are responsible for the larger scale operationalising of the designed interventions. A360 is co-funded by the Bill & Melinda Gates Foundation and the Children’s Investment Fund Foundation. A360 set out to take a transdisciplinary approach to intervention development which involves the different discipline ‘lenses’ working jointly to create new innovations that integrate and move beyond discipline-specific approaches to address a common problem ( Figure 1).
Figure 1.

The A360 approach.

A360 aims to create context-specific, youth-driven solutions that respond to the needs of young people. Through an iterative design process, unique solutions were created in each of the A360 countries. In Ethiopia, Smart Start uses financial planning as an entry point to discuss contraception with newly married couples [17]. In southern Nigeria, 9JA girls provides physical safe spaces (branded spaces in public health clinics) for unmarried girls [18], and in northern Nigeria, Matasa Matan Arewa targets married girls and their husbands with counselling and contraceptive services using maternal and child health as an entry point [19]. In Tanzania ‘Kuwa Mjanja’ (be smart) delivers life and entrepreneurial skills training alongside opt-out counselling sessions and on-site contraceptive service provision to both married and unmarried adolescent girls [20].

External evaluation

The independent external evaluation aims to understand whether the A360 approach leads to improved reproductive health outcomes; generate evidence about A360’s cost-effectiveness in relation to other programming approaches; and capture how A360 is implemented in different contexts, and the experience of the participating young people and communities. It has four components ( Figure 2) [21]:
Figure 2.

A360 multi-component external evaluation.

Outcome evaluation which has a quasi-experimental design with before-after cross-sectional surveys in all four settings (Oromia, Ethiopia; Mwanza, Tanzania; Ogun and Nasarawa, Nigeria) and a comparison group in the two Nigerian settings [22]. Process evaluation incorporating both traditional qualitative methodologies and youth participatory research [23]. Costing study to understand the drivers of cost and a more extensive cost-effectiveness study within the outcome evaluation study geographies. Engagement and research uptake.

HCD process and timeline

HCD takes a phased approach to intervention development. In the ‘inspiration phase’ the designers immerse themselves with girls and their influencers to understand the issues. In the ideation phase they attempt to make sense of what has been learnt, identify opportunities for design and conduct an iterative cycle of prototyping of possible solutions and strategies. The most promising prototypes are then piloted. In the case of A360, implementation at scale involved a further ‘optimisation’ period where the selected solutions were further modified to maximise scalability and affordability. In A360, the inspiration and ideation phases took place over approximately 12 months (Sept 16–Aug 17) with piloting for a further 3–4 months (September–December 17). Optimisation and scale-up started towards the end of 2017/early 2018. Among other things, the length of the intervention design period presented some challenges for the evaluation, which we explore in the sections that follow.

Outcome evaluation

The outcome evaluation aimed to measure the impact of A360 on the voluntary use of modern contraceptives among adolescent girls aged 15–19 years in the study geographies. The outcome evaluation was designed when the implementers were at the ‘inspiration phase’ with final study protocols submitted for ethical approval when prototyping was still ongoing and before the intervention had been finalised ( Figure 3). This outcome evaluation timeline was necessary in order to conduct baseline surveys in 2017 prior to scale-up of the finalised interventions in the four study settings. The implications of the concurrent intervention development and evaluation design were that key pieces of information about the intervention components and the theory of change were unclear and changed over the protocol development period. Detailed information on the intervention were needed to inform the setting for baseline surveys and the kinds of questions to be asked during the surveys. The outcome evaluation study design was finalised before the interventions themselves so there remained a risk that the evaluation was not optimal.
Figure 3.

Timeline of A360 intervention development and implementation and the external evaluation.

A key consideration for the baseline survey design was who would be targeted with the A360 intervention in terms of geographical location and demographic characteristics. Early on in the intervention design process it became clear that A360 would be targeted towards only married or unmarried adolescent girls in some locations though uncertainty remained through early phases of the design process. For example, in Northern Nigeria, the outcome evaluation focused on married adolescent girls as the initial intention of A360 was to target only those who were married, however, 9ja Girls, the intervention that was designed in Southern Nigeria to target unmarried adolescent girls, was eventually also implemented in Northern Nigeria. In Ethiopia, initial HCD insights pointed towards a focus on unmarried adolescents in school but other programming pressures including donor preferences led to a later shift in focus to married rural women. The outcome evaluation was designed to target all married 15–19-year-old girls, but the final intervention focused on ‘newly married’ adolescent girls. While some change in target population might be expected with non-HCD programmatic interventions, the HCD process provided the implementers with more freedom and confidence to make additional changes. It was challenging to determine the geographies that would receive A360 as the implementers were in the middle of an intense phase of solution prototyping and were not yet thinking about larger scale implementation of A360. The level of implementation of the intervention was also relevant for study design, e.g. would A360 be implemented at health facilities, in schools or in the community? Given the uncertainty as to how A360 would be implemented, we opted for a community-based survey, which turned out to be appropriate for the final A360 interventions. However, if A360 had been implemented primarily in health facilities or schools then evaluating the intervention at the broader community level might not have been the most efficient study design. The uncertainties in key study design parameters meant that we had to develop multiple study design scenarios which were repeatedly revised as new information came to light. For example, initially the intervention was to have distinct implementation phases as it was rolled out across Nigeria and we explored the idea of conducting a stepped-wedge trial. Following further discussions with the implementers, it became clear that there was no certainty that roll-out would be in a phased manner as they did not know yet what the intervention would comprise. In Tanzania, we explored whether a regression discontinuity design might be possible if the intervention were implemented in one or several of the existing demographic surveillance sites, however, given the uncertainty around the nature of the intervention, the implementers were reluctant to commit to implementing in those areas. This prolonged study design process entailed some added costs and was at times frustrating for all involved.

Process evaluation

The process evaluation was aligned to the HCD-driven phases of the design process for A360 and had the primary objective of presenting a descriptive and analytical account of how the implementation of A360 has played out in relation to its Theory of Change [23]. The process evaluation attempted to understand both the intervention development process and intervention implementation. During intervention development the process evaluation team faced challenges as the fast-paced, highly iterative HCD process meant that the ‘design energy’ i.e. how decisions were made at key points in the design process, often went undocumented [4]. This challenge was also noted in another evaluation of a HCD design process [8]. As a result, the process evaluation team needed to be flexible in order to align closely with the work plan of the implementers and methodologies such as direct participant observation were key to capturing the depth of the HCD process. The potential for research fatigue was observed among target community members as their views were solicited by both the implementers designing the intervention and the process evaluation team interested in understanding the HCD process from the participants’ point of view. The process evaluation team, therefore, needed to balance the importance of capturing the views of community members with the potential for research fatigue. During the design of the process evaluation, the intention was that the findings would feed into and inform the intervention design at key moments. However, there was limited uptake of process evaluation findings by implementers. For example, the process evaluation highlighted the need for the programme to do more to address broader community and social norms, but this finding had a limited impact on intervention design. Poor uptake of findings was partly a result of the fast-paced nature of A360 and the resultant demands on the implementers’ time which did not allow them sufficient time to pause and reflect on the process evaluation findings. Country implementing teams differed in how they engaged with external recommendations, with some teams receptive and willing to listen and adjust, while others were more protective of their ‘solutions’. Uptake of the process evaluation findings improved when the evaluation team introduced participatory action research activities which focused on operational questions that were important for the country teams (e.g. health care provider attitudes in Ethiopia [24] and Nigeria [25]).

Cost-effectiveness evaluation

The cost-effectiveness study faced similar challenges as the outcome evaluation in terms of complicated and delayed development of the cost-effectiveness study protocols. Although all costing exercises face some unknowns about the intervention to be costed, the flexible and iterative HCD process increased the number of unknowns. Furthermore, because proponents hypothesized that it was the design process itself that would be the key factor in producing an effective (and cost-effective) intervention, an important challenge was both measuring the total design cost and isolating the cost of HCD specifically as HCD activities of the design process were tightly intertwined with the other A360 ‘lenses’. In addition, because A360 was creating a new trans-disciplinary approach, there was also interest in separating out the costs associated with the ‘creation’ of the A360 approach from the costs associated with implementing and scaling A360 in countries. Interviews were held with intervention staff in order to get a sense of the distribution of activities (and associated costs) across the study ‘lenses’ and between efforts to create versus implement A360. The implementing agencies planned continuous changes or tweaks to the intervention once it was up and running and so more frequent cost data draws were required during the scale-up period to capture how those changes might affect the “production process” and associated costs.

Lessons learnt

Robust evaluation of a new and promising approach such as HCD is warranted yet challenging. At the start of the programme, there was a clear methodologic gap between the evaluation team who had a background in public health and the social sciences, and the design team at IDEO who were leading the HCD process. However, PSI and the other implementers, who were working closely with IDEO, also had a background in public health and the social sciences and helped bridge this gap. Also, the evaluation team had been thoroughly briefed about the HCD methodologies that would be used but they had only limited practical experience evaluating programmes incorporating human-centered design. Some of the challenges faced are not unique to interventions designed with HCD and will be recognisable by those who have conducted interdisciplinary research and/or led evaluations of programmatic interventions. In comparison to research-led studies, evaluation of programmatic interventions is associated with a reduced level of control and increased uncertainty. A challenge for evaluators is to find ways to deal with this uncertainty while still retaining scientific rigor. An additional long-standing challenge, not specific to HCD, is to convince implementers that rigorous evaluation, while at times intrusive, will improve programme design and implementation. We have the following recommendations for others who would like to evaluate programmes incorporating HCD: Evaluators and implementers/designers should take time to familiarise themselves with the methodologies used by the different disciplines and have open discussions about the potential challenges of interdisciplinary research, and how they will be addressed/mitigated against. Implementers should allow adequate time to participate in the process evaluation, as well as work with the process evaluation team to ensure that findings are timed to feed into key decisions. The process evaluation team should maximise secondary analysis of data collected by the implementers, and joint data collection could be considered where additional data collection is needed and participant research fatigue is anticipated. Like HCD, an iterative and adaptive process evaluation approach is required. In A360, the process evaluation paused after the pilot phase and our team worked with the A360 implementers and donors to develop evaluation questions that reflected the solutions that were being developed, reflecting an iterated A360 Theory of Change. Methodologies such as participant action research may identify process evaluation questions that are more relevant for the implementers. For the evaluators direct observations can be instrumental to capturing the fast-paced, highly-iterative HCD process, and to understand the ‘design energy’ i.e. how decisions were made at key points in the design process. To avoid a time-consuming and resource-intensive design process, future HCD-based initiatives should consider a phased evaluation approach: Conduct process evaluation during the HCD inspiration, ideation, and pilot phases. Wait until the implementers have a better understanding of the emerging programme and have finalised the target geographies, target population, and intended outcomes before planning an outcome evaluation and cost-effectiveness study. During the implementation phase, conduct a comprehensive process evaluation that can capture whether, how, and why the intervention changed during implementation. The advantages of a phased approach need to be balanced against the disadvantages of lengthening of the time between implementation and the availability of evaluation findings.

Data availability

No data are associated with this article. The authors have been very responsive in their revisions and I have no further concerns. Again, this article makes a valuable contribution to the literature. I confirm that I have read this submission and believe that I have an appropriate level of expertise to confirm that it is of an acceptable scientific standard. The article presents lessons learnt from an external evaluation of the A360 programme in three African countries. A360 uses the human-centred design and hence applied three types of evaluations to evaluate it: whether the programme leads to improved reproductive health outcomes (outcome evaluation), capture how it is implemented (process evaluation) and study its cost-effectiveness. Additionally engagement and research uptake. The article honestly presents the known tension between effective implementation of a context-specific and flexible interventions versus an experimental evaluation design. It is refreshing to read an honest reflection on this, and as someone who has done similar evaluations, it clearly recognize the challenges. However, it would be even better if more specific recommendations would have been given on how to actually deal with this tension. Limited uptake of the process evaluation results by the implementers: it would have been interesting to understand why some implementers were eager to take on recommendations and others were not, and how this affected the effectiveness. Cost-effectiveness: since the effectiveness study was somewhat compromised, it is difficult to link it to the costs. It feels as if the cost-effectiveness study mainly mapped the costs, without linking it to the effectiveness. Reflections: The lessons learnt are important, but stem from common sense and are not new. While it is often clear what should be done, the challenge remains on how to do it: how to convince donors and implementers of the importance of proper evaluations and its consequences for the programme design and implementation. Cost-effectiveness of the evaluation: a lot of time and energy was given to do three types of evaluations. However, it seems as it there value was not optimal. It may be interesting to reflect on the cost-effectiveness of the evaluation itself. How were the different evaluations lied together? They seem to be presented quite separately, while for a clear understanding of the programme, they should be tied together. How is the evaluation approach chosen better/worse than e.g. realist evaluation, which seems quite appropriate for this type of intervention/programme. I confirm that I have read this submission and believe that I have an appropriate level of expertise to confirm that it is of an acceptable scientific standard. The article presents lessons learnt from an external evaluation of the A360 programme in three African countries. A360 uses the human-centred design and hence applied three types of evaluations to evaluate it: whether the programme leads to improved reproductive health outcomes (outcome evaluation), capture how it is implemented (process evaluation) and study its cost-effectiveness. Additionally engagement and research uptake. The article honestly presents the known tension between effective implementation of a context-specific and flexible interventions versus an experimental evaluation design. It is refreshing to read an honest reflection on this, and as someone who has done similar evaluations, it clearly recognize the challenges. However, it would be even better if more specific recommendations would have been given on how to actually deal with this tension. Limited uptake of the process evaluation results by the implementers: it would have been interesting to understand why some implementers were eager to take on recommendations and others were not, and how this affected the effectiveness. Cost-effectiveness: since the effectiveness study was somewhat compromised, it is difficult to link it to the costs. It feels as if the cost-effectiveness study mainly mapped the costs, without linking it to the effectiveness. Reflections: The lessons learnt are important, but stem from common sense and are not new. While it is often clear what should be done, the challenge remains on how to do it: how to convince donors and implementers of the importance of proper evaluations and its consequences for the programme design and implementation. See previous response. Cost-effectiveness of the evaluation: a lot of time and energy was given to do three types of evaluations. However, it seems as it there value was not optimal. It may be interesting to reflect on the cost-effectiveness of the evaluation itself. How were the different evaluations lied together? They seem to be presented quite separately, while for a clear understanding of the programme, they should be tied together. How is the evaluation approach chosen better/worse than e.g. realist evaluation, which seems quite appropriate for this type of intervention/programme. In this article, the authors discuss challenges and opportunities associated with evaluating HCD interventions. In this case, the intervention was associated with A360. Overall, the honesty of the authors is appreciated as rarely does one get to hear about problems that occurred during project implementation. Reporting on successes and failures is important as it provides an opportunity for others to learn. Nevertheless, the descriptions of the challenges that arose were a little cryptic and it was not always clear. I was also disappointed that the authors did not reach another logical conclusion which is that there was a methodologic gap between the public health researchers/evaluators and designers which lead to the challenges. Conducting interdisciplinary research can be challenging and requires each team to be very familiar with the other’s methods and processes so they can carefully design the collaboration. For example, when the study populations grow frustrated with the multiple inquiries it suggests that the data collection was not harmonized across teams. Similarly, the breakdown in communication emerges. These are the classic challenges of interdisciplinary research and may not only apply to HCD. The authors might discuss how they prepared to work with human centered designers or vice versa. What type of training took place and how familiar were the evaluators with the design process or were the designers with public health methods? Specific comments are as follows: It is interesting that the intervention design phase appears long. If one were to compare it to drug development or another multiphase intervention development process, then two years would be quite short. It is not clear why baseline data were collected prior to developing the intervention. How does collecting data in the four settings, the explanation provided here, constrain when data are collected? In looking at the diagram of the timeline, it seems that the baseline survey could have been moved up to July 2018 once all decisions were made regarding what the study would be about. The process evaluation is very interesting as is the objective of including insights into the final intervention. However, I was not clear what was meant by including PAR improved uptake. Is the author saying that they collected data using PAR that was of interest to the designers? If so, that would suggest that a recommendation might be that each team needs more familiarity with the array of methods available for process research prior to starting the program so data collection could be designed in collaboration. Cost effectiveness - it is not clear what the authors mean by a layer of uncertainty created by HCD. Further, it is not clear why the cost of HCD was being assessed rather than the cost of the intervention. Lessons learnt, item 5 is very important as is anticipating issues of interdisciplinary research from the beginning. I confirm that I have read this submission and believe that I have an appropriate level of expertise to confirm that it is of an acceptable scientific standard, however I have significant reservations, as outlined above. In this article, the authors discuss challenges and opportunities associated with evaluating HCD interventions. In this case, the intervention was associated with A360. Overall, the honesty of the authors is appreciated as rarely does one get to hear about problems that occurred during project implementation. Reporting on successes and failures is important as it provides an opportunity for others to learn. Nevertheless, the descriptions of the challenges that arose were a little cryptic and it was not always clear. I was also disappointed that the authors did not reach another logical conclusion which is that there was a methodologic gap between the public health researchers/evaluators and designers which lead to the challenges. Conducting interdisciplinary research can be challenging and requires each team to be very familiar with the other’s methods and processes so they can carefully design the collaboration. For example, when the study populations grow frustrated with the multiple inquiries it suggests that the data collection was not harmonized across teams. Similarly, the breakdown in communication emerges. These are the classic challenges of interdisciplinary research and may not only apply to HCD. The authors might discuss how they prepared to work with human centered designers or vice versa. What type of training took place and how familiar were the evaluators with the design process or were the designers with public health methods? Specific comments are as follows: It is interesting that the intervention design phase appears long. If one were to compare it to drug development or another multiphase intervention development process, then two years would be quite short. It is not clear why baseline data were collected prior to developing the intervention. How does collecting data in the four settings, the explanation provided here, constrain when data are collected? In looking at the diagram of the timeline, it seems that the baseline survey could have been moved up to July 2018 once all decisions were made regarding what the study would be about. The process evaluation is very interesting as is the objective of including insights into the final intervention. However, I was not clear what was meant by including PAR improved uptake. Is the author saying that they collected data using PAR that was of interest to the designers? If so, that would suggest that a recommendation might be that each team needs more familiarity with the array of methods available for process research prior to starting the program so data collection could be designed in collaboration. Cost effectiveness - it is not clear what the authors mean by a layer of uncertainty created by HCD. Further, it is not clear why the cost of HCD was being assessed rather than the cost of the intervention. Lessons learnt, item 5 is very important as is anticipating issues of interdisciplinary research from the beginning. Overview This letter offers reflections on the authors’ experience evaluating an HCD-led global health program called A360. The letter is timely in that it raises questions that many global health programs have grappled with of late. I particularly appreciate the authors’ frankness as they describe disconnects between design/implementation work and research. On the whole I am very enthusiastic about this letter moving forward in the publication process. I would like to offer a few reflections of my own in the hope of helping the author’s improve their letter and of helping our research community identify better solutions to these deeply important challenges. Major comments This letter seems to focus on the relationship between health research and design practice, and my primary question has to do with how the authors’ experiences and conclusions might have differed if their project had emphasized linking health research and design research. To be clear, I’m not referring to the way that practitioners refer to doing ‘design research’ or ‘user research’ to investigate user experiences and contexts. By design research, I’m referring to the program of empirical study that has been carried out by human-centered design researchers in the fields of design, information systems, computer supported cooperative work (CSCW), human computer interaction, and design anthropology (and beyond) in the last few decades. These fields draw heavily on ethnographic, participatory, and action research methods that are familiar to many process evaluators in global health (no doubt including the authors of this letter). Design researchers in these fields also often draw on domain-specific craft skills that enable visual expression and prototyping, intervention development and implementation, and iteration based on ‘hands-on’ engagement with end users. That is to say, in fields with a more established tradition of design research, the researchers often are designers. For methods of contemporary importance, see e.g., action design research [1] and research through design [2]. For a paper that is older but perhaps more relevant to my comment here, see Spinuzzi’s [3] argument that participatory design is a design method, yet also is an action-research method. While some of these design-led research fields started with collaborations across disciplines (e.g. ethnographers and computer scientists founded CSCW in the early 1990s), today doctoral students typically learn to integrate design practices and research methods in the conduct of their scholarly endeavours. Integrating design practice with research method thus becomes less a matter of coordinating two teams within an organizational bureaucracy, and something more akin to how an individual ethnographer strives for reflexivity concerning her dual roles as participant and as observer. This seems to contrast with the research described in this letter, in which the evaluations were conducted by researchers who did not consider themselves as ‘doers’ of the design work (if I’m mistaken about this point, I’d welcome being corrected!). In light of the example set by human-centered design researchers in CSCW and adjacent fields, my more precise question is this: how might the challenges you encountered be reconsidered as having to do with the all too common gap between researchers and practitioners? In particular I’m thinking of these challenges: In my experience, these are common problems/risks wherever researchers try to plug into (but aren’t leading) a process led by practitioners, regardless of the field of practice. Often this is no fault of the practitioners - understanding and accommodating a scientific agenda is often beyond their remit. These gaps can be alleviated to some degree when there is one integrated design team that has the skills, methodological background, and remit to conduct a design process, produce evidence about user experiences to iteratively inform intervention development, and use this evidence to evaluate and write about their process (see e.g., [4] , [5] , [6] , [7]). The frustrating communications that resulted from a prolonged design process. Research fatigue among participants, and the importance of jointly collecting data for design and for research so that people need not be interviewed twice. The need for evaluators to be present, as participant-observers in design work in order to understand the design-energy and design decisions. Limited uptake of findings generated by process evaluators. Implementers/designers needing to set aside time to participate in a process evaluation. The importance of ensuring that the process evaluation is phased and iterative in a manner that matches the design work. To be fair, the approach outlined in this letter seems at present to be the norm rather than the exception in global health journals: enlisting a team of researchers, who are not designers, to evaluate design practitioners, who are not researchers. The reason I find this letter so valuable is that it surfaces a set of challenges that may be inherent in this approach, and the authors should be commended for their frankness. The authors will need to discern among themselves whether, based on their experience, they would encourage future evaluators of design-led health initiatives to include researchers from disciplines with more established traditions of integrating design research and practice. If this does resonate with your experience, I would offer two sources that you might reference that speak to a way forward for the design for global health community. The ISO standard on human-centered design [8] and more recent work on design for global health [9] emphasize that human-centered design is a highly multidisciplinary approach to research and practice - it is centrally concerned with the practical task of synthesizing diverse perspectives. This point might be used to argue that involving design researchers alongside the range of other research disciplines outlined in the A360 approach is a way forward for streamlining adoption of your five recommendations. As a final point, I would acknowledge that while the tradition of design research with which I am most familiar can well serve the role of process evaluation as it plays out in the global health field, outcomes evaluations and cost-effectiveness studies remain more the domain of public health researchers and health economists. Such studies can be conducted relatively straightforwardly if they focus on the interventions that resulted from a design process. While good enough for most policy purposes, that approach doesn’t address the question of what proportion/aspect of good outcomes we might attribute specifically to the application of human-centered design. Given the complexity of the design process I don’t know the answer here, and while the letter raises the question, the authors’ five recommendations don’t seem to address it either. One could follow the approach taken in evaluating the impact of applying behavior change theories in the design of communication interventions - meta-analyses of many studies with and without HCD. Such an approach would likely surface a further round of questions, unaddressed in this letter, about the extent to which the impacts of design work can/should be attributed to the design practitioner’s degree of expertise, or whether a project was structured to enable the design practice to thrive, versus something integral to human-centered design as a discipline. Minor Comments: It might be helpful if you clarified your use of the terms “implementer” and “designer” throughout the paper. Are you referring to implementers of the design work (i.e., designers), or to other implementers who were operationalizing the interventions that the designers designed, or both? The reference below to ‘design energy’ is intriguing; I think readers would appreciate slightly more explanation of what this refers to and in what sense the term is dealt with in the cited reference. “During intervention development the process evalu- ation team faced challenges as the fast-paced, highly iterative HCD process meant that the decision making process, the ‘design energy’, often went undocumented 4.” In the statement below, I wasn’t able to follow how starting with a blank slate may have blocked the the design team from drawing on public health evidence. Design researchers and ethnographers typically strive to approach observational fieldwork in a manner that minimizes the influence of preconceived notions, but this does not prevent them from drawing on relevant theories or evidence during the process of analysis. Was disregarding public health evidence explicitly described as typical of the design process (a point I would disagree with in general, but that may well have been true to your experience), or might it have just been another researcher to practitioner disconnect? “Designers add value to public health as they take a different approach when address- ing an issue, however, in the case of A360, a preference to work from a blank slate free of the constraints of public health evidence may have impacted on an ability to respond to the PE findings.” Typos “However, the iterative and flexible nature of HCD potentially poses some unique chal- lenges for evaluation 1,3,7.[ remove period] and examples of evaluations of HCD public health interventions have so far been limited 7,8.” “In the ‘inspiration phase’ the designers immerse themselves with girls and her their influencers to understand the issues.” “The OE study design was finalised before the interventions themselves so there remained a risk that the evaluation was not optimal.” I think this refers to Outcome Evaluation study design; it would be helpful if you could clarify on first use of the acronym. “distinct imple- mentation phases as it was rolled out across Nigeria” “because A360 was creating a new trans-disciplin ary approach” “more frequent cost data draws were required during the scale-up period to capture how those change s might affect the “production process” and associated costs.” I confirm that I have read this submission and believe that I have an appropriate level of expertise to confirm that it is of an acceptable scientific standard. Overview This letter offers reflections on the authors’ experience evaluating an HCD-led global health program called A360. The letter is timely in that it raises questions that many global health programs have grappled with of late. I particularly appreciate the authors’ frankness as they describe disconnects between design/implementation work and research. On the whole I am very enthusiastic about this letter moving forward in the publication process. I would like to offer a few reflections of my own in the hope of helping the author’s improve their letter and of helping our research community identify better solutions to these deeply important challenges. Major comments This letter seems to focus on the relationship between health research and design practice, and my primary question has to do with how the authors’ experiences and conclusions might have differed if their project had emphasized linking health research and design research. To be clear, I’m not referring to the way that practitioners refer to doing ‘design research’ or ‘user research’ to investigate user experiences and contexts. By design research, I’m referring to the program of empirical study that has been carried out by human-centered design researchers in the fields of design, information systems, computer supported cooperative work (CSCW), human computer interaction, and design anthropology (and beyond) in the last few decades. These fields draw heavily on ethnographic, participatory, and action research methods that are familiar to many process evaluators in global health (no doubt including the authors of this letter). Design researchers in these fields also often draw on domain-specific craft skills that enable visual expression and prototyping, intervention development and implementation, and iteration based on ‘hands-on’ engagement with end users. That is to say, in fields with a more established tradition of design research, the researchers often are designers. For methods of contemporary importance, see e.g., action design research   and research through design  . For a paper that is older but perhaps more relevant to my comment here, see Spinuzzi’s  argument that participatory design is a design method, yet also is an action-research method. While some of these design-led research fields started with collaborations across disciplines (e.g. ethnographers and computer scientists founded CSCW in the early 1990s), today doctoral students typically learn to integrate design practices and research methods in the conduct of their scholarly endeavours. Integrating design practice with research method thus becomes less a matter of coordinating two teams within an organizational bureaucracy, and something more akin to how an individual ethnographer strives for reflexivity concerning her dual roles as participant and as observer. This seems to contrast with the research described in this letter, in which the evaluations were conducted by researchers who did not consider themselves as ‘doers’ of the design work (if I’m mistaken about this point, I’d welcome being corrected!). In light of the example set by human-centered design researchers in CSCW and adjacent fields, my more precise question is this: how might the challenges you encountered be reconsidered as having to do with the all too common gap between researchers and practitioners? In particular I’m thinking of these challenges: In my experience, these are common problems/risks wherever researchers try to plug into (but aren’t leading) a process led by practitioners, regardless of the field of practice. Often this is no fault of the practitioners - understanding and accommodating a scientific agenda is often beyond their remit. These gaps can be alleviated to some degree when there is one integrated design team that has the skills, methodological background, and remit to conduct a design process, produce evidence about user experiences to iteratively inform intervention development, and use this evidence to evaluate and write about their process (see e.g.,  , , , ). The frustrating communications that resulted from a prolonged design process. Research fatigue among participants, and the importance of jointly collecting data for design and for research so that people need not be interviewed twice. The need for evaluators to be present, as participant-observers in design work in order to understand the design-energy and design decisions. Limited uptake of findings generated by process evaluators. Implementers/designers needing to set aside time to participate in a process evaluation. The importance of ensuring that the process evaluation is phased and iterative in a manner that matches the design work. To be fair, the approach outlined in this letter seems at present to be the norm rather than the exception in global health journals: enlisting a team of researchers, who are not designers, to evaluate design practitioners, who are not researchers. The reason I find this letter so valuable is that it surfaces a set of challenges that may be inherent in this approach, and the authors should be commended for their frankness. The authors will need to discern among themselves whether, based on their experience, they would encourage future evaluators of design-led health initiatives to include researchers from disciplines with more established traditions of integrating design research and practice. If this does resonate with your experience, I would offer two sources that you might reference that speak to a way forward for the design for global health community. The ISO standard on human-centered design  and more recent work on design for global health   emphasize that human-centered design is a highly multidisciplinary approach to research and practice - it is centrally concerned with the practical task of synthesizing diverse perspectives. This point might be used to argue that involving design researchers alongside the range of other research disciplines outlined in the A360 approach is a way forward for streamlining adoption of your five recommendations. As a final point, I would acknowledge that while the tradition of design research with which I am most familiar can well serve the role of process evaluation as it plays out in the global health field, outcomes evaluations and cost-effectiveness studies remain more the domain of public health researchers and health economists. Such studies can be conducted relatively straightforwardly if they focus on the interventions that resulted from a design process. While good enough for most policy purposes, that approach doesn’t address the question of what proportion/aspect of good outcomes we might attribute specifically to the application of human-centered design. Given the complexity of the design process I don’t know the answer here, and while the letter raises the question, the authors’ five recommendations don’t seem to address it either. One could follow the approach taken in evaluating the impact of applying behavior change theories in the design of communication interventions - meta-analyses of many studies with and without HCD. Such an approach would likely surface a further round of questions, unaddressed in this letter, about the extent to which the impacts of design work can/should be attributed to the design practitioner’s degree of expertise, or whether a project was structured to enable the design practice to thrive, versus something integral to human-centered design as a discipline. Minor Comments: It might be helpful if you clarified your use of the terms “implementer” and “designer” throughout the paper. Are you referring to implementers of the design work (i.e., designers), or to other implementers who were operationalizing the interventions that the designers designed, or both? The reference below to ‘design energy’ is intriguing; I think readers would appreciate slightly more explanation of what this refers to and in what sense the term is dealt with in the cited reference. “During intervention development the process evaluation team faced challenges as the fast-paced, highly iterative HCD process meant that the decision making process, the ‘design energy’, often went undocumented 4.” In the statement below, I wasn’t able to follow how starting with a blank slate may have blocked the the design team from drawing on public health evidence. Design researchers and ethnographers typically strive to approach observational fieldwork in a manner that minimizes the influence of preconceived notions, but this does not prevent them from drawing on relevant theories or evidence during the process of analysis. Was disregarding public health evidence explicitly described as typical of the design process (a point I would disagree with in general, but that may well have been true to your experience), or might it have just been another researcher to practitioner disconnect? “Designers add value to public health as they take a different approach when addressing an issue, however, in the case of A360, a preference to work from a blank slate free of the constraints of public health evidence may have impacted on an ability to respond to the PE findings.” Typos “However, the iterative and flexible nature of HCD potentially poses some unique chal- lenges for evaluation 1,3,7.[ remove period] and examples of evaluations of HCD public health interventions have so far been limited 7,8.” “In the ‘inspiration phase’ the designers immerse themselves with girls and  her their influencers to understand the issues.” “The OE study design was finalised before the interventions themselves so there remained a risk that the evaluation was not optimal.” I think this refers to Outcome Evaluation study design; it would be helpful if you could clarify on first use of the acronym. “distinct imple- mentation phases as it was rolled  out across Nigeria” “because A360 was creating a new trans-disciplin ary approach” “more frequent cost data draws were required during the scale-up period to capture how those change s might affect the “production process” and associated costs.”
  7 in total

1.  Evaluation of quality improvement programmes.

Authors:  J Øvretveit; D Gustafson
Journal:  Qual Saf Health Care       Date:  2002-09

Review 2.  Evaluation design for large-scale HIV prevention programmes: the case of Avahan, the India AIDS initiative.

Authors:  Padma Chandrasekaran; Gina Dallabetta; Virginia Loo; Stephen Mills; Tobi Saidel; Rajatashuvra Adhikary; Michel Alary; Catherine M Lowndes; Marie-Claude Boily; James Moore
Journal:  AIDS       Date:  2008-12       Impact factor: 4.177

Review 3.  Human-centred design in global health: A scoping review of applications and contexts.

Authors:  Alessandra N Bazzano; Jane Martin; Elaine Hicks; Maille Faughnan; Laura Murphy
Journal:  PLoS One       Date:  2017-11-01       Impact factor: 3.240

4.  A framework for learning about improvement: embedded implementation and evaluation design to optimize learning.

Authors:  Danika Barry; Leighann E Kimble; Bejoy Nambiar; Gareth Parry; Ashish Jha; Vijay Kumar Chattu; M Rashad Massoud; Don Goldmann
Journal:  Int J Qual Health Care       Date:  2018-04-20       Impact factor: 2.038

5.  Evaluating the impact of an intervention to increase uptake of modern contraceptives among adolescent girls (15-19 years) in Nigeria, Ethiopia and Tanzania: the Adolescents 360 quasi-experimental study protocol.

Authors:  Christina Joanne Atchison; Emma Mulhern; Saidi Kapiga; Mussa Kelvin Nsanya; Emily E Crawford; Mohammed Mussa; Christian Bottomley; James R Hargreaves; Aoife Margaret Doyle
Journal:  BMJ Open       Date:  2018-05-31       Impact factor: 2.692

6.  Effects of the EQUIP quasi-experimental study testing a collaborative quality improvement approach for maternal and newborn health care in Tanzania and Uganda.

Authors:  P Waiswa; F Manzi; G Mbaruku; A K Rowe; M Marx; G Tomson; T Marchant; B A Willey; J Schellenberg; S Peterson; C Hanson
Journal:  Implement Sci       Date:  2017-07-18       Impact factor: 7.327

7.  Evaluating the impact of the DREAMS partnership to reduce HIV incidence among adolescent girls and young women in four settings: a study protocol.

Authors:  Isolde Birdthistle; Susan B Schaffnit; Daniel Kwaro; Maryam Shahmanesh; Abdhalah Ziraba; Caroline W Kabiru; Penelope Phillips-Howard; Natsayi Chimbindi; Kenneth Ondeng'e; Annabelle Gourlay; Frances M Cowan; James R Hargreaves; Bernadette Hensen; Tarisai Chiyaka; Judith R Glynn; Sian Floyd
Journal:  BMC Public Health       Date:  2018-07-25       Impact factor: 3.295

  7 in total
  4 in total

1.  Human-centred design of a new microneedle-based hormonal contraceptive delivery system.

Authors:  Benedetta Gualeni; Louise Hughes; Isabelle Stauber; Louise Ackers; Angela Gorman; Dorothy Gashuga; Nettie Dzabala; Frider Chimimba; Ibrahim Chikowe; Sion A Coulman; James C Birchall
Journal:  Gates Open Res       Date:  2021-07-30

2.  Cost-effectiveness analysis of human-centred design for global health interventions: a quantitative framework.

Authors:  Chen Liu; Jae Hyoung Lee; Patricia Turimumahoro; J Lucian Davis; David Dowdy; Amanda J Gupta; Austin Tucker; Chris Larkin; Achilles Katamba
Journal:  BMJ Glob Health       Date:  2022-03

3.  Innovative demand creation strategies to increase voluntary medical male circumcision uptake: a pragmatic randomised controlled trial in Zimbabwe.

Authors:  Webster Mavhu; Melissa Neuman; Karin Hatzold; Stephen Buzuzi; Galven Maringwa; Sungai T Chabata; Collin Mangenah; Noah Taruberekera; Ngonidzashe Madidi; Malvern Munjoma; Getrude Ncube; Sinokuthemba Xaba; Owen Mugurungi; Cheryl C Johnson; Elizabeth L Corbett; Helen A Weiss; Katherine Fielding; Frances M Cowan
Journal:  BMJ Glob Health       Date:  2021-07

4.  Elevating Social and Behavior Change as an Essential Component of Family Planning Programs.

Authors:  Joanna Skinner; Hope Hempstone; Laura Raney; Christine Galavotti; Benedict Light; Michelle Weinberger; Lynn Van Lith
Journal:  Stud Fam Plann       Date:  2021-07-15
  4 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.