Literature DB >> 26600614

A national system for monitoring the performance of hospitals in Ethiopia.

Zahirah McNatt1, Erika Linnander1, Abraham Endeshaw2, Dawit Tatek1, David Conteh3, Elizabeth H Bradley1.   

Abstract

Many countries struggle to develop and implement strategies to monitor hospitals nationally. The challenge is particularly acute in low-income countries where resources for measurement and reporting are scarce. We examined the experience of developing and implementing a national system for monitoring the performance of 130 government hospitals in Ethiopia. Using participatory observation, we found that the monitoring system resulted in more consistent hospital reporting of performance data to regional health bureaus and the federal government, increased transparency about hospital performance and the development of multiple quality-improvement projects. The development and implementation of the system, which required technical and political investment and support, would not have been possible without strong hospital-level management capacity. Thorough assessment of the health sector's readiness to change and desire to prioritize hospital quality can be helpful in the early stages of design and implementation. This assessment may include interviews with key informants, collection of data about health facilities and human resources and discussion with academic partners. Aligning partners and donors with the government's vision for quality improvement can enhance acceptability and political support. Such alignment can enable resources to be focused strategically towards one national effort - rather than be diluted across dozens of potentially competing projects. Initial stages benefit from having modest goals and the flexibility for continuous modification and improvement, through active engagement with all stakeholders.

Entities:  

Year:  2015        PMID: 26600614      PMCID: PMC4645435          DOI: 10.2471/BLT.14.151399

Source DB:  PubMed          Journal:  Bull World Health Organ        ISSN: 0042-9686            Impact factor:   9.408


Introduction

Improvement in the quality of hospital care is a fundamental aspect of health system strengthening– that is directly linked to the service delivery dimension of the World Health Organization (WHO) building blocks of a health system. While the monitoring of hospital performance is a key ingredient to such improvement,,, many countries struggle to develop and implement feasible strategies to monitor hospitals nationally. The challenge is particularly acute in low-income countries where resources for measurement and reporting are scarce. In the field of global health, research on performance monitoring often focuses broadly on health systems– rather than on hospitals. The literature on the development and implementation of systems for monitoring hospital performance is largely dominated by case studies and reports from high-income countries with national health systems – e.g. Canada and Denmark, the United Kingdom of Great Britain and Northern Ireland and other countries in western Europe.– Although there has also been some relevant research in the United States of America, it has tended to focus on a narrow set of quality measures in specific populations., The WHO performance assessment tool for quality improvement in hospitals is a donor-led, externally designed measurement project rather than a country-led, internally developed initiative., This tool has been applied in only one middle-income country (South Africa)., Most attempts to monitor hospital performance in low-income settings have involved small numbers of facilities and narrowly defined clinical measures of performance.– When creating their accreditation systems for hospitals, both Liberia and Zambia monitored hospital performance for just a year, to collect baseline data., We could find no peer-reviewed studies done in low-income countries that described the development and sustained implementation of a national system for monitoring hospital performance, based upon a comprehensive set of key performance indicators. We therefore sought to describe the creation and implementation of such a national system in a low-income country. We considered Ethiopia to be a good setting in which to conduct our case study because of recent hospital reform in the country. The reform led to the creation of: (i) the role of hospital chief executive officer – qualified through a master’s degree programme in hospital and health-care administration;, (ii) private wings in hospitals that allowed revenue generation and (iii) hospital governing boards., The many new government hospitals that were built during the ongoing reform process led to improved hospital access in both rural and urban settings. We describe the development of key performance indicators, the process of monitoring hospital performance relative to these indicators and the trend in performance since 2010, which marked the implementation of Ethiopia’s national system of hospital monitoring. Findings from this case study may be helpful to other low-income countries seeking to elevate the quality of facility-based health care through performance monitoring and accountability.

Key performance indicators

Development

We developed performance indicators that were relevant for hospitals and consistent throughout the country. The first indicator developed was the most fundamental – adherence to national guidelines on hospital management. In 2009, Ethiopia partnered with the Clinton Health Access Initiative and the Yale Global Health Leadership Institute to develop national guidelines for the management of hospitals: the Ethiopian Hospital Reform Implementation Guidelines., These guidelines included 124 hospital management standards, each of which was a statement – e.g. “the hospital conducts a physical inventory of all pharmaceuticals in the store and each dispensing unit at a minimum of once per year.” Hospitals were asked to report quarterly whether each standard was met. The success of the rollout of Ethiopia’s first attempt to monitor hospital performance, in 2010, was probably the result of simplicity and focus on hospital management. The guidelines leveraged the Ethiopian Federal Ministry of Health’s investment in the training of hospital chief executive officers via the master’s of hospital and health-care administration degree programme., The guidelines, the associated scoring sheet, the promotion of adherence to the guidelines and the building of management capacity were made integral parts of the two-year programme. The students in the programme were selected by regional health bureaus. At the time of writing, more than 90% of those who successfully completed the degree programme remain employed in the Ethiopian health-care sector (D Tatek, unpublished observations, 2014). Given the reality that, in 2009–2010, government hospitals were understaffed, financially limited and often did not have 24-hour access to basic resources such as water and electricity, the ministry of health agreed that, before launching reporting on other aspects of hospital performance, such as efficiency, cost, clinical outcomes and patient experience – government hospitals should be accountable for meeting a set of minimum management standards. By 2011, 40% of government hospitals were reporting data on their adherence to operational standards – to both the ministry of health and the appropriate regional health bureau. Improvements were already apparent in the establishment of hospital quality committees, drug formularies, pharmacy inventory control systems and a host of other quality-improvement efforts. Staff from regional health bureaus and development partners visited hospitals to corroborate the reported information and to provide coaching on the operational standards. The environment was poised for the introduction of a more robust monitoring system based upon multiple key performance indicators. In 2011, the ministry of health negotiated a standardized set of performance indicators, in partnership with the regional health bureaus and with technical assistance as before, from the Clinton Health Access Initiative and the Yale Global Health Leadership Institute. The process of selecting such indicators for the nation’s hospitals was rigorous and included reviewing other country experiences, development of thematic areas and frequent reviews with federal, regional and external stakeholders. Given the need for these indicators to be reliable – and collection of data on them to be feasible – the ministry of health held sessions with the regional health bureaus to determine which indicators were most important to the bureaus and what human resource training and infrastructure development were needed to enable collection of the corresponding data. Six months of research and negotiation resulted in the establishment of 36 national indicators for the assessment of hospital performance. These indicators covered 11 aspects of hospital operations: hospital management, outpatient, emergency, inpatient, maternity, referral and pharmacy services, productivity, human resources, finance and patient experience (Table 1). Together, the indicators encompassed measures of operational functioning, clinical quality, financial activity and patient experience. The ministry of health worked with partners to limit the number of indicators that could potentially create perverse incentives (e.g. mortality rates) and to explain, to hospital and ministry of health staff, the potential unintended effects of indicator measurement.
Table 1

Hospital key performance indicators, Ethiopia, 2010

Category, indicator codeIndicator
Hospital management
KPI 1Proportion of EHRIG operational standards met
Outpatient services
KPI 2Outpatient attendees
KPI 3Outpatient attendees seen by private-wing service
KPI 4Outpatient waiting time to treatment
KPI 5Outpatients not seen on same day
Emergency services
KPI 6ED attendees
KPI 7ED patients triaged within 5 minutes of arrival at ED
KPI 8ED attendances with stay longer than 24 hours
KPI 9ED mortality
Inpatient services
KPI 10Inpatient admissions
KPI 11Inpatient admissions to private wing
KPI 12Inpatient mortality
KPI 13Delay for elective surgical admission
KPI 14Bed occupancy
KPI 15Mean length of stay
KPI 16Incidence of pressure ulcer
KPI 17Percentage of surgical sites infected
KPI 18Completeness of inpatient medical records
Maternity services
KPI 19Deliveries – i.e. live births and stillbirths – attended
KPI 20Births by surgical, instrumental or assisted vaginal delivery
KPI 21Institutional maternal mortality
KPI 22Institutional neonatal deaths within 24 hours of birth
Referral services
KPI 23Referrals made
KPI 24Rate of referrals
KPI 25Emergency referrals, as a proportion of all referrals made
Pharmacy services
KPI 26Mean stock-out duration of hospital-specific tracer drugs
Productivity
KPI 27Patient-day equivalents per physician
KPI 28Patient-day equivalents per nurse or midwife
KPI 29Major surgeries per surgeon
KPI 30Major surgeries conducted in private wing
Human resources
KPI 31Attrition rate among physicians
KPI 32Staff experience, as a staff satisfaction rating
Finance
KPI 33Cost per patient-day equivalent
KPI 34Raised revenue, as a proportion of total operating revenue
KPI 35Revenue utilization – i.e. the proportion of budget used
Patient experience
KPI 36Patient experience, as a patient satisfaction rating

ED: emergency department; EHRIG: Ethiopian hospital reform implementation guidelines; KPI: key performance indicator.

ED: emergency department; EHRIG: Ethiopian hospital reform implementation guidelines; KPI: key performance indicator. The performance indicators included process measures that were directly related to patient outcomes. For example, one indicator – the average time during which stocks of basic drugs were unavailable – highlighted how often inpatients and outpatients were unable to purchase medications and therefore had to remain untreated or source medications from private pharmacies. Another indicator – the percentage of patients triaged within five minutes of arrival in the emergency department – was particularly important to all stakeholders as it directly responded to community concerns about mortality and morbidity resulting from delayed treatment. The success of the development of the indicators was largely due to the simplicity, flexibility and transparency of the process. The number of indicators was kept small and the focus was on measures that could be calculated reasonably easily by hospital staff. The ministry of health required commitment to reporting the 36 national indicators but allowed regional health bureaus to create additional indicators, which stimulated regional ownership. A National Hospital Performance Monitoring and Improvement Manual, which outlined each indicator thoroughly and specified precise definitions and data sources, was disseminated through a series of national workshops funded by the United States Centers for Disease Control and Prevention (CDC) and the ministry of health.

Implementation and monitoring

The implementation of the monitoring system included substantial investments in both human resources and information technology. In terms of human resources, new roles were developed at the hospital level and in the regional health bureaus and ministry of health. Each hospital had several individuals – so-called data owners – who were each dedicated to collecting data on the performance indicators that were relevant to their department. For example, a midwife could be the data owner for neonatal mortality. In addition, each hospital had an indicator collator who worked closely with each data owner and was responsible for the collation of data on all the indicators. Instead of hiring new personnel to undertake these tasks, most hospitals modified the job descriptions of current employees and provided additional short-term, on-site training. Data on the indicators were initially collected on paper forms and then compiled and submitted as spreadsheet computer files. Health and development partners provided technical support for designing data entry and reporting applications. At bureau and ministry level, the curative and rehabilitation teams and the medical services directorate were dedicated to the performance indicators and hospital operations. These teams were responsible for training hospital data owners and indicator collators, troubleshooting problems with data collection and reporting and synthesizing the hospital-level data into a national database for comparing hospital performance within and across regions (Table 2). The ministry of health used the summary databases during discussions of trends in hospital performance, at quarterly joint sessions of the regional and federal health leadership.
Table 2

National summary data on nine key performance indicators for 121 government hospitals, Ethiopia, 2013

IndicatorCodeQuarter of year
FirstSecondThirdFourthAll
Hospital management
Proportion of EHRIG operational standards met, %KPI 170.674.775.377.574.5
Outpatient services
Outpatient attendees, No.KPI 2586 337618 442648 910648 125625 453
Outpatient attendees seen by private-wing services, %KPI 37.06.65.96.06.4
Outpatient waiting time to treatment, minutesKPI 437.140.344.941.441.0
Outpatients not seen on same day, %KPI 50.50.50.20.20.3
Emergency services
ED attendees, No.KPI 6198 078203 496212 982213 570828 126
ED patients triaged within 5 minutes of arrival at ED, %KPI 793.676.394.9NR93.0
ED attendees with stay longer than 24 hours, %KPI 82.42.12.32.02.2
ED mortality, %KPI 90.30.20.20.20.2

ED: emergency department; EHRIG: Ethiopian hospital reform implementation guidelines; KPI: key performance indicator; NR, not reported.

ED: emergency department; EHRIG: Ethiopian hospital reform implementation guidelines; KPI: key performance indicator; NR, not reported. The approach used to establish the Ethiopian system for monitoring hospital performance was designed to fit the Ethiopian context. Many hospital employees were initially unfamiliar with the methods used for reliable and valid data collection and few had adequate experience with computer software. As many of the computers available in hospitals functioned poorly, the system was designed to use relatively simple software programmes. The main challenges that arose during implementation were errors in data collection and calculation at hospital level and the fear of reprisal for poor performance. For instance, some hospital employees were unsure which denominators or patient populations they should be using. Some hospitals repeatedly failed to report data on particular indicators and some were afraid to report data that highlighted poor performance – especially poor clinical indicators. In the first year of the system, rates of surgical site infection and neonatal mortality were often found to be underreported. Hospitals that appeared to be struggling in reporting reasonably accurate data on the key performance indicators were offered additional on-site training and one-on-one coaching. In their hospital-wide meetings, hospital chief executive officers were encouraged to cultivate an accountable but non-punitive environment. Regional health bureaus reinforced the importance of the data-collection efforts and, by improving the timeliness of their feedback on the summary data to hospitals, helped prompt more immediate exploration and correction of data errors. The costs of the monitoring system were originally covered by a grant from the United States CDC. Implementing partners were unable to quantify such costs accurately or to separate them from those of other programmatic activities. In addition to the efforts of the nongovernmental organization and university partners, the ministry of health and regional health bureaus made both financial and in-kind contributions to the establishment and maintenance of the monitoring system. Future efforts would benefit from a more explicit analysis of costs.

Impact of monitoring

As the national monitoring system was fully implemented, rates of hospital reporting of performance indicators increased. This trend indicated changes in hospital functioning and encouraged improvements in performance. In September 2011, 40% of the 114 government hospitals then in Ethiopia were regularly reporting their performance in terms of all 36 key indicators; by September 2013 this had risen to 78%, and by September 2014, 84%.,, The collection and analysis of performance data reportedly motivated hospital-based performance-improvement projects – e.g. the introduction of hourly nurse rounding, distinct staff uniforms, continuous pharmaceutical stock reporting and outpatient appointment systems. Between 2012 and 2013, mean adherence to the operational standards increased from 68.2% to 74.5% while the mean number of deliveries attended each month increased from 12 187 to 16 001., The national monitoring system also improved evidence-based decision-making at both hospital and government level. Comparative performance results were presented at quarterly meetings with hospitals and regional health bureau staff and this allowed for the open review of performance results, feedback and problem solving. Managers at all levels of the health sector aimed to sustain the enthusiasm for performance monitoring. This required continuous investment in the use of data for tangible improvements, media attention and team and organizational rewards and may, in the long term, include institutional accreditation by national bodies. One example was Ethiopia’s recent integration of the 36 performance indicators into a national quality campaign: the Ethiopian Hospital Alliance for Quality. In 2012, the alliance financially rewarded the 15 hospitals that, according to the relevant performance indicators, offered the most positive patient experiences – with about 55 000 United States dollars each. In 2014, the ministry of health began the alliance’s second cycle and prioritized institutional maternal mortality.

General observations

Our five-year experience of the development and implementation of a national system for monitoring hospital performance led to several key observations. First, technical investment was critical throughout the process. Many hours of research, writing and development of guidelines were needed to develop a core set of performance indicators that were evidence-based, comprehensive but not overwhelming, and precisely described to allow their consistent calculation and reporting. Ethiopia’s ministry of health led the initiative between 2009 and 2014 and now has full operational responsibility. The ministry has a department exclusively charged with overseeing the country’s management of hospital performance – with support from key champions, including the Minister of Health. Second, while technical support was critical in the development of the indicators and related documentation, political support was paramount to successful implementation. The ministry of health set a consistent direction and held partners accountable to deliver on its vision for quality improvement. The regional health bureaus also demonstrated strong leadership in advocating for additional performance indicators that fit their regional needs and ensured government and hospital ownership of the monitoring system. Although disagreement emerged, senior government officials continued discussions until a negotiated consensus brought a stable solution that all parties could then support. The process of identifying the best key indicators conferred momentum and helped sustain the monitoring efforts. Although such characteristics may be key to making lasting changes, they can be challenging to embed in any large-scale national efforts. Lastly, both the technical and political inputs were accomplished because of the ability to leverage strong management capacity – which was built at hospital level and supported by the executive master’s degree programme. The importance of management capacity has been highlighted by many studies.,,,– The chief-executive-officer model – i.e. the establishment of a dedicated, qualified person in each hospital who is trained in hospital management and supported by a hospital governing board – was pivotal in the successful implementation of the system for monitoring performance. Without the management capacity provided by this model, the ideas and strategies written in technical and political arenas would not have been translated into practice at the hospital level. Once adequate management capacity has been built, performance management and reporting become achievable – and even desirable for facility-level staff who wish to assess their own progress. The combination of leverage from existing hospital management capacity, technical inputs and political support provided the conditions and tools needed to enable success in this country-led effort to elevate the performance of hospitals in Ethiopia.

Conclusion

Ethiopia’s implementation of a national system for monitoring hospital performance serves as an example of a low-income country that aims to improve health service delivery via the creation of a culture of accountability. A limitation of our study is that we lacked outcome data and thus were unable to evaluate the impact of the monitoring system on population health. Such an evaluation would require a long and comprehensive follow-up of patients. Despite this limitation, our observations may be helpful to other low-income countries that are seeking to improve the quality of their hospital care. We offer several recommendations. First, a thorough assessment of the health sector’s readiness to change and desire to prioritize hospital quality can be helpful in the early stages of design and implementation. Such an assessment may include interviews with key informants, collection of data about health facilities and human resources and investigation of local university capacity to offer academic programmes in health-care management. Second, partner and donor alignment with the government’s national vision for quality improvement can enhance acceptability and political support. This alignment can enable resources to be focused strategically towards one national effort – rather than be diluted across dozens of potentially competing projects. Finally, early phases of implementation benefit from having modest early goals and the facility for continuous modification and improvement to the performance monitoring system, through active engagement with all stakeholders.
  40 in total

1.  Quality improvement in the developing world.

Authors:  Helen L Smits; Sheila Leatherman; Donald M Berwick
Journal:  Int J Qual Health Care       Date:  2002-12       Impact factor: 2.038

Review 2.  Lessons from developing nations on improving health care.

Authors:  Donald M Berwick
Journal:  BMJ       Date:  2004-05-08

3.  A performance assessment framework for hospitals: the WHO regional office for Europe PATH project.

Authors:  J Veillard; F Champagne; N Klazinga; V Kazandjian; O A Arah; A-L Guisset
Journal:  Int J Qual Health Care       Date:  2005-09-09       Impact factor: 2.038

4.  U.S. health system performance: a national scorecard.

Authors:  Cathy Schoen; Karen Davis; Sabrina K H How; Stephen C Schoenbaum
Journal:  Health Aff (Millwood)       Date:  2006-09-20       Impact factor: 6.301

5.  Hospital management in the context of health sector reform: a planning model in Ethiopia.

Authors:  Kari Hartwig; Josh Pashman; Emily Cherlin; Martha Dale; Marguerite Callaway; Cindy Czaplinski; W Edward Wood; Yigeremu Abebe; Timothy Dentry; Elizabeth H Bradley
Journal:  Int J Health Plann Manage       Date:  2008 Jul-Sep

Review 6.  Supporting the delivery of cost-effective interventions in primary health-care systems in low-income and middle-income countries: an overview of systematic reviews.

Authors:  Simon Lewin; John N Lavis; Andrew D Oxman; Gabriel Bastías; Mickey Chopra; Agustín Ciapponi; Signe Flottorp; Sebastian García Martí; Tomas Pantoja; Gabriel Rada; Nathan Souza; Shaun Treweek; Charles S Wiysonge; Andy Haines
Journal:  Lancet       Date:  2008-09-13       Impact factor: 79.321

7.  Improving quality in resource poor settings: observational study from rural Rwanda.

Authors:  Meera Kotagal; Patrick Lee; Caste Habiyakare; Raymond Dusabe; Philibert Kanama; Henry M Epino; Michael L Rich; Paul E Farmer
Journal:  BMJ       Date:  2009-10-30

8.  The Philippines Field Management Training Program (FMTP): strengthening management capacity in a decentralized public health system.

Authors:  Nemia L Sucaldito; Enrique A Tayag; Maria Concepcion R Roces; Michael D Malison; Brian D Robie; Elizabeth H Howze
Journal:  Int J Public Health       Date:  2014-09-20       Impact factor: 3.380

9.  Application of balanced scorecard in the evaluation of a complex health system intervention: 12 months post intervention findings from the BHOMA intervention: a cluster randomised trial in Zambia.

Authors:  Wilbroad Mutale; Jeffrey Stringer; Namwinga Chintu; Roma Chilengi; Margaret Tembo Mwanamwenge; Nkatya Kasese; Dina Balabanova; Neil Spicer; James Lewis; Helen Ayles
Journal:  PLoS One       Date:  2014-04-21       Impact factor: 3.240

10.  Advancing the application of systems thinking in health: realist evaluation of the Leadership Development Programme for district manager decision-making in Ghana.

Authors:  Aku Kwamie; Han van Dijk; Irene Akua Agyepong
Journal:  Health Res Policy Syst       Date:  2014-06-16
View more
  13 in total

1.  Maternal and newborn outcomes at a tertiary care hospital in Lusaka, Zambia, 2008-2012.

Authors:  Bellington Vwalika; Marie C D Stoner; Mulindi Mwanahamuntu; K Cherry Liu; Eugene Kaunda; Getrude G Tshuma; Somwe W Somwe; Yusuf Ahmed; Elizabeth M Stringer; Jeffrey S A Stringer; Benjamin H Chi
Journal:  Int J Gynaecol Obstet       Date:  2016-11-24       Impact factor: 3.561

2.  Factors Influencing the Efficiency of Public Hospitals in Saudi Arabia: A Qualitative Study Exploring Stakeholders' Perspectives and Suggestions for Improvement.

Authors:  Ahmed D Alatawi; Louis W Niessen; Minakshi Bhardwaj; Yussif Alhassan; Jahangir A M Khan
Journal:  Front Public Health       Date:  2022-06-16

3.  Race to the Top: evaluation of a novel performance-based financing initiative to promote healthcare delivery in rural Rwanda.

Authors:  Evrard Nahimana; Ryan McBain; Anatole Manzi; Hari Iyer; Alice Uwingabiye; Neil Gupta; Gerald Muzungu; Peter Drobac; Lisa R Hirschhorn
Journal:  Glob Health Action       Date:  2016-11-28       Impact factor: 2.640

4.  Professionalizing Healthcare Management: A Descriptive Case Study.

Authors:  Erika L Linnander; Jeannie M Mantopoulos; Nikole Allen; Ingrid M Nembhard; Elizabeth H Bradley
Journal:  Int J Health Policy Manag       Date:  2017-10-01

5.  Identifying Key Performance Indicators for Holistic Hospital Management with a Modified DEMATEL Approach.

Authors:  Sheng-Li Si; Xiao-Yue You; Hu-Chen Liu; Jia Huang
Journal:  Int J Environ Res Public Health       Date:  2017-08-19       Impact factor: 3.390

6.  Maternal and neonatal services in Ethiopia: measuring and improving quality.

Authors:  Maureen E Canavan; Marie A Brault; Dawit Tatek; Daniel Burssa; Ayele Teshome; Erika Linnander; Elizabeth H Bradley
Journal:  Bull World Health Organ       Date:  2017-04-25       Impact factor: 9.408

7.  Is bed turnover rate a good metric for hospital scale efficiency? A measure of resource utilization rate for hospitals in Southeast Nigeria.

Authors:  Henry E Aloh; Obinna E Onwujekwe; Obianuju G Aloh; Chijioke J Nweke
Journal:  Cost Eff Resour Alloc       Date:  2020-07-01

8.  Use of a national collaborative to improve hospital quality in a low-income setting.

Authors:  Erika Linnander; Zahirah McNatt; Heather Sipsma; Dawit Tatek; Yigeremu Abebe; Abraham Endeshaw; Elizabeth H Bradley
Journal:  Int Health       Date:  2016-01-20       Impact factor: 2.473

9.  Using the World Health Organization health system building blocks through survey of healthcare professionals to determine the performance of public healthcare facilities.

Authors:  Tsegahun Manyazewal
Journal:  Arch Public Health       Date:  2017-08-31

Review 10.  The development of hospital accreditation in low- and middle-income countries: a literature review.

Authors:  Wesam Mansour; Alan Boyd; Kieran Walshe
Journal:  Health Policy Plan       Date:  2020-07-01       Impact factor: 3.344

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.