Literature DB >> 30137381

Process evaluation of the effects of patient safety auditing in hospital care (part 2).

Mirelle Hanskamp-Sebregts1, Marieke Zegers2,3, Wilma Boeijen4, Hub Wollersheim3, Petra J van Gurp5, Gert P Westert3.   

Abstract

OBJECTIVE: To identify factors that explain the observed effects of internal auditing on improving patient safety. DESIGN SETTING AND PARTICIPANTS: A process evaluation study within eight departments of a university medical centre in the Netherlands. INTERVENTION(S): Internal auditing and feedback for improving patient safety in hospital care. MAIN OUTCOME MEASURE(S): Experiences with patient safety auditing, percentage implemented improvement actions tailored to the audit results and perceived factors that hindered or facilitated the implementation of improvement actions.
RESULTS: The respondents had positive audit experiences, with the exception of the amount of preparatory work by departments. Fifteen months after the audit visit, 21% of the intended improvement actions based on the audit results were completely implemented. Factors that hindered implementation were short implementation time: 9 months (range 5-11 months) instead of the 15 months' planned implementation time; time-consuming and labour-intensive implementation of improvement actions; and limited organizational support for quality improvement (e.g. insufficient staff capacity and time, no available quality improvement data and information and communication technological (ICT) support).
CONCLUSIONS: A well-constructed analysis and feedback of patient safety problems is insufficient to reduce the occurrence of poor patient safety outcomes. Without focus and support in the implementation of audit-based improvement actions, quality improvement by patient safety auditing will remain limited.
© The Author(s) 2018. Published by Oxford University Press in association with the International Society for Quality in Health Care.

Entities:  

Keywords:  audit; hospitals; patient safety; process evaluation study

Mesh:

Year:  2019        PMID: 30137381      PMCID: PMC6819993          DOI: 10.1093/intqhc/mzy173

Source DB:  PubMed          Journal:  Int J Qual Health Care        ISSN: 1353-4505            Impact factor:   2.038


Introduction

Healthcare providers are faced with the challenge of improving patient safety by detecting and preventing the occurrence of adverse events (AEs) [1, 2]. There is a growing interest in effective and sustainable interventions for reducing AEs [3, 4]. Audit and feedback is a widely used intervention for improving quality of care in numerous settings; however, the effectiveness is variable [5-7]. After a serious event on hospital level, we implemented an internal audit system to continuously improve patient safety in hospital care [8]. Central to this audit system is the use of a peer-to-peer evaluation approach to engage healthcare providers in the plan-do-check-act (PDCA) quality-improvement cycle [8]. Audits can be seen as the ‘check’ stage of the PDCA quality improvement cycle: auditors check whether quality standards have been established (‘plan’) and applied in practice (‘do’). Based on the audit results improvement actions are implemented (‘act’) by healthcare providers and management to improve patient safety [8]. Despite this approach, we observed in a previous effect evaluation study no statistically significant effect on the primary outcome, AEs and on team climate and patient safety culture [9]. Only patients’ experiences regarding their safety, observed medication safety and information security increased [9]. We incorporated a process evaluation in our effect evaluation study to understand the effects observed [10]. In this process evaluation, we evaluated whether the internal audit, including the implementation of improvement actions and revisit 15 months after the audit visit, had been executed as planned and whether the audit results triggered departments to improve patient safety. If the internal audit in its ultimate form differed considerably from the original plan, this could lead to ‘implementation error’, meaning incorrect implementation or insufficient execution of the intervention (in this case internal auditing) can lead to incorrect conclusions about the effectiveness of the intervention [10]. The effectiveness of patient safety auditing also depends on the context in which patient safety auditing is performed [7]. Therefore, we identified internal and external factors that might influence the effectiveness of patient safety auditing. The main research question of our process evaluation was: ‘Which combination of factors explains the degree of patient safety improvement by internal auditing in hospital care?’ Based on the results of our process evaluation, we provide recommendations on how organizations can support staff to improve the audit process. This process evaluation is the second part of a mixed-method evaluation study of patient safety auditing. The results of the effect evaluation are described in part 1 of this study [9].

Methods

Study design and setting

We used the process evaluation framework described by Hulscher et al. [10] to describe the actual performance of the patient safety auditing activities and to gain insight into the experiences with patient safety auditing of those who participated in the audit (auditees) and the factors that hinder or facilitate effective patient safety auditing in detail. We formulated four research questions: (1) Were the internal audits (including revisit) executed as planned? (2) What are the auditees’ experiences with internal audits? (3) Were the improvement actions implemented as planned and what are the results regarding patient safety? and (4) What are the barriers and facilitators for the implementation of the improvement actions? The study was performed between 2011 and 2014 on eight clinical departments of a 953-bed university hospital [8]. We used the Consolidated criteria for reporting qualitative research (COREQ) [11] guidelines and the Standards for QUality Improvement Reporting Excellence (SQUIRE) [12] for reporting our study.

Context of patient safety auditing

Patient safety auditing in a Dutch hospital was introduced after a serious event in 2006 that occurred due to inadequate clinical governance, lack of leadership, poor collaboration between hospital department medical staff, a blame culture and the lack of use of patient outcomes for quality improvement [13, 14]. Based on these findings, the internal audit system was optimized from 2007 to 2009 (Fig. 1). In this audit system, quality control is central and patient outcomes are the basis for quality improvements instead of controlling the presence of preconditions for good patient care as in previous audits [13]. Team climate measurements and appraisal and assessments by internal and external collaborating partners became a standard part of the internal audits. Healthcare providers were trained as auditors, and an independent board for Quality Assurance and Safety was installed. Since 2011, hospital departments, which deliver patient care, undergo an upgraded internal audit according to a fixed audit scheme of 21 months (Appendix 1) [8, 9]. The department heads write an improvement plan tailored to the audit results by using a standard improvement plan (Appendix 2). This standard should facilitate the implementation of the intended improvement actions and monitoring of progress by the department heads.
Figure 1

Timeline of development of patient safety auditing in hospital care.

Timeline of development of patient safety auditing in hospital care.

Measurements and data collection

We used a mixed-method approach for this process evaluation: (1) survey, (2) semi-structured interviews and (3) quantitative and qualitative content analysis of audit documents (Table 1).
Table 1

Research questions, measurements and data source

FocusResearch questionsMethodsTimingData source
Execution of internal audit1) Were the internal audits (including revisit) executed as planned?Quantitative content analysis of documents>1 month after revisit

Audit schedules

Audit visit programmes

Audit interview frameworks

Audit reports

Revisit reports

2) What are the experiences of auditees with internal audits?

Survey

Interviews

<2 weeks after the internal audit

1 month after revisit

Web-based questionnaire

Transcripts

Implementation of improvement actions

3) Were the improvement actions implemented as planned?

What are the results regarding patient safety?

Quantitative and qualitative content analysis of documents

Interviews

>1 month after revisit

1 month after revisit

Improvement plans

Revisit reports

Access database with classification of audit results and scores of improvement actions

Transcripts

4) What are the barriers and facilitators for the implementation of improvement actions?

Qualitative content analysis of documents

Interviews

>1 month after revisit

1 month after revisit

Revisit reports

Transcripts

Research questions, measurements and data source Audit schedules Audit visit programmes Audit interview frameworks Audit reports Revisit reports Survey Interviews <2 weeks after the internal audit 1 month after revisit Web-based questionnaire Transcripts 3) Were the improvement actions implemented as planned? What are the results regarding patient safety? Quantitative and qualitative content analysis of documents Interviews >1 month after revisit 1 month after revisit Improvement plans Revisit reports Access database with classification of audit results and scores of improvement actions Transcripts Qualitative content analysis of documents Interviews >1 month after revisit 1 month after revisit Revisit reports Transcripts (1) Survey We developed a web-based questionnaire to evaluate the auditees’ experiences (Appendix 3). The questions covered the elements of the audit process [8, 9] and its perceived strengths and limitations. The face validity of the questionnaire was examined by expert review (n = 10) and one plenary expert discussion meeting, and was tested in two departments. This led to refining of the formulations and the removal of three items from the questionnaire. The auditees received the questionnaire online within 2 weeks after the internal audit. Non-respondents received one reminder to reduce non-response bias.

(2) Interviews To cover the topics experiences in-depth with (strengths and limitations of) patient safety auditing, implementation of improvement actions and the barriers and facilitators of the implementation, an experienced interviewer conducted 24 semi-structured face-to-face interviews (60 min each) from March 2013 to October 2014 using an interview topic guide (Appendix 4). The interviews took place at the workplace within 1 month after the revisit. Purposive sampling was used to select a varied group of healthcare professionals to ensure diversity regarding involvement in the internal audits and job type [15]. The interviewees were informed about the study and its aim by e-mail and provided verbal informed consent at the beginning of the interview. (3) Quantitative and qualitative content analysis of audit documents The audit results were classified by an independent fixed team of four analysts into four quality domains [16]: management structure, preconditions, care processes and outcomes. The degree of implementation of the audit-based improvement actions was independently assessed by the two auditors who revisited the department. They used a 4-point rating scale [16]: 1 = not implemented, 2 = partly implemented, 3 = nearly implemented, 4 = completely implemented (5 = not assessed), and tabulated the final scores in the revisit reports. Disagreements on the classification and scores were discussed until consensus was reached.

Data analysis

The questionnaire data was analysed descriptively using the statistical software IBM SPSS V.22. The percentages of positive and negative experiences were calculated for each question by combining two positively phrased ratings, i.e. ‘agreed and strongly agreed’, and two negatively phrased ratings, i.e. ‘disagree and strongly disagree’, respectively. A positive response >80% and a negative response >10% were considered a strength and limitation, respectively, of patient safety auditing. This was a data driven decision meaningful for practice. The barriers and facilitators from the interviews (n = 24) and the revisit reports (n = 8) were qualitatively analysed by thematic analysis [15, 17] and directed content analysis [18], respectively. The interviews were audio-taped and transcribed verbatim according to a standardized format. Two researchers (M.H.-S. and E.P.) analysed the transcripts (n = 24) independently using the coding framework based on the integrated checklist of determinants of practice (TICD) of Flottrop [19] (see Appendix 5 for the coding tree). The interview data were discussed, replaced or recoded until consensus was reached. One researcher (M.H.-S.) systematically read the revisit reports, coded the barriers and facilitators using the interview coding framework, recoded in consultation with a second researcher (P.Jv.G.) and organized the final codes into a hierarchical structure if possible. M.H.-S tabulated all perceived barriers and facilitators and described the prominent sub-categories with illustrative quotes. M.H.-S., M.Z., P.Jv.G., H.W. and G.P.W. studied and discussed the prominent sub-categories in two team meetings. Atlas.ti.7 was used for coding, annotating and interpreting the results in the primary data material of the interviews and revisit reports. IBM SPSS V.22 was used to calculate the median time (in months) between audit visit and audit preparation, audit report handover, completion of the implementation plan, actual implementation of improvement actions and revisit by the audit team, respectively, and the classification of the audit results and the implementation rating scores. The 4-point implementation rating scale was therefore split into two scores: ‘completely implemented’ (4 points) and ‘not implemented’ (1–3 points). Improvement actions that were not assessed (n = 2) were excluded from the analysis.

Results

Execution of the internal audit

The internal audits (n = 8) were performed by 34 auditors in total: 27 physicians, three nurses and four allied healthcare workers. In total, 402 auditees were interviewed at the audit visit: 37.6% were physicians, 27.6% were nurses, 13.7% were managers, 8.5% were healthcare researchers, 5.5% were secretaries, 4.5% were allied healthcare workers and 2.7% were outpatient clinic employees. The median times were 5 months (range, 4–6 months) for audit preparation, 3 months (range, 2–3 months) for audit report handover to the department head, 3 months (range, 2–7.5 months) for completing the improvement plan, 9 months (range, 5–11 months) for starting of the implementation of the improvement plan and 15 months (range, 15–18 months) for performing the revisit. A pool of 10 auditors consisting of seven physicians, two nurses and one allied healthcare worker revisited the eight departments. They interviewed a total of 88 auditees, who were purposive selected on their involvement in the implementation of the audit-based improvement actions: 38.6% were physicians, 17.0% were nurses, 33.0% were managers, 1.1% were secretaries, 4.5% were allied healthcare workers, 1.1% were outpatient clinic employees and 4.5% were others.

Auditees’ experiences with internal audits

Of the 402 invited auditees, 213 (53.0%) filled in the questionnaire. Respondents were mainly physicians (32.4%), nurses (23.5%) and managers (23.9%). The other respondents were allied healthcare workers (2.8%), administrative staff (5.2%), researchers (3.3%), quality officers (6.6%) and unknown (2.3%). All heads of the departments involved (n = 16) responded. The distribution of job types of the respondents (e.g. physicians, nurses) corresponded to the distribution in the total group of auditees. Overall, 81.7% of the respondents had positive experiences with the internal audit; 84.1% indicated that the purpose of the audit was clear to them, 80.3% said that the internal audit was well-tailored and 80.7% considered the performance of the internal auditors professional (Appendix 3). Of the department heads, 86.4% indicated that the internal audit added value for improving quality of care. However, 36.4% felt that the amount of preparatory work required of the department was disproportional to the benefits, and 13.6% disagreed that the audit preparation was efficient. An overview of the results for each item is shown in Appendix 3.

Implementation of the intended improvement actions

In total, the eight departments formulated 81 improvement actions varying from 6 to 13 actions per department. Fifteen months after the internal audit, 21% (n = 17) of the 81 actions were implemented; 79% (n = 64) were partly implemented or not implemented (Fig. 2). The completed improvement actions were mainly related to medical devices and sterile medical aids management, documenting care processes in protocols and the use of existing indicators for monitoring the quality of care processes. The partly implemented or not implemented actions were related to the reorganization of multidisciplinary care processes (e.g. physician visit structure, outdoor clinic), improving inter-departmental collaboration, implementation of the national patient safety programme for Dutch hospitals [20], medical and nursing staff recruitment and evaluation of the departmental quality policy.
Figure 2

The number of improvement actions and percentage degree of implementation per department and overall 15 months after the audit visit.

The number of improvement actions and percentage degree of implementation per department and overall 15 months after the audit visit.

Perceived facilitators and barriers for implementing quality improvement actions

In total, 24 auditees were interviewed (interviewees), i.e. three per department (n = 8): six medical heads of departments, four department managers, five staff physicians, five head nurses, two senior nurses and two quality officers. Fifty-eight percent was male. The average years of experience in the current function were 6.5 and ranged 1–25 years. The interviewees mentioned 58 prominent barriers and facilitators that influenced the implementation of the audit-based improvement actions (Table 2). Most factors were related to internal auditing (n = 22), implementation processes (n = 14) and professional (n = 9). Eight factors were organizational-related and, five social-related. Prominent auditing-related factors were quality of audit results, efficiency of the audit process and competence of internal auditors. A prominent implementation-related factor was the feasibility of improvement actions and a prominent professional-related factor was the attitude of healthcare providers for quality improvement. The two prominent social-related factors were leadership of the department heads and improvement culture; organizational structure for quality improvement was a prominent organizational-related factor. Generally, interviewees considered the audit results objective, relevant and recognizable due to the competence of the internal auditors, which facilitated the translation of the audit results into improvement actions. The interviewees wondered if the same audit results could be achieved with a less extensive audit system to reduce their amount of preparatory work. They also noted that a number of the intended patient safety improvement actions were difficult to implement due to the feasibility of the actions, the quality of both the patient safety improvement and implementation plan and the organizational structure for quality improvement (e.g. sufficient time, capacity, management support, information and communication technological [ICT] support). Despite these barriers, the interviewees also mentioned that the patient safety interventions implemented were due to the positive attitude and intrinsic motivation of the healthcare providers and the urgency of the department heads to improve that patient safety problem. According to the interviewees, a learning culture facilitated continuing patient safety improvement even when the improvement action proved inadequate (Table 2).
Table 2

Perceived prominent facilitators and barriers for the implementation of improvement actions by the interview respondents

CategorySub-categoryFactorsFBIllustrative quotes
1. Internal auditing (including revisit)Quality of audit results

– Recognizable

Department manager: The recognizability of the audit results is definitely necessary for the support of your employees to initiate improvement actions.

Head nurse: How can you do 13 interviews in one day, how can you extrapolate the audit results while there are 225 people working in our department?

Staff physician: If I look at the content and quality of the audit results, they reflected well on the actual quality and safety problems. Looking at the audit report as a whole, the main bottlenecks were extracted by the audit team.

– Objective

– (Not) relevant

– (In)sufficient advice on how to improve

– Information bias (due to auditee selection and number of auditees, interview questions and low response rate on surveys)

Efficiency Process

– Time-consuming

Head of the department: We had nine surveys and visits in one and a half years, which is too much! Yes, I can hardly look into you as an organization, but I think you need to work more efficiently; I think that should be a bit effective, you should handle it.

Staff physician: Such a system rigging is fine, but you also need to be able to downsize at the moment that you find a very big step forward in quality improvement.

– Labour-intensive (lot of paperwork)

– Clear communication

– Use of quantitative measurement instruments (such as TCI)

– Use of expert auditors (e.g. pharmacist for medication safety)

– No coordination with site visits

– Not tailor made auditing to the quality level of the department

– No interviews with patients

– IA frequency

– Timing IA and revisit

Revisit

– Objective information of improvement plan progress

Senior nurse: The timing of the revisit is questionable given the fact that it easily takes half a year before your improvement plan is approved and you can proceed.

Department manager: You have to wonder if you have to do all kinds of intermediate things such as a revisit, because that gives us all the extra tasks at the expense of our available change capacity. Because we have to prepare the revisit, we have to collect documents again, instead of being able to work on the improvement actions.

– Information on PSI action effectiveness

– At the expense of change capacity

Auditors

– Work in hospital practice

Department manager: I think the audit has just been performed very thoroughly, with many specialized audits and quality topics. And I am just amazed, sometimes, that something comes up, that the audit team has discovered.

– Knowledge of healthcare processes

– Communication and social skills for safe interview climate

– Continuity within the fixed audit team

2. Implementation processFeasibility of improvement actions

– Time-consuming

Department manager: The strategic reorientation of our quality policy often takes a lot of time, because you have to live through with each other. That has not changed in the hearts and minds within 2 years.

Head nurse: Just to keep the improvement action small and practical, and the nurses perceived what the improvement gives them, or what benefits the improvement provides for the patient.

From revisit report: Not all intended results and steps to be taken for this purpose are SMART-formulated. This makes it difficult for the department to assess whether the improvement actions have led to the desired result.

– Labour-intensive

– Proposed /developed by healthcare professionals

– Involvement of stakeholders

– Evidence-based

– Not formulated cleverly

– Requires much behavioural change or work process reorganization

– Depends on the willingness and capacity of collaborating partners

– Perceived as threatening by staff

Implementation plan

– (No) prioritization

From revisit report: The PDCA cycle is not complete yet. From bottom-up new initiatives arise (plan/do phase); there is no structural attention for implementation and quality assurance (check/act phase).

– (No) use of the PDCA cycle

– No implementation plan

– (No) Pragmatic implementation approach

– Delayed by the hospital-wide introduction of a new EPD

3. ProfessionalAttitudeOpinion that IA:

Head of the department: I think it will clearly show the current state of affairs. It makes the department aware of all issues which need improvement. External control is necessary and helpful to bring the department to a higher level of performance.

Head nurse: I think that there are effects, but I’m not sure if they can all be contributed to the audit. I think we are currently in an upward move due to many different factors of which the audit report is one, since we are triggered in many ways with respect to quality.

– stimulates communication, transparency and awareness about quality of care

– stimulates a sense of urgency, interest and commitment to improve

– breaks through an inadequate (hierarchical) collaborative culture

– reduces business blindness

– produces checklists and tends to bureaucracy

– is more a monitoring instrument for board of directors than a PSI instrument for healthcare providers

Motivation

– Feeling responsible for PSI

Head nurse: We immediately realized this improvement action by well-driven people in the right place, who want to do their work qualitatively well.

– Intrinsic motivation for PSI

– Ownership of PSI

4. SocialLeadership

– Department head is role model (commitment to PSI, demonstrating approachable quality culture, proactive policy)

Head nurse: I think of the strength of our current heads of the department, who are interested in quality improvement, with a positive attitude, no blaming culture, and you improve immediately.

– (No) Support for PSI and behaviour changes

– Poor communication about PSI

Culture

– (Un)safe culture during the audit (‘feels like an exam’)

Head nurse: We learn if we have not taken the right decisions. We say that afterwards, we could have chosen better for that approach and not for this approach. That is what you are learning and carries that with us for the next time.

– (No) learning culture regarding PSI (e.g. learn from incorrect decisions, support each other, open dialogue, zero tolerance policy)

5. Organizational (= hospital, department)Organizational structure

– Insufficient staff capacity, time, finance, training

From revisit report: Various quality initiatives are visible within the divisions of the department. These initiatives often have an ad hoc character, generally lacking consistency within the department and are not part of an overarching quality system.

Quality officer: In our department, with high production standards and timetables running alongside each other, it is not always easy to have time to implement improvement actions.

– Poor monitoring implementation progress

– No available quality and safety data

– Lack of ICT support

– Unclear communication structure

– Support of quality officers

– No sharing of best practices

– Established responsibility by the board of directors for monitoring PSI

F = facilitator; B = barrier; IA = internal audit; PSI = patient safety improvement; EPD = electronic patient record; TCI = Team Climate Inventory; SMART = specific, measurable, attainable, realistic, timely; QI = quality improvement.

Perceived prominent facilitators and barriers for the implementation of improvement actions by the interview respondents – Recognizable Department manager: The recognizability of the audit results is definitely necessary for the support of your employees to initiate improvement actions. Head nurse: How can you do 13 interviews in one day, how can you extrapolate the audit results while there are 225 people working in our department? Staff physician: If I look at the content and quality of the audit results, they reflected well on the actual quality and safety problems. Looking at the audit report as a whole, the main bottlenecks were extracted by the audit team. – Objective – (Not) relevant – (In)sufficient advice on how to improve – Information bias (due to auditee selection and number of auditees, interview questions and low response rate on surveys) – Time-consuming Head of the department: We had nine surveys and visits in one and a half years, which is too much! Yes, I can hardly look into you as an organization, but I think you need to work more efficiently; I think that should be a bit effective, you should handle it. Staff physician: Such a system rigging is fine, but you also need to be able to downsize at the moment that you find a very big step forward in quality improvement. – Labour-intensive (lot of paperwork) – Clear communication – Use of quantitative measurement instruments (such as TCI) – Use of expert auditors (e.g. pharmacist for medication safety) – No coordination with site visits – Not tailor made auditing to the quality level of the department – No interviews with patients – IA frequency – Timing IA and revisit – Objective information of improvement plan progress Senior nurse: The timing of the revisit is questionable given the fact that it easily takes half a year before your improvement plan is approved and you can proceed. Department manager: You have to wonder if you have to do all kinds of intermediate things such as a revisit, because that gives us all the extra tasks at the expense of our available change capacity. Because we have to prepare the revisit, we have to collect documents again, instead of being able to work on the improvement actions. – Information on PSI action effectiveness – At the expense of change capacity – Work in hospital practice – Knowledge of healthcare processes – Communication and social skills for safe interview climate – Continuity within the fixed audit team – Time-consuming Department manager: The strategic reorientation of our quality policy often takes a lot of time, because you have to live through with each other. That has not changed in the hearts and minds within 2 years. Head nurse: Just to keep the improvement action small and practical, and the nurses perceived what the improvement gives them, or what benefits the improvement provides for the patient. From revisit report: Not all intended results and steps to be taken for this purpose are SMART-formulated. This makes it difficult for the department to assess whether the improvement actions have led to the desired result. – Labour-intensive – Proposed /developed by healthcare professionals – Involvement of stakeholders – Evidence-based – Not formulated cleverly – Requires much behavioural change or work process reorganization – Depends on the willingness and capacity of collaborating partners – Perceived as threatening by staff – (No) prioritization – (No) use of the PDCA cycle – No implementation plan – (No) Pragmatic implementation approach – Delayed by the hospital-wide introduction of a new EPD Head of the department: I think it will clearly show the current state of affairs. It makes the department aware of all issues which need improvement. External control is necessary and helpful to bring the department to a higher level of performance. Head nurse: I think that there are effects, but I’m not sure if they can all be contributed to the audit. I think we are currently in an upward move due to many different factors of which the audit report is one, since we are triggered in many ways with respect to quality. – stimulates communication, transparency and awareness about quality of care – stimulates a sense of urgency, interest and commitment to improve – breaks through an inadequate (hierarchical) collaborative culture – reduces business blindness – produces checklists and tends to bureaucracy – is more a monitoring instrument for board of directors than a PSI instrument for healthcare providers – Feeling responsible for PSI – Intrinsic motivation for PSI – Ownership of PSI – Department head is role model (commitment to PSI, demonstrating approachable quality culture, proactive policy) – (No) Support for PSI and behaviour changes – Poor communication about PSI – (Un)safe culture during the audit (‘feels like an exam’) – (No) learning culture regarding PSI (e.g. learn from incorrect decisions, support each other, open dialogue, zero tolerance policy) – Insufficient staff capacity, time, finance, training From revisit report: Various quality initiatives are visible within the divisions of the department. These initiatives often have an ad hoc character, generally lacking consistency within the department and are not part of an overarching quality system. Quality officer: In our department, with high production standards and timetables running alongside each other, it is not always easy to have time to implement improvement actions. – Poor monitoring implementation progress – No available quality and safety data – Lack of ICT support – Unclear communication structure – Support of quality officers – No sharing of best practices – Established responsibility by the board of directors for monitoring PSI F = facilitator; B = barrier; IA = internal audit; PSI = patient safety improvement; EPD = electronic patient record; TCI = Team Climate Inventory; SMART = specific, measurable, attainable, realistic, timely; QI = quality improvement.

Discussion

We assumed that patient safety in hospital care would be significantly improved 15 months after the audit visit, as after the serious event in our hospital, there was a great sense of urgency among the healthcare providers to prevent such events, and they believed that patient safety auditing could be an effective intervention for assuring and improving the safety of hospital care [13]. Auditing is an appropriate method for identifying patient safety problems and their underlying causes and would aid the formulation of improvement actions tailored to the detected safety problems [7, 21]. However, in a previous study, we measured the effects of internal auditing in hospital care and observed no convincing effect [9]. In this process evaluation, we studied why the observed effects are small. In general, healthcare providers’ experiences with the internal audit process were positive, with the exception of the amount of departmental preparatory work, which was perceived as inefficient and disproportional in relation to the audit results. The audit results provided sufficient information for writing an improvement plan. The department heads and the healthcare providers implemented 21% of the intended improvement actions. Important factors that hindered the implementation of improvement actions were mainly in the follow-up phase: (1) the implementation time was <9 months instead of the planned 15 months; (2) it is unrealistic to expect to change healthcare providers’ behaviour and reorganize multidisciplinary care processes within 15 months; (3) there was limited implementation of interventions focused on specific patient safety problems, such as team training [22]; (4) an unfeasible improvement plan (e.g. too time-consuming and labour-intensive improvement actions) and (5) inadequate infrastructure for quality improvement (e.g. insufficient staff capacity and time, lack of quality evaluation data and ICT support). Previous comparative studies, such as that by De Vos et al. [23] have also demonstrated difficulties in translating feedback information into effective actions; the same study also reported that the relatively short follow-up of 14 months resulted in no effects on quality improvement. Benning et al. [3] learned that great resources are necessary for effecting departmental change, and time is needed to entrenched patient safety improvements to have an impact on patient safety at department level. In most studies, data were typically gathered within 12 months of the intervention, while it is recognized that culture is a slow-changing phenomena [24]. Improved patient safety outcomes are observed when measurements are taken at least 2 years after the intervention [24]. The time required for healthcare providers to implement the quality improvement changes is underestimated [25]. Factors as staff capacity, time, data infrastructure, management and ICT support are well-known facilitators for successful quality improvement [24, 26]. A methodological strength of this study is the use of a mixed-method approach, which included a survey, interviews and document analysis for providing in-depth information on internal audit execution and experiences and the implementation of improvement actions [15]. This approach is recommended for capturing the effects of diffuse and complex innovations [27]. With the interviews more in-depth information is gathered regarding patient safety auditing; information that is difficult to capture with questionnaires [15]. Several limitations of this study have to be taken into account. The response to the questionnaire was just over 50%. This appears to be a low response rate; however, response rates of <20% among healthcare professionals are not uncommon [28, 29]. We also did not interview the internal auditors. Possibly, the auditors’ perspectives would provide new information on the audit process and the influencing factors. This process evaluation resulted in a broad overview of factors that influence the effectiveness of patient safety auditing, which hospitals can use to support their staff to improve the audit process (Box 1). PDCA = Plan, Do, Check, Act; ICT = Information and Communication Technology. Conclusively, a well-constructed analysis and feedback of patient safety problems is insufficient to reduce the occurrence of poor patient safety outcomes. Our process evaluation showed that a lack of focus and support in the implementation of audit-based improvement actions prevents the completion of the PDCA improvement cycle. Quality improvement by patient safety auditing will remain limited. Click here for additional data file. Click here for additional data file. Click here for additional data file. Click here for additional data file. Click here for additional data file.
FacilitatorsRecommendations
Reliable audit results Provide accurate feedback of the identified patient safety problems by objective, relevant, recognizable and reliable audit results.
Competent auditors Detect patient safety problems by competent internal auditors who are also working at the hospital and who therefore know the care processes very well.
Efficient audit process Prevent time consuming and labour intensive audit process by: tailor-made auditing on the quality level of the audited department/healthcare process, clear communication and coordination with other quality visits and measures.
Feasible improvement actions Formulate and prioritize improvement actions with healthcare providers and stakeholders for feasibility and commitment.
Effective implementation Choose tailor made, pragmatic implementation approach (e.g. PDCA cycle). Use quality officers’ support.
Leadership, attitude and learning culture Combine leadership of the department heads, healthcare providers’ positive attitude and intrinsic motivation, and a learning culture for initiating and perpetuating improvements to address patient safety problems successfully.
Organization-wide advice and support Give organization-wide advice and support for those responsible for improvement actions to achieve successfully implementation, among other things: calculating and facilitating the staff, time and finance required, the availability of quality improvement data, ICT support and sharing good patient safety improvement practices.

PDCA = Plan, Do, Check, Act; ICT = Information and Communication Technology.

  22 in total

1.  Three approaches to qualitative content analysis.

Authors:  Hsiu-Fang Hsieh; Sarah E Shannon
Journal:  Qual Health Res       Date:  2005-11

2.  Surveying clinicians by web: current issues in design and administration.

Authors:  Jennifer Dykema; Nathan R Jones; Tara Piché; John Stevenson
Journal:  Eval Health Prof       Date:  2013-09       Impact factor: 2.651

3.  An epistemology of patient safety research: a framework for study design and interpretation. Part 4. One size does not fit all.

Authors:  C Brown; T Hofer; A Johal; R Thomson; J Nicholl; B D Franklin; R J Lilford
Journal:  Qual Saf Health Care       Date:  2008-06

Review 4.  Safety and risk management interventions in hospitals: a systematic review of the literature.

Authors:  Michel Dückers; Marjan Faber; Juliette Cruijsberg; Richard Grol; Lisette Schoonhoven; Michel Wensing
Journal:  Med Care Res Rev       Date:  2009-09-16       Impact factor: 3.929

5.  Trends in adverse events over time: why are we not improving?

Authors:  Kaveh G Shojania; Eric J Thomas
Journal:  BMJ Qual Saf       Date:  2013-04       Impact factor: 7.035

6.  Case studies of patient safety research classics to build research capacity in low- and middle-income countries.

Authors:  Anne Andermann; Albert W Wu; Angela Lashoher; Peter Norton; Narendra Kumar Arora; David W Bates; Itziar Larizgoitia
Journal:  Jt Comm J Qual Patient Saf       Date:  2013-12

7.  Improving quality of care through improved audit and feedback.

Authors:  Sylvia J Hysong; Cayla R Teal; Myrna J Khan; Paul Haidet
Journal:  Implement Sci       Date:  2012-05-18       Impact factor: 7.327

8.  How effective are patient safety initiatives? A retrospective patient record review study of changes to patient safety over time.

Authors:  Rebecca Baines; Maaike Langelaan; Martine de Bruijne; Peter Spreeuwenberg; Cordula Wagner
Journal:  BMJ Qual Saf       Date:  2015-07-06       Impact factor: 7.035

9.  Effects of patient safety auditing in hospital care: results of a mixed-method evaluation (part 1).

Authors:  Mirelle Hanskamp-Sebregts; Marieke Zegers; Gert P Westert; Wilma Boeijen; Steven Teerenstra; Petra J van Gurp; Hub Wollersheim
Journal:  Int J Qual Health Care       Date:  2019-08-01       Impact factor: 2.038

10.  Response audit of an Internet survey of health care providers and administrators: implications for determination of response rates.

Authors:  Mark J Dobrow; Margo C Orchard; Brian Golden; Eric Holowaty; Lawrence Paszat; Adalsteinn D Brown; Terrence Sullivan
Journal:  J Med Internet Res       Date:  2008-10-16       Impact factor: 5.428

View more
  3 in total

1.  Effects of patient safety auditing in hospital care: results of a mixed-method evaluation (part 1).

Authors:  Mirelle Hanskamp-Sebregts; Marieke Zegers; Gert P Westert; Wilma Boeijen; Steven Teerenstra; Petra J van Gurp; Hub Wollersheim
Journal:  Int J Qual Health Care       Date:  2019-08-01       Impact factor: 2.038

2.  Understanding how and why audits work in improving the quality of hospital care: A systematic realist review.

Authors:  Lisanne Hut-Mossel; Kees Ahaus; Gera Welker; Rijk Gans
Journal:  PLoS One       Date:  2021-03-31       Impact factor: 3.240

3.  Reducing Injection-Related Safety Events in Retina Clinics.

Authors:  Denise A John; Beth Hansemann; Philip Lieu; Jennifer Weizer
Journal:  Clin Ophthalmol       Date:  2022-04-22
  3 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.