Literature DB >> 32107249

Revitalising audit and feedback to improve patient care.

Robbie Foy1, Mirek Skrypak2, Sarah Alderson1, Noah Michael Ivers3, Bren McInerney4, Jill Stoddart5, Jane Ingham5, Danny Keenan5.   

Abstract

Entities:  

Mesh:

Year:  2020        PMID: 32107249      PMCID: PMC7190377          DOI: 10.1136/bmj.m213

Source DB:  PubMed          Journal:  BMJ        ISSN: 0959-8138


× No keyword cloud information.
Clinical audit and feedback entail reviewing clinical performance against explicit standards and delivering feedback to enable data driven improvement The impact of audit could be maximised by applying implementation science, considering the needs of clinicians and patients, and emphasising action over measurement Embedding research on how to improve audit and feedback in large scale programmes can further enhance their effectiveness and efficiency Healthcare systems face challenges in tackling variations in patient care and outcomes.1 2 Audit and feedback aim to improve patient care by reviewing clinical performance against explicit standards and directing action towards areas not meeting those standards.3 It is a widely used foundational component of quality improvement, included in around 60 national clinical audit programmes in the United Kingdom. Ironically, there is currently a gap between what audit and feedback can achieve and what they actually deliver, whether led locally or nationally. Several national audits have been successful in driving improvement and reducing variations in care, such as for stroke and lung cancer, but progress is also slower than hoped for in other aspects of care (table 1).4 5 Audit and feedback have a chequered past.6 Clinicians might feel threatened rather than supported by top-down feedback and rightly question whether rewards outweigh efforts invested in poorly designed audit. Healthcare organisations have limited resources to support and act on audit and feedback. Dysfunctional clinical and managerial relationships undermine effective responses to feedback, particularly when it is not clearly part of an integrated approach to quality assurance and improvement. Unsurprisingly, the full potential of audit and feedback has not been realised.
Table 1

Examples of national clinical audit programmes and randomised trials evaluating audit and feedback

ObjectiveMethodsIllustrative findings
National clinical audit programmes
To measure and improve the structure, processes, and outcomes of stroke careThe National Clinical Audit for Stroke operates a prospective, continuous audit of the processes and outcomes of NHS funded stroke care and rehabilitation in acute and post-acute settings in England and Wales. It also reviews care at six months and beyond to assess how longer term needs are metStroke unit performance in key aspects of care improved over five years; eg, the proportion of patients assessed by a stroke specialist consultant physician within 24 hours rose from 74% to 83%, whereas the proportion of applicable patients screened for nutrition and seen by a dietitian by discharge rose from 66% to 81%.4 However, significant gaps in provision remain; eg, fewer than one in three patients receive a six month review
To measure and improve care and outcomes for lung cancerFor the National Lung Cancer Audit, secondary and tertiary care NHS hospitals in England and Wales submit data via the National Cancer Registration and Analysis Service as part of the Cancer Outcomes and Services Dataset. The data are linked to Hospital Episode Statistics, the National Radiotherapy Dataset, the Systemic Anti-Cancer Dataset, pathology reports, and death certificate dataThe proportion of patients alive at least one year after diagnosis rose from 31% in 2010 to 37% in 2017.5 However, almost a third of patients still lack access to the benefits of specialist nursing support
Randomised trials of audit and feedback
To assess the effect of adding an action implementation toolbox to electronic audit and feedback targeting quality of pain management in intensive care units16 21 Dutch intensive care units were randomly assigned to receive usual electronic feedback only or to feedback with an implementation toolbox suggesting practical actions staff could take to improve pain managementOver six months, the proportion of patient shifts with adequate pain management increased by 14.8% compared with 4.8% in the feedback only group
To assess the effects of feedback including “social norm” persuasive messaging and patient focused information on antibiotic prescribing in higher prescribing general practices7 26 1581 English general practices whose prescribing rate for antibiotics was in the top 20% for their locality were randomly assigned to receive feedback including a letter from England’s chief medical officer highlighting the higher rate of antibiotic prescribing or to no communication. They were then randomly assigned to receive patient focused information promoting reduced use of antibiotics or to no communicationOver six months, the rate of antibiotic items dispensed per 1000 population was 127 in the feedback intervention group and 131 in the control group, representing an estimated 73 406 fewer antibiotic items dispensed. The patient focused intervention did not significantly affect prescribing
Examples of national clinical audit programmes and randomised trials evaluating audit and feedback Clinical, patient, and academic communities might need to have more sophisticated conversations about audit and feedback to achieve substantial, data driven, continuous improvement. They can also act now. There are ways to maximise returns from the considerable resources, including clinician time, invested in audit programmes. These include applying what is already known, paying attention to the whole audit cycle, getting the right message to the right recipients, making more out of less data, embedding research to improve impact, and harnessing public and patient involvement.

Apply what is already known

Audit and feedback generally work. A Cochrane review of 140 randomised trials found that they produced a median 4.3% absolute improvement (interquartile range 0.5% to 16%) in healthcare professionals’ compliance with desired practice, such as recommended investigations or prescribing.3 This is a modest effect, but cumulative incremental gains through repeated audit cycles can deliver transformative change. Audit and feedback also influence reach and population through scaled up national programmes, which other quality improvement approaches (such as financial incentives or educational outreach visits) might not achieve with similar resources; for example, social norm feedback (presenting information to show that individuals are outliers in their behaviour) from a high profile messenger can reduce antibiotic prescribing in primary care at low cost and at national scale (table 1).7 The interquartile range in the Cochrane review indicates that a quarter of audit and feedback interventions had a relatively large, positive effect of up to 16% on patient care, whereas a quarter had a negative or null effect. The effects of feedback can be amplified by ensuring that it is given by a supervisor or colleague, provided more than once, delivered in both verbal and written formats, and includes both explicit targets for change and action plans.3 A synthesis of expert interviews and systematic reviews identified 15 “state of the science,” theory informed suggestions for effective feedback (box 1).8 These are practical ways to maximise the impact and value of existing audit programmes. Can you recommend actions that are consistent with established goals and priorities? Can you recommend actions that can improve and are under the recipient’s control? Can you recommend specific actions? Can you provide multiple instances of feedback? Can you provide feedback as soon as possible and data frequency informed by the number of new patient cases? Can you provide individual rather than general data? Can you choose comparators that reinforce desired behaviour change? Can you closely link the visual display and summary message? Can you provide feedback in more than one way? Have you minimised extraneous cognitive load for feedback recipients? Have you addressed barriers to feedback use? Can you provide short, actionable messages followed by optional detail? Have you addressed credibility of the information? Can you prevent defensive reactions to feedback? Can you construct feedback through social interaction?

Pay attention to the whole cycle

The audit and feedback process comprises one or more cycles of establishing best practice criteria, measuring current practice, feeding back findings, implementing changes, and further monitoring. This chain is only as strong as its weakest link. Feedback effects can be weakened by information-intention gaps (feedback fails to convince recipients that change is necessary), intention-behaviour gaps (intentions are not translated into action), or behaviour-impact gaps (actions do not yield the desired effect on patient care).9 The success of national audit programmes depends on local arrangements that promote action as well as measurement.10 A synthesis of 65 qualitative evaluations proposed ways of designing audit programmes to better align with local capacity, identity, and culture and to promote greater changes in clinical behaviour.11 Healthcare organisations have finite capacity, so audit programmes should be designed so that they require less work, make best use of limited local resources, and clearly state why any investment is justified. Clinician beliefs about what constitutes best practice can influence how they respond to feedback, so audit programmes need to consider these while also challenging the status quo. All aspects of audit programmes should be designed with a focus on the desired changes in behaviour by recipients to achieve better outcomes; for example, feedback tackling unnecessary blood transfusions could include suggested alternative approaches to minimise blood loss during surgery.12 Because the purpose of an audit programme is not measurement alone but using data to inform quality improvement, we need to understand existing barriers to desired change and have a plan for how feedback helps to tackle those barriers. Without functioning local networks and systems, national audit programmes can become echo chambers, where good intentions and blame for limited progress reverberate. Audit and feedback will flounder if local quality improvement is based on repeated, unconnected, and inappropriately delegated projects conducted in isolation from mainstream pursuits and if any learning is dissipated in collective amnesia. Clinical and managerial leaders should ask questions about their organisational performance in response to feedback (box 2)13 and set clear goals, mobilise resources, and promote continuous improvement.14 Audit and feedback by themselves cannot solve ingrained deficiencies but can emphasise priorities for change, inform focused actions, and evaluate progress. Do we know how good we are? Do we know where we stand relative to the best? Do we know where and understand why variation exists in our organisation? Over time, where are the gaps in our practice that indicate a need for change? In our efforts to improve, what’s working?

Get the right message to the right recipients

Feedback comparing performance among different healthcare organisations and clinicians can leverage competitive instincts. This might not always work as intended. Nobody likes being told they are getting it wrong, repeatedly. Yet this is how clinicians and organisations often experience feedback suggesting suboptimal performance. Low baseline performance is associated with greater improvement after feedback3 but can elicit defensive reactions (“I don’t believe these data”), especially if feedback does not align with recipient perceptions (”My patients are different”). Such responses are not uncommon given that clinicians tend to overestimate their own performance.15 Continued negative feedback perceived as punitive can also be demotivating and risk creating burnout (“What else can I do?”). Giving feedback to professionals who take pride in their work requires careful thought. Consider, for example, providing feedback to high performers—will positive feedback lead to reduced effort or increase motivation? Should audit programmes switch attention to new topics where performance is poorer, at risk of inducing fatigue in higher performers? Given the law of diminishing returns, attempts to improve already high levels of performance might be less fruitful than switching attention to other priorities. Many clinical actions have a “ceiling” beyond which improvement is restricted because healthcare organisations or clinicians are functioning at or near their maximum capabilities. A range of approaches can help tailor feedback to recipients’ needs. First, feedback can include comparators that show like for like (such as similar types of organisations with similar case mixes) and set realistic goals for change relative to performance levels (such as lower but more achievable targets for poorer performers). Second, feedback can be delivered alongside a range of tangible action plans to support improvement; for example, an implementation toolbox improved pain management in intensive care units.16 17 Third, new audit criteria need to be convincing, based on robust evidence and with scope for patient and population benefit.

Make more out of less data

Healthcare organisations and clinicians need to juggle competing priorities and therefore struggle to act on all feedback from national and local audit programmes. A 2012 snapshot identified 107 National Institute for Health and Care Excellence clinical guidelines relevant to primary care, resulting in 2365 recommendations.18 Audit programmes can help to identify which recommendations have the greatest potential to benefit patients and populations. One of the highest costs associated with audit programmes is the time and effort involved in data collection, particularly the manual review of patient records. The burden of this data collection can be compounded by temptations to add in more variables for analyses that marginally improve precision.19 The resulting feedback might reinforce the credibility of data and enable recipients to explore associations in the data. Providing larger amounts of complex data, however, risks cognitive overload and distracting recipients from key messages. The diminishing returns of continuing efforts to perfect data come at the expense of focusing energy on improvement.19 The increasing availability of electronic patient record systems and routinely collected data on quality of care offer opportunities for large scale, efficient feedback programmes. Such approaches offer greater population coverage, which can overcome risks of biased sampling associated with manual review, such as the loss of records of patients with poorer outcomes. Routine data can also be collected and analysed in real time, thereby enabling faster, continuous feedback and countering objections voiced by clinicians (“These data are out of date”). Data quality is only as good as coding at the point of care. Validity checks and quality control of the data might compound the burden on clinical teams. Data linkage and extraction across different information requires compliance with data protection and information governance requirements. Even with all this in place, we must acknowledge Einstein’s advice that not everything that counts can be counted, and not everything that can be counted counts.

Embed research to improve impact

Poor research design, conduct, and dissemination contribute to “research waste.”20 Implementation science aims to translate research evidence into routine practice and policy but is also affected by research waste. A cumulative meta-analysis of the Cochrane review of audit and feedback indicated that the effect size stabilised in 2003 after 30 trials.21 By 2011, 47 more trials of audit and feedback versus control were published that did not substantially advance knowledge, many omitting feedback features likely to enhance effectiveness. This indicated a growing literature but “stagnant science.” Implementation laboratories offer a means of enhancing the impact of audit and feedback while also producing generalisable knowledge about how to optimise effectiveness.22 A “radical incrementalist” approach entails making serial, small changes, supported by tightly focused evaluations to cumulatively improve outcomes.23 It is already used in public policy and in business. Amazon and eBay randomly assign potential customers to see different presentations of their products online to understand what drives purchases. It is also applicable to healthcare24 and can help answer many questions about how best to organise and deliver feedback (such as, does feedback on performance indicating an organisation’s position against top performing peers stimulate more improvement than showing its position against average performance? What is the effect of shorter versus longer feedback reports? Does adding additional persuasive messages have any effect?). Embedding sequential head-to-head trials testing different feedback methods in an audit programme provides a robust empirical driver for change. Modifications identified as more effective than the current standard become the new standard; those that are not are discarded.

Harness public and patient involvement

Healthcare providers and researchers are still learning how to work meaningfully with patients and the public, and there are opportunities in audit programmes. This means moving beyond current models of involvement—typically advisory group roles to ensure accountability and contribute to strategy—towards active participation in feedback and service improvement. Patients and the public are often surprised by the extent of unwarranted variations in healthcare delivery, which is the core business of audit programmes.25 They express frustration at the difficulties in routinely measuring less technical aspects of care, such as consultation skills and patient centredness. Involving patients and the public, including seldom heard communities, early in the process of developing indicators is important. Audit programmes can be at the forefront of innovating and evaluating different approaches to involvement, asking questions such as, does incorporating the patient voice in feedback lead to greater improvement? Can feedback reports be better designed to improve understanding for both lay and professional board members of healthcare organisations? Patients and the public represent an underexplored and untapped force for change, which audit programmes can learn to harness.

Conclusion

Audit and feedback are widely used, sometimes abused, and often under-realised in healthcare. More imaginative design and responses are overdue; these require evidence informed conversations between clinicians, patients, and academic communities. It is time to fully leverage national audits to accelerate data guided improvement and reduce unwarranted variations in healthcare. The status quo is no longer ethical.
  21 in total

Review 1.  Avoidable waste in the production and reporting of research evidence.

Authors:  Iain Chalmers; Paul Glasziou
Journal:  Lancet       Date:  2009-06-12       Impact factor: 79.321

Review 2.  Healthcare outcomes and quality in the NHS: how do we compare and how might the NHS improve?

Authors:  Azeem Majeed; Dominique Allwood; Kim Foley; Andrew Bindman
Journal:  BMJ       Date:  2018-07-13

3.  Are postal questionnaire surveys of reported activity valid? An exploration using general practitioner management of hypertension in older people.

Authors:  M Eccles; G A Ford; S Duggan; N Steen
Journal:  Br J Gen Pract       Date:  1999-01       Impact factor: 5.386

4.  Creating a Learning Health System through Rapid-Cycle, Randomized Testing.

Authors:  Leora I Horwitz; Masha Kuznetsova; Simon A Jones
Journal:  N Engl J Med       Date:  2019-09-19       Impact factor: 91.245

5.  How does audit and feedback influence intentions of health professionals to improve practice? A laboratory experiment and field study in cardiac rehabilitation.

Authors:  Wouter T Gude; Mariëtte M van Engen-Verheul; Sabine N van der Veer; Nicolette F de Keizer; Niels Peek
Journal:  BMJ Qual Saf       Date:  2016-04-11       Impact factor: 7.035

6.  Provision of social norm feedback to high prescribers of antibiotics in general practice: a pragmatic national randomised controlled trial.

Authors:  Michael Hallsworth; Tim Chadborn; Anna Sallis; Michael Sanders; Daniel Berry; Felix Greaves; Lara Clements; Sally C Davies
Journal:  Lancet       Date:  2016-02-18       Impact factor: 79.321

7.  Clinical Performance Feedback Intervention Theory (CP-FIT): a new theory for designing, implementing, and evaluating feedback in health care based on a systematic review and meta-synthesis of qualitative research.

Authors:  Benjamin Brown; Wouter T Gude; Thomas Blakeman; Sabine N van der Veer; Noah Ivers; Jill J Francis; Fabiana Lorencatto; Justin Presseau; Niels Peek; Gavin Daker-White
Journal:  Implement Sci       Date:  2019-04-26       Impact factor: 7.327

8.  The Quality of Outpatient Care Delivered to Adults in the United States, 2002 to 2013.

Authors:  David M Levine; Jeffrey A Linder; Bruce E Landon
Journal:  JAMA Intern Med       Date:  2016-12-01       Impact factor: 21.873

9.  Developing 'high impact' guideline-based quality indicators for UK primary care: a multi-stage consensus process.

Authors:  Bruno Rushforth; Tim Stokes; Elizabeth Andrews; Thomas A Willis; Rosemary McEachan; Simon Faulkner; Robbie Foy
Journal:  BMC Fam Pract       Date:  2015-10-28       Impact factor: 2.497

Review 10.  Growing literature, stagnant science? Systematic review, meta-regression and cumulative analysis of audit and feedback interventions in health care.

Authors:  Noah M Ivers; Jeremy M Grimshaw; Gro Jamtvedt; Signe Flottorp; Mary Ann O'Brien; Simon D French; Jane Young; Jan Odgaard-Jensen
Journal:  J Gen Intern Med       Date:  2014-11       Impact factor: 5.128

View more
  19 in total

1.  A multi-faceted intervention to improve screening of erectile dysfunction for men with diabetes mellitus in public health clinics: A pilot study.

Authors:  Chai Li Tay; Chirk Jenn Ng
Journal:  Malays Fam Physician       Date:  2022-05-15

2.  Outliers from national audits: their analysis and use by the Care Quality Commission in quality assurance and regulation of healthcare services in England.

Authors:  Helen Grote; Keiko Toma; Laura Crosby; Catherine Robson; Clare Palmer; Claire Land; Jessica Ball; Edward Baker
Journal:  Clin Med (Lond)       Date:  2021-08-13       Impact factor: 5.410

3.  Enhancing national audit through addressing the quality improvement capabilities of feedback recipients: a multi-phase intervention development study.

Authors:  Michael Sykes; Elaine O'Halloran; Lucy Mahon; Jenny McSharry; Louise Allan; Richard Thomson; Tracy Finch; Niina Kolehmainen
Journal:  Pilot Feasibility Stud       Date:  2022-07-08

4.  'More than just numbers on a page?' A qualitative exploration of the use of data collection and feedback in youth mental health services.

Authors:  Craig Hamilton; Kate Filia; Sian Lloyd; Sophie Prober; Eilidh Duncan
Journal:  PLoS One       Date:  2022-07-20       Impact factor: 3.752

5.  Repeated analyses of national clinical audit reports demonstrate improvements in feedback methods.

Authors:  Tasneem Khan; Sarah Alderson; Jill J Francis; Fabiana Lorencatto; John Grant-Casey; Simon J Stanworth; Robbie Foy
Journal:  Implement Sci Commun       Date:  2020-11-25

6.  Opportunities to enhance ward audit: a multi-site qualitative study.

Authors:  Michael Sykes; Richard Thomson; Niina Kolehmainen; Louise Allan; Tracy Finch
Journal:  BMC Health Serv Res       Date:  2021-03-12       Impact factor: 2.655

7.  Embedded trials within national clinical audit programmes: A qualitative interview study of enablers and barriers.

Authors:  Sarah Alderson; Thomas A Willis; Su Wood; Fabiana Lorencatto; Jill Francis; Noah Ivers; Jeremy Grimshaw; Robbie Foy
Journal:  J Health Serv Res Policy       Date:  2021-12-09

8.  Analysis of a Web-Based Dashboard to Support the Use of National Audit Data in Quality Improvement: Realist Evaluation.

Authors:  Natasha Alvarado; Lynn McVey; Mai Elshehaly; Joanne Greenhalgh; Dawn Dowding; Roy Ruddle; Chris P Gale; Mamas Mamas; Patrick Doherty; Robert West; Richard Feltbower; Rebecca Randell
Journal:  J Med Internet Res       Date:  2021-11-23       Impact factor: 5.428

9.  Making the most of audit and feedback to improve diabetes care: a qualitative study of the perspectives of Australian Diabetes Centres.

Authors:  Matthew Quigley; Sophia Zoungas; Edward Zimbudzi; Natalie Wischer; Sofianos Andrikopoulos; Sally E Green
Journal:  BMC Health Serv Res       Date:  2022-02-24       Impact factor: 2.655

Review 10.  A short cultural history of the UK Renal Registry 1995-2020.

Authors:  Eric John Will
Journal:  BMC Nephrol       Date:  2020-08-12       Impact factor: 2.388

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.