Literature DB >> 35777872

Feedback of aggregate patient-reported outcomes (PROs) data to clinicians and hospital end users: findings from an Australian codesign workshop process.

Olivia Francis Ryan1, Shaun L Hancock1, Violet Marion1, Paulette Kelly2, Monique F Kilkenny1,3, Benjamin Clissold4,5, Penina Gunzburg6, Shae Cooke7, Lauren Guy8, Lauren Sanders9,10, Sibilah Breen1, Dominique A Cadilhac11,3.   

Abstract

OBJECTIVES: Patient-reported outcomes (PROs) are increasingly used to measure the patient's perspective of their outcomes following healthcare interventions. The aim of this study was to determine the preferred formats for reporting service-level PROs data to clinicians, researchers and managers to support greater utility of these data to improve healthcare and patient outcomes.
SETTING: Healthcare professionals receiving PRO data feedback at the health service level. PARTICIPANTS: An interdisciplinary Project Working Group comprised of clinicians participated in three workshops to codesign reporting templates of summarised PRO data (modified Rankin Scale, EuroQol Five Dimension Descriptive System, EuroQol Visual Analogue Scale and Hospital Anxiety and Depression Scale) using a modified Delphi process. An electronic survey was then distributed to short list the preferred templates among a broad sample of clinical end users. A final workshop was undertaken with the Project Working Group to review results and reach consensus on the final templates. PRIMARY AND SECONDARY OUTCOME MEASURES: The recommendation of preferred PRO summary data feedback templates and guiding principles for reporting aggregate PRO data to clinicians was the primary outcome. A secondary outcome was the identification of perceived barriers and enablers to the use of PRO data in hospitals. For each outcome measure, quantitative and qualitative data were summarised.
RESULTS: 31 Working Group members (19 stroke, 2 psychology, 1 pharmacy, 9 researchers) participated in the workshops, where 25/55 templates were shortlisted for wider assessment. The survey was completed by 114 end users. Strongest preferences were identified for bar charts (37/82 votes, 45%) and stacked bar charts (37/91 votes, 41%). At the final workshop, recommendations to enhance communication of PROs data for comparing health service performance were made including tailoring feedback to professional roles and use of case-mix adjustment to ensure fair comparisons.
CONCLUSIONS: Our research provides guidance on PROs reporting for optimising data interpretation and comparing hospital performance. © Author(s) (or their employer(s)) 2022. Re-use permitted under CC BY-NC. No commercial re-use. See rights and permissions. Published by BMJ.

Entities:  

Keywords:  epidemiology; health & safety; public health; quality in health care; stroke

Mesh:

Year:  2022        PMID: 35777872      PMCID: PMC9252210          DOI: 10.1136/bmjopen-2021-055999

Source DB:  PubMed          Journal:  BMJ Open        ISSN: 2044-6055            Impact factor:   3.006


We used iterative codesign methods with interdisciplinary clinical or policy stakeholders who were considered the end users of the evidence we were to generate from this study. A modified Delphi process was used to reach consensus on the PRO summary data feedback templates and to establish guidance for the preferred reporting formats. Strengths of our study include the engagement of clinical, academic, patient and government representatives who were members of the Working Group for the Delphi process and involvement of a wider group of stakeholders who completed an electronic survey. The use of different commonly used PRO measures was a strength, as we trialled both condition-specific and generic PRO measures in an effort to increase the generalisability of our results across different clinical populations. A limitation was our inability to include examples of the presentation of baseline and follow-up PRO data due to the sample data being derived from the stroke registry where PROs are only collected at a single time point after the acute event. We acknowledge that these data are important for a range of conditions such as chronic kidney disease or surgical interventions where progression may be important to monitor. Due to the use of the condition of stroke as a case study our findings may lack generalisability for other clinical conditions. However, the principles are largely transferable. The next steps are to pilot and refine the recommended templates.

Introduction

Patient-reported outcomes (PROs) are data collected directly from the patient about their own health, without interpretation by a clinician or any other person.1 The use of PROs in comparative-effectiveness research and clinical trials is well established.2 There is increased recognition that PROs have the potential to enhance clinical practice through directing the need for improvements in the safety and quality of healthcare.3–5 PRO data can be used at the individual-patient level as a means to inform clinicians on aspects of their health that are important to them6 or at the health service level to enable clinicians to compare hospital performance or against achievable benchmarks to identify quality improvement opportunities.7 8 The interpretation of aggregate, health service-level PROs data by clinicians and other hospital end users is the focus of our research. Known challenges related to the interpretation and use of PRO data include scoring and scaling differences between measures as well as inconsistencies in the methods used to report PRO data back to clinicians.9 To address these challenges, our preliminary research involved a scoping review to summarise the existing evidence related to preferred formats for PRO data feedback to clinicians.10 While a single preferred format or approach to feedback PRO data to healthcare professionals was not identified, we could summarise some general guidance on how to design feedback formats to enhance the interpretation of summarised PRO data. The aim of the current study was to build on our preliminary research and establish guidance for preferred formats for hospital comparisons of PRO data to support greater utility of these data to improve healthcare and patient outcomes. In terms of quality of care, PROs can have relevance to clinicians if these data are considered in addition to information about performance on standards of clinical care (eg, expected interventions to be provided) and other information such as mortality and readmissions. For example, multicomponent audit and feedback approaches whereby PROs data are used to complement clinical indicator data can have an additive effect and contribute to better care through improving adherence to clinical guidelines.11 Other authors12 have acknowledged that there is currently a lack of primary studies and randomised controlled trials related to the effectiveness of aggregate PRO data for benchmarking and quality improvement purposes and the associated impact on patient outcomes.13 Despite this, PRO feedback is an emerging field and our research constitutes an important development in this area. We also sought to gain preliminary insights into the perceived barriers and enablers to the use of PRO data in hospitals.

Methods

Study design

This was a codesigned project, whereby the methods included several interdisciplinary end users working together to produce the outcomes.14 We defined end users as hospital clinicians from a range of disciplines, academics who analyse and report PRO data and government representatives who use PRO data to inform quality improvement and policy decisions. We used a modified Delphi process15 to reach consensus on PRO summary data templates and establish guidance for preferred reporting formats. The Delphi method is a group communication process, which aims to achieve convergence of opinions on a specific real-world issue.15 The current study was designed in three stages: (1) a series of three workshops whereby members of a Project Working Group (herein referred to as Working Group) iteratively refined several PRO data feedback templates, (2) an anonymous electronic survey disseminated widely across Australia via clinical and academic networks to enable the short listing of the codesigned PRO summary data templates, (3) a final workshop with the Working Group members to achieve consensus on the recommended set of PRO summary data reporting format characteristics (figure 1).
Figure 1

Diagrammatic representation of project stages. EOI: expression of interest; WS: workshop

Diagrammatic representation of project stages. EOI: expression of interest; WS: workshop

Setting

This study was commissioned by the Victorian Agency for Health Information (VAHI) as part of a broader strategy to facilitate the comprehensive collection of PRO data in Victoria, Australia. To inform the state-wide strategy, VAHI commissioned consultancy work by Paxton Partners in 2018, whereby the authors identified that regular feedback of PRO data had the potential to add substantial value to improve clinical engagement with these data (Paxton Partners, PRO measures (PROMs): Literature Scan, personal communication, 2019). In response, VAHI appointed a project team comprising representatives from The Florey, Safer Care Victoria and Monash University to build on this work (see online supplemental file 1). Currently, published evidence related to the systematic use of aggregated PRO data for quality improvement initiatives and policy decisions among clinicians and other stakeholders, such as government representatives, is sparse. The increasing ability to collect PROs and the proliferation of Clinical Quality Registries internationally,16 has increased the need to establish rigorous evidence regarding the best feedback methods and preferred graphical formats for reporting PROs data to clinicians and other stakeholders. This current knowledge gap exists even though there is a strong policy rationale to support high functioning, mature Clinical Quality Registries with the ability to feedback data (both clinical and PROs)17 and despite the fact that promoting the collection of PRO data is published as a priority for Australia in the 2020–2025 National Health Reform Agreement.18 The current study was initiated to assess the perceived barriers and opportunities for uptake of PRO data among clinicians and to establish best practice feedback approaches to be adopted by Clinical Quality Registries. Since 2012, PROMs (generic and stroke-specific) have been collected by The Florey as part of its role in data management processes of the Australian Stroke Clinical Registry (AuSCR).19 Therefore, the condition of stroke was chosen as a case study.

Participants

The Working Group was established following the circulation of an expression of interest by the Victorian Stroke Clinical Network. Our objective was to achieve broad, interdisciplinary representation of experts and those with direct interests in the transparent feedback of PRO data. Individuals who were based in metropolitan or regional areas in Victoria and working in stroke care from acute, subacute and community settings were eligible to self-nominate for membership. We also included a consumer representative who was a survivor of stroke and had experience working with the Victorian government on various committees (see online supplemental file 1).

Stage 1: codesign workshops

In stage 1, we conducted three workshops between August and October 2019, with the option for participants to attend face-to-face or via the videoconferencing platform Zoom.20 The meetings were chaired by authors DAC/SB and the content discussion, for example, data templates were facilitated by an experienced moderator (VM). The first workshop oriented the Working Group members to the aims of the current study and was used to provide an outline of the general reporting principles from the existing literature.10 The identification of priorities and opinions was facilitated through the use of open-ended questions (eg, ‘Who has used PROs data reported for their health service?’, ‘Do you think PROs data add value, why/why not?’). At the conclusion of the first workshop, it was agreed that the Florey project team (authors OFR, SLH, VM, SB and DAC) would develop a set of templates based on three PROMs commonly used in clinical practice. These included: the modified Rankin Scale (mRS),21 EuroQol Five Dimensions Three Levels (EQ-5D-3L)22 and the Hospital Anxiety and Depression Scale (HADS)23 (see online supplemental table I). The PROMs were pragmatically chosen based on the experience of the project team with input from VAHI representatives.24 All templates used fictitious hospital, state and/or national-level data. For the templates designed to facilitate benchmarking of PRO data, peer hospitals were defined as those facilities with a similar bed size/number of admissions per year and population served. For these types of templates, the data from all eligible peer hospitals were included as a comparison. The number of peer hospitals is specified in the relevant graph axis label, figure legend or instructional footnote. The initial set of templates developed after the first workshop were circulated to the Working Group prior to the second workshop, via an electronic survey developed using SurveyMonkey software.25 The Working Group members then ranked the templates according to preference. During the second workshop, members discussed the results of the survey and critiqued the templates to identify the prevailing preferences using Zoom live polling.20 Refinements to the templates were made, then circulated via an electronic survey prior to the third workshop. The third workshop mirrored that of the second, whereby the survey results were discussed and members could revise their preferences. Agreement was reached for the final set of templates to be shortlisted in stage 2 (see online supplemental files 2–9). All workshops were audio recorded. At the conclusion of each workshop, formal minutes and the aggregated survey results were circulated to all members.

Stage 2: anonymous electronic survey

An anonymous electronic survey, from a broad sample of potential PRO data end users who had not been part of stage 1, was used to evaluate the feedback templates finalised in stage 1. The survey was designed using SurveyMonkey software.25 Survey respondents were recruited using a convenience sampling approach via invitations sent from several relevant clinical and academic networks. The survey link included instructions and outlined the voluntary participation requirements. Eligibility was self-determined by individuals who identified as a healthcare professional or academic/clinical researcher interested in the use of PRO data. Where feasible, snowball sampling was also encouraged, whereby respondents could share the link with other colleagues. The survey remained open for four weeks. Reminders to the main dissemination groups were sent one week prior to the closure of the survey. The survey included questions on the characteristics of respondents, as well as feedback on each of the PRO templates for each specific measure, including mRS, EQ-5D-3L and the HADS. A set of templates that presented PRO data over time were also included. Using previous PRO-related stakeholder survey methods reported by Brundage et al,26 a series of scale response questions (0=least, 10=most) were asked to ascertain: (a) the respondents’ rating of the perceived utility: ‘How useful do you find this graph?’, (b) the respondents’ rating of their ability to understand each sample format: ‘How easy it is for you to understand this graph?’. The respondents then selected the preferred template out of all the options for each category and could provide a free-text comment. Multiple-choice and open-ended questions were included to elicit perceived barriers and enablers related to the use of PRO data. There were three survey questions specific to barriers and enablers, for example, ‘What are the main barriers to the use of PROMs in your clinical practice?’.

Stage 3: final consensus workshop

The Working Group was reconvened for a fourth (final) workshop to review the results of the anonymous electronic survey. The workshop discussion was guided by the same moderator (VM) using open-ended questions to achieve a final consensus on the recommended PRO feedback templates (eg, ‘Could this graph be modified?’, ‘Do you agree this format is the preferred template for this PROM, why/why not?’).

Statistical analysis

The survey data collected as part of stage 1 (preworkshop and live polls) and stage 2 (electronic survey) were extracted and summarised using Microsoft Excel (V.2016).27 The quantitative data were summarised using descriptive analyses (totals and proportions). The stage 2 anonymous survey scale data were summarised as medians and IQRs. The stage 2 survey data were analysed overall, as well as according to clinical versus non-clinical respondents. Clinical respondents were defined as those with a medical, nursing or allied health background. Non-clinical respondents were defined as health system administrators, policymakers or researchers. The free-text qualitative data collected as part of the stage 2 survey were analysed using thematic analysis.28 An inductive approach was used, whereby one member of the project team identified broad theme categories related to each question (VM). A second reviewer (OFR) cross checked the qualitative responses to verify the themes. Both the qualitative and quantitative data were used with the objective of determining prevailing preferences and preferred formatting attributes.

Patient and public involvement

The main consumer group for this research were clinicians for which the data summary reports are focused. A patient consumer representative was included as part of the Working Group for this project. No other public representatives or patients were involved in the overall study. Clinician representatives were asked to advise on the interpretation and write up of the results, and several members of our working group were included as coauthors on this article.

Results

The Working Group comprised 33 members (19 clinicians, 1 consumer representative, 13 project team members) (table 1). The majority of clinical representatives were women (74%) and from an allied health (42%) or nursing (26%) profession. Three members were from other professional backgrounds, including psychology and pharmacy. Working Group members from a clinical background had a median of 12.5 years in clinical practice. Stroke management was the field of expertise for 84% (n=16) of the clinical representatives.
Table 1

Demographic characteristics of the project working group (N=33)*

CharacteristicCliniciansN=19n (%)Project teamN=13n (%)
Male5 (26)2 (15)
Profession
 Doctor4 (21)
 Nurse5 (26)
 Allied health8 (42)
 Other†2 (11)13 (100)
Clinical area of expertise
 Stroke16 (84)13 (100)
 Other3 (16)
Median years in clinical practice (Q1, Q3)12.5 (8, 15)

*Includes one male consumer representative.

†Other: refers to professional backgrounds including psychology and pharmacy.

Demographic characteristics of the project working group (N=33)* *Includes one male consumer representative. †Other: refers to professional backgrounds including psychology and pharmacy.

Stage 1: codesign workshops

Overall, 31 Working Group members actively participated in one or more workshops (see online supplemental file 1). The number and type of templates presented during each workshop are summarised in online supplemental table II. The templates were refined or retired based on the preferences identified during each workshop. The formatting preferences included: simple layouts or symbols to reduce complexity, use of definitions and instructions (where appropriate), use of normative population data, CIs and denoting sample sizes (see online supplemental table III). Several templates were determined to be unsuitable for service-level PRO reporting, including spider plots, heat maps and pictographs. The Working Group agreed these formats were more complex and would limit the ability to readily interpret the results, while pictographs were considered too simplistic with insufficient detail for this audience. Following the third workshop, a separate category containing templates used for reporting PROs data over time was supported for wider testing, resulting in six additional templates.

Stage 2: anonymous electronic survey

There were 114 respondents for this survey (table 2). The majority of respondents were women (76%) and aged between 30 and 49 years (56%). Respondents were predominantly clinical representatives (91%), including allied health (38%) or nursing (32%). Overall, 87 respondents (71%) reported stroke as their area of clinical expertise. The respondents had a median of 20 years in clinical practice, and 23% reported using PROs at the time of the survey. The following sections detail the results for each PRO category assessed as part of the survey. The denominators vary since not all respondents completed every question used to evaluate the templates. The overall response rate for the survey was 61%.
Table 2

Characteristics of the stage 2 electronic survey respondents

Respondent characteristicsSurvey respondentsN=114 n (%)
Female87 (76)
Age group (years)
 <306 (5)
 30–4964 (56)
 50–6440 (35)
 65+4 (4)
Profession
Clinical:104 (91)
 Doctor24 (21)
 Nurse37 (32)
 Allied health43 (38)
Non-clinical:10 (9)
 Management1 (1)
 Researcher4 (3)
 Government representative1 (1)
 Health system administration2 (1)
 Quality/safety2 (1)
Clinical area of expertise
 Stroke81 (71)
 Cancer5 (4)
 Other (ie, rehabilitation, geriatrics etc.)28 (25)
Current work role*
 Clinical care82 (71)
 Education27 (23)
 Research47 (41)
 Management26 (23)
 Other (eg, quality/safety, integrated cancer service etc.)6 (5)
Median years in practice (Q1, Q3)20 (10, 29)
Completed a higher degree (master’s or doctorate)71 (62)
Currently use PROMs in clinical practice26 (23)

*Multiple responses were permitted for this question therefore percentages do not add up to 100.

PROMs, patient reported outcome measures.

Characteristics of the stage 2 electronic survey respondents *Multiple responses were permitted for this question therefore percentages do not add up to 100. PROMs, patient reported outcome measures.

Preferences for graphical formats of summary data

When interpreting mRS data, overall, 41% (n=37/91) of survey respondents preferred format A (see online supplemental files 2 & 3) which reflected a two-bar horizontal stacked bar chart with comparative data (table 3). The prevailing template had the greatest rating for ability to understand the data being presented (median=9), despite only a two-point median difference between the five templates presented (see online supplemental table IV). The responses of non-clinical respondents were similar to clinical respondents, with a preference for an additional table with mRS data values paired with the stacked bar chart.
Table 3

Respondent preferences for templates displayed in stage 2 survey, overall and by profession

Patient-reported outcome measureTemplate reference*OverallN=114n/NMedicalN=19n/NNursingN=33n/NAllied HealthN=43n/NOtherN=8n/N
Modified Rankin Scale Format A 37/91 (41%) 10/19 (53%) 15/33 (46%) 12/31 (39%) 0
Format B21/91 (23%)4/19 (21%)5/33 (15%)7/31 (23%) 5/8 (63%)
Format C12/91 (13%)1/19 (5%)5/33 (15%)6/31 (19%)0
Format D11/91 (12%)1/19 (5%)4/33 (12%)5/31 (16%)1/8 (12%)
Format E10/91 (11%)3/19 (16%)4/33 (12%)1/31 (3%)2/8 (25%)
EuroQol Visual Analogue Scale Format A 37/82 (45%) 9/19 (47%) 18/30 (60%) 9/26 (35%)1/7 (14%)
Format B29/82 (35%)6/19 (32%)10/30 (33%) 10/26 (38%) 3/7 (43%)
Format C16/82 (20%)4/19 (21%)2/30 (7%)7/26 (27%) 3/7 (43%)
EuroQol Five Dimension Descriptive System Format A15/78 (19%)2/19 (11%) 9/30 (30%) 4/23 (17%)0
Format B10/78 (13%)1/19 (6%)5/30 (17%)4/23 (17%)0
Format C3/78 (4%)03/30 (10%)00
Format D7/78 (9%)4/19 (22%)1/30 (3%)2/23 (9%)0
Format E14/78 (18%)4/19 (22%)4/30 (13%)2/23 (9%) 4/6 (67%)
Format F 22/78 (28%) 5/19 (28%) 7/30 (23%) 8/23 (35%) 2/6 (33%)
Format G6/78 (8%)2/19 (11%)1/30 (3%)3/23 (13%)0
Hospital Anxiety and Depression Scale Format A14/75 (19%)2/19 (11%) 9/28 (32%) 3/22 (14%)0
Format B 24/75 (32%) 7/19 (39%) 7/28 (25%) 8/22 (36%) 2/6 (33%)
Format C17/75 (23%)5/19 (28%)5/28 (18%)6/22 (27%)1/6 (17%)
Format D19/75 (26%)4/19 (22%)7/28 (25%)5/22 (23%) 3/6 (50%)
Longitudinal data Format A7/69 (10%)1/17 (6%)5/28 (18%)1/19 (5%)0
Format B14/69 (20%) 5/17 (29%) 5/28 (18%)4/19 (21%)0
Format C3/69 (4%)02/28 (7%)1/19 (5%)0
Format D12/69 (17%)4/17 (24%)3/28 (11%)4/19 (21%)1/5 (20%)
Format E 21/69 (30%) 4/17 (24%) 9/28 (32%) 5/19 (26%) 3/5 (60%)
Format F12/69 (17%)3/17 (17%)4/28 (14%)4/19 (21%)1/5 (20%)

Bolded figures represent the overall preference of ranking for each respondent type (displayed by profession) category according to the survey data. Other: refers to non-clinical professions such as health service/department managers, researchers, etc. The denominators used for each row proportion differ due to survey respondents terminating the survey at different levels of completion.

*See online supplemental files 2-9 for a copy of the templates used for each format.

Respondent preferences for templates displayed in stage 2 survey, overall and by profession Bolded figures represent the overall preference of ranking for each respondent type (displayed by profession) category according to the survey data. Other: refers to non-clinical professions such as health service/department managers, researchers, etc. The denominators used for each row proportion differ due to survey respondents terminating the survey at different levels of completion. *See online supplemental files 2-9 for a copy of the templates used for each format. For EQ-VAS data, 45% of respondents (37/82) preferred format A (see online supplemental file 4) comprised of a bar chart with the inclusion of CIs and stratification by patient sex and age group. Compared with the more complex formats (eg, format C: caterpillar plot), the bar chart resulted in the greatest rating (median=7) for both perceived ease of understanding and usefulness (see online supplemental table IV). The median ratings across the three EQ-VAS templates were similar, with a maximum median difference of two points only. The responses of non-clinical respondents differed to clinical respondents, with a preference for a dot or caterpillar plot. When interpreting EQ-5D-3L data, format F (see online supplemental files 5 & 6), which comprised a stacked bar chart presenting data for all five dimensions (with categorisation according to the proportion of patients reporting ‘no problems’ vs those reporting ‘problems’), was identified as the preferred format (22/78, 28%). Despite the greatest proportion of respondents preferring format F, there was a diverse range of median ratings across all seven templates. Several anomalies were identified whereby four other templates (eg, formats A and B with vertical stacked bar chart and traffic light colour coding) received a median rating of ≥8 for perceived ease of understanding, which was greater than that of the prevailing template. The preference of non-clinical respondents differed from clinical respondents, with format E comprised of a table of data values paired with a horizontal bar chart prevailing. In addition, the preferred format (24/75, 32%) for presenting HADS data was format B, a bar chart with the inclusion of a dotted line to depict a clinical cut-off point, which indicated possible anxiety or depression (see online supplemental file 7). A total of 21/69 (30%) respondents preferred format E (see online supplemental files 8 & 9), a line graph template for presenting data over time. The line graph that was short listed as the preference displayed hospital data and comparative state-level hospital data, with additional metrics to stratify results according to patient characteristics (eg, stroke type) as well as a dotted line to indicate normative population estimates. Although this template prevailed, there were two other variations of a line graph (formats A and B) that had identical median ratings (understanding=8, usefulness=7), along with two bar graphs (formats C and D). The consistent median ratings across all formats did not align with the differences observed between the overall rankings of preference. There was no difference for the overall preference between clinical and non-clinical respondents. Several themes emerged from the qualitative data related to formatting preferences, including: the use of colours, particularly intuitive colours such as green=better/good outcome and red=worse/poorer outcome (see online supplemental table III). Though, the use of dark/shaded colours was cautioned to ensure templates would be interpretable in grayscale (eg, to facilitate printed versions for hospital end users). In addition, the inclusion of clear labelling (namely graph axes and titles), sample sizes and proportions were also identified as preferred features. Though, the inclusion of too much explanatory text was cautioned. The respondents considered stratification of PRO summary data by age, sex and other variables to be useful, particularly to assist with targeted quality improvement strategies. The inclusion of benchmarked peer-level, state-level and/or national-level data was also supported to facilitate hospital performance monitoring. The respondent’s qualitative responses identified that benchmarks provided information about targets while peer data allowed for comparison against similar hospitals and within jurisdictions. The respondents also preferred the term ‘peer hospitals’ to be clearly defined. The need to provide case-mix adjusted data was identified by several respondents across various templates, with the importance of this statistical adjustment highlighted for the presentation of longitudinal PRO data (eg, otherwise variation could be due to the difference in patient characteristics each year).

PRO data reporting interval preferences and perceived barriers and enablers

The majority of respondents (60/69, 87%) preferred a 3 to 6-month interval for PRO data reporting. The main perceived barrier to using PRO data in clinical practice from the summarised qualitative data, included resource and time constraints (35/68, 51%) (including staff time to read and interpret the information). Other perceived barriers included a lack of tools available to facilitate meaningful reporting of PROs data, a lack of understanding about how to use the information and lack of organisational support to use the data or a perceived need for culture change. Two of the most common themes identified as enabling factors included: an interest to use PROs as a mechanism to deliver patient-centred care, and the development of enhanced feedback methods to facilitate greater PRO data uptake and future use. Other reported enablers included the need for adequate resources such as time and funding for clinicians to interpret and use the data and enhanced education about how best to interpret and use the data to improve patient care.

Stage 3: Final consensus workshop

The Working Group agreed with the outcomes of the stage 2 survey results and the shortlisted templates. The median ratings for each template, especially for formats where anomalies, were identified in contrast to the overall preference, were used as a supplement for the Working Group to consider. Figure 2 illustrates two templates recommended for presenting: data at a single time point versus data over time, along with a summary of preferred formatting features. An example of a template that was identified to not be appropriate for routine feedback purposes is presented in figure 3. The Working Group agreed that the professional roles of the recipients of these data would require consideration, and that certain recipients (eg, hospital executives) may require further detail such a pairing a graphical display with a table of data values. The Working Group discussed several data quality aspects of PROs (eg, the need for case-mix adjustment) which were supported for incorporation into the final project recommendations (see online supplemental table V).
Figure 2

Example illustrating two recommended templates to use for reporting aggregate, service-level patient reported outcomes data to clinicians.

Figure 3

Example illustrating a non-preferred template for reporting aggregate, service-level patient reported outcomes data to clinicians.

Example illustrating two recommended templates to use for reporting aggregate, service-level patient reported outcomes data to clinicians. Example illustrating a non-preferred template for reporting aggregate, service-level patient reported outcomes data to clinicians.

Discussion

PROs data have the potential to improve healthcare and patient outcomes,29 30 but in order for this to occur, it is essential that PROs data are presented in a manner that is both useful and understandable.31 32 The presentation of aggregated PROs data is often done in the absence of best practice guidance.5 Guided by evidence from our previous literature review,10 this study aimed to use a codesign process to create a set of PRO summary data templates for health service-level reporting. The recommended templates and preferred formatting attributes identified from our study complement existing mechanisms proposed by Snyder et al31 and contribute much needed evidence for methods to use when communicating service-level PRO data to clinicians. Despite some mixed opinions among the clinical representatives related to each template, a clear preference emerged to support the provision of these data. Our findings indicate that the simpler formats reduced the cognitive burden for clinicians, and there was a preference for use of explanatory attributes to aid interpretation of the data. For data presented at a single time point, a bar or stacked bar chart with up to four bars prevailed. A caveat to this recommendation was for PROMs with multiple dimensions, such as the EQ-5D, where clinicians preferred to review local and comparison hospital data within the one graph (up to 10 bars). A line graph with a maximum of four lines and minimal additional constructs was identified as the preference for data presented over time. A construct can include statistical or formatting features such as: stratification according to patient demographic/clinical characteristics or providing a reference to normative population data or clinical cut-offs.33 Our main findings are consistent with the preliminary principles for graphical display of PRO data proposed by Bantug et al5 and builds on the ‘less is more’ guidance recommended for effective communication of performance monitoring34 or clinical data in general.35 The diversity of opinions among clinicians in the Working Group and broader survey sample in our study is not dissimilar to previous research. Brundage et al1 evaluated the preferences of graphically displayed PRO results among 233 clinicians and found differing preferences for bar and pie charts. In the current study, we also found that clinicians had differing preferences related to the use of pie charts compared with bar charts as well as intuitive colour coding schemes36 37 (eg, some clinicians indicated that traffic light colours added visual clarity, whereas others disliked them). Furthermore, the discrepancies identified from the median ratings in this current study, in which several templates received a rating of >8 despite not being identified as the prevailing format, highlight the challenge of recommending a single format style for all target audiences. Despite some mixed views, and mostly representation from professionals working in the clinical area of stroke, the recommendations from this current study form an empirical basis for future work in this field. We recommend that there is a need to involve hospital end users, including representatives from quality improvement departments, if aggregate PRO data are being prepared by academics, and that these data should be presented in more than one way to address the needs of specific audience subgroups. The barriers and enablers identified in this current study, including resource and time constraints and education for clinicians, align with the findings of our previously conducted scoping review into PRO feedback.10 We also found a preference for the majority of the shortlisted templates to include comparative hospital data. In our examples, we followed the approach used in the AuSCR to provide peer-hospital, state-level and national-level benchmarks. Case-mix adjustment for PROs data is an emerging field38 and the application of case-mix methods is a prerequisite for quality reporting and benchmarking purposes. Further empirical research is needed to identify patient-related (eg, age, sex, life events, new healthcare episodes) and hospital-related (workload, volume of patients, hospital type, etc) factors requiring adjustment for reporting benchmarked PRO data.12 It was intriguing that only a single survey respondent considered the importance of case-mix adjustment, and this might suggest that education for clinicians about the importance of case-mix adjustment when reviewing comparisons of PROs at the service level be undertaken. This current study contributes to the discourse in confirming the need for ongoing methodological research to determine the most appropriate analytic methods and the variables needed to enable reliable case-mix adjustment. In the interim, it is recommended that when providing descriptive PROs comparisons without the ability to adjust for differences in patient characteristics, a concise and explicit definition of what constitutes a peer hospital must be included in the template.

Strengths and limitations

The strengths of our study include the engagement of an interdisciplinary sample of clinical, academic and government representatives who were members of the Working Group, and other potential end users who completed the electronic survey. This is one of only a few studies in the field, outside of a cancer setting,1 9 32 to iteratively develop and evaluate the preferences of clinicians when considering the use of PROs for comparing hospital performance. Applying a modified Delphi process throughout each stage was also a strength as well as the use of multiple engagement tools to optimise participation of Working Group members (eg, live polling). An additional strength was the use of different commonly used PROs measures as part of the templates. While previous work has focused on the development of templates for PRO measures for a single condition,26 we trialled both condition-specific and generic PRO measures in an effort to increase the generalisability of our results across different clinical populations. The limitations of our study include that our templates were based on experience reporting PROs as part of the AuSCR and informed by the literature review. Some subjective interpretation of how to display data and how many templates to produce for assessment was unavoidable. Although we sought broad representation for the survey by advertising widely, we acknowledge that the majority of the Working Group were experienced in the field of stroke and were located within a single jurisdiction. A further limitation was our inability to include examples of the presentation of baseline and follow-up PRO data due to the sample data being derived from the stroke registry where PROs are only collected at a single time point after the acute event. We acknowledge these data are important for a range of conditions such as chronic kidney disease or surgical interventions, where progression may be important to monitor. We acknowledge the presence of responder bias in our survey data and that the ability to detect preferences among different professional audience groups (eg, medical/nursing staff vs government representatives/researchers) was limited by the composition of mostly female respondents who self-selected to participate in the survey. Despite this, our findings align with those from preliminary work conducted in other patient populations,9 39 which suggests that there are general principles for the presentation of aggregate PRO data that are applicable across clinical specialities. We acknowledge that the next steps are to pilot and refine the recommended templates and encourage others to consider evaluating these as well.

Conclusions

We have illustrated the iterative process and outcomes of a codesigned approach to establishing summary data templates for reporting aggregate service-level PROs data to clinicians. Simple graphical templates, with accompanying instructions and formatting attributes to aid data interpretation, were identified as the preferred format characteristics. This work provides important evidence for Clinical Quality Registries and other organisations that routinely feedback aggregate PRO data, to ensure that the potential of these data to support quality improvement efforts is fully realised.
  28 in total

1.  Engaging stakeholders to improve presentation of patient-reported outcomes data in clinical practice.

Authors:  Katherine C Smith; Michael D Brundage; Elliott Tolbert; Emily A Little; Elissa T Bantug; Claire F Snyder
Journal:  Support Care Cancer       Date:  2016-05-10       Impact factor: 3.603

Review 2.  Implementing patient-reported outcomes assessment in clinical practice: a review of the options and considerations.

Authors:  Claire F Snyder; Neil K Aaronson; Ali K Choucair; Thomas E Elliott; Joanne Greenhalgh; Michele Y Halyard; Rachel Hess; Deborah M Miller; Bryce B Reeve; Maria Santana
Journal:  Qual Life Res       Date:  2011-11-03       Impact factor: 4.147

3.  Integrating the use of patient-reported outcomes for both clinical practice and performance measurement: views of experts from 3 countries.

Authors:  Philip J Van Der Wees; Maria W G Nijhuis-Van Der Sanden; John Z Ayanian; Nick Black; Gert P Westert; Eric C Schneider
Journal:  Milbank Q       Date:  2014-12       Impact factor: 4.911

4.  Patient-reported outcomes in the Swedish National Quality Registers.

Authors:  E Nilsson; L Orwelius; M Kristenson
Journal:  J Intern Med       Date:  2015-08-26       Impact factor: 8.989

5.  Presenting comparative study PRO results to clinicians and researchers: beyond the eye of the beholder.

Authors:  Michael Brundage; Amanda Blackford; Elliott Tolbert; Katherine Smith; Elissa Bantug; Claire Snyder
Journal:  Qual Life Res       Date:  2017-11-02       Impact factor: 4.147

Review 6.  Feedback of patient-reported outcomes to healthcare professionals for comparing health service performance: a scoping review.

Authors:  Shaun L Hancock; Olivia F Ryan; Violet Marion; Sharon Kramer; Paulette Kelly; Sibilah Breen; Dominique A Cadilhac
Journal:  BMJ Open       Date:  2020-11-23       Impact factor: 2.692

7.  Making a picture worth a thousand numbers: recommendations for graphically displaying patient-reported outcomes data.

Authors:  Claire Snyder; Katherine Smith; Bernhard Holzner; Yonaira M Rivera; Elissa Bantug; Michael Brundage
Journal:  Qual Life Res       Date:  2018-10-10       Impact factor: 4.147

8.  The Hospital Anxiety And Depression Scale.

Authors:  R Philip Snaith
Journal:  Health Qual Life Outcomes       Date:  2003-08-01       Impact factor: 3.186

9.  How do aggregated patient-reported outcome measures data stimulate health care improvement? A realist synthesis.

Authors:  Joanne Greenhalgh; Sonia Dalkin; Elizabeth Gibbons; Judy Wright; Jose Maria Valderas; David Meads; Nick Black
Journal:  J Health Serv Res Policy       Date:  2017-12-20

10.  Collection and use of EQ-5D for follow-up, decision-making, and quality improvement in health care - the case of the Swedish National Quality Registries.

Authors:  Olivia Ernstsson; Mathieu F Janssen; Emelie Heintz
Journal:  J Patient Rep Outcomes       Date:  2020-09-16
View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.