| Literature DB >> 30100565 |
Matthew Barclay1, Mary Dixon-Woods1, Georgios Lyratzopoulos1,2.
Abstract
'The Problem with…' series covers controversial topics related to efforts to improve healthcare quality, including widely recommended but deceptively difficult strategies for improvement and pervasive problems that seem to resist solution. © Author(s) (or their employer(s)) 2019. Re-use permitted under CC BY. Published by BMJ.Entities:
Keywords: health services research; pay for performance; quality measurement; report cards
Year: 2018 PMID: 30100565 PMCID: PMC6559782 DOI: 10.1136/bmjqs-2018-007798
Source DB: PubMed Journal: BMJ Qual Saf ISSN: 2044-5415 Impact factor: 7.035
Common issues with selected composite indicators of care quality
| CMS Overall Hospital Star Rating | AHRQ PSI90 | Leapfrog Composite Patient Safety Score | MyNHS Overall Stroke Care Rating | NHS England Overall Patient Experience Score | ||
| Transparency | Are all important m | Yes, but spread across several documents and web pages | Yes, but spread across several documents | Yes | No, searches of MyNHS and SSNAP websites did not find a comprehensive methods | Yes |
| Selection of individual measures | Are the measures used equally applicable across all rated hospitals? | No, some hospitals do not report all measures | Yes | Yes | Yes | Yes |
| Underlying measures and data | Is missing measure information handled in a way that can introduce bias? | Yes, pairwise deletion is used | Yes, effectively using mean imputation | Yes, pairwise deletion is used where proxy measures are not available | Yes, pairwise deletion is used | No missing measure information |
| Are component measures adequately adjusted for case-mix? | Some but not all measures | Yes | Yes | Not discussed in identified methods | Yes | |
| Use of banding onto consistent scales | Are measures standardised using banding? | No | No | No | Yes | No |
| Choice of weights | Is there an apparent justification for the weights used? | Yes, but reason for the precise weights used is unclear | Yes | Yes | No | No |
| Is any sensitivity analysis of the choice of weights reported? | No | Yes | No | No | No | |
| Uncertainty | Is the uncertainty in the final composite rating presented? | Not in the star rating | Yes | No | No | Yes |
Requirements, steps forward and remaining challenges for robust and useful composite indicators
| Requirement | Steps forward | Remaining challenges |
| Transparency | Being clear about who is involved in making decisions in developing the composite indicator. | Many stakeholders may be involved. The design may evolve in unexpected ways over time. |
| Fully describing the decision-making process, reporting the reasons and justifications for the decisions made. | ||
| Purpose-led design | Selecting individual measures to cover the full range of services intended to be measured by the composite. | Identifying appropriate individual measures. Appropriate measures may not exist for all areas included in the composite. |
| Choosing weights that reflect the relative importance of the different quality measures. | Balancing the weighting system against competing priorities. | |
| Technical reproducibility | Providing clear and comprehensive technical documentation. | |
| Reporting full definitions of the individual underlying measures and how they are combined. | Individual measures may only be available from sources that do not fully document the details, but these measures should not be used in the composite. | |
| Publishing the code used in data processing and statistical analysis. | ||
| Statistical fitness | Performing appropriate statistical case-mix adjustment. | Accurate patient-level data may not exist for important case-mix factors. Adequate statistical case-mix adjustment may not be possible. Interpretable results may require further processing. |
| Using reporting periods long enough to give acceptable reliability. | Longer reporting periods may be necessary to increase reliability, but impedes use in driving quality improvement. | |
| Standardising measures to consistent scales in a principled way that preserves the useful information in the underlying measures. | Understanding what good and bad performance in the real world looks like on each measure. |