| Literature DB >> 34674415 |
Wolf E Hautz1, Moritz M Kündig2, Roger Tschanz2, Tanja Birrenbach1, Alexander Schuster3, Thomas Bürkle2, Stefanie C Hautz1, Thomas C Sauter1, Gert Krummrey1.
Abstract
OBJECTIVES: Identification of diagnostic error is complex and mostly relies on expert ratings, a severely limited procedure. We developed a system that allows to automatically identify diagnostic labelling error from diagnoses coded according to the international classification of diseases (ICD), often available as routine health care data.Entities:
Keywords: decision support; diagnostic error; quality improvement
Mesh:
Year: 2021 PMID: 34674415 PMCID: PMC9125795 DOI: 10.1515/dx-2021-0039
Source DB: PubMed Journal: Diagnosis (Berl) ISSN: 2194-802X
Figure 1:Example of an ICD taxonomy.
Dotted line: distance between two given exemplary diagnosis within this taxonomy is determined by counting the edges of the shortest path connecting them. Bold line: alternatively, the edges on the shortest connecting path between any two diagnoses can be determined by summing the weights of the edges connecting them.
Datasets used to validate the automated identification of diagnostic labelling errors.
| Data-set | Type | Size | Pairs of diagnosesa | Cases with diagnoses | Prevalence of error | |
|---|---|---|---|---|---|---|
| Similar | Discrepant | |||||
| 1 | Clinical | Seven hundred and fifty five patients, admission and discharge diagnosis | 755 (100%) | 662 | 93 | 12.30% |
| 2 | Educational | Twenty advanced students diagnosing six virtual patients each | 109 (90.34%) | 14 | 95 | 87.20% |
| 3 | Educational | Fifty one advanced students diagnosing 8 virtual patients alone [ | 272 (100%) | 180 | 92 | 66.2% |
|
|
|
|
|
| ||
aPercentages refer to pairs of diagnoses with complete data available for analysis. Bold values signify the column total.
Performance of four algorithms overall, and by type of dataset.
| Algorithm | Area under the curve AUC and (95% confidence interval) | ||
|---|---|---|---|
| Overall | Clinical dataset only | Educational datasets only | |
| Steps | 0.835 (0.812–0.859) | 0.853 (0.822–0.884) | 0.92 (0.88–0.959) |
| Weights | 0.830 (0.806–0.854) | 0.856 (0.826–0.887) | 0.924 (0.887–0.962) |
| Lietal | 0.821 (0.797–0.845) | 0.856 (0.825–0.887) | 0.923 (0.855–0.962) |
| WuPalmer | 0.821 (0.797–0.846) | 0.856 (0.825–0.887) | 0.923 (0.855–0.962) |
Figure 2:Sensitivity and specificity to identify diagnostic labelling errors, by algorithm and type of dataset.
Cut-off values for all for algorithms to achieve a given sensitivity (or specificity).
| Sensitivity | Specificity | Classify as discrepant if larger or equal to | Classify as discrepant if smaller or equal to | ||
|---|---|---|---|---|---|
| Steps | Weights | WuPalmer | Lietal | ||
| 0.00 | 1.00 | 11.00 | 26.00 | −1.00 | −1.00 |
| 0.01 | 1.00 | 9.50 | 24.85 | ||
| 0.08 | 0.97 | 8.50 | 24.55 | ||
| 0.38 | 0.88 | 7.50 | 23.65 | ||
| 0.51 | 0.80 | 12.85 | 0.1 | ||
| 0.51 | 0.81 | 17.80 | 0.05 | ||
| 0.66 | 0.80 | 6.50 | |||
| 0.67 | 0.80 | 12.55 | 0.24 | 0.15 | |
| 0.80 | 0.78 | 5.50 | 11.65 | 0.27 | 0.18 |
| 0.82 | 0.76 | 11.4 | 0.31 | 0.22 | |
| 0.83 | 0.76 | 4.50 | 5.85 | 0.42 | 0.28 |
| 0.89 | 0.73 | 3.50 | 4.65 | 0.54 | 0.44 |
| 0.91 | 0.71 | 2.50 | 1.95 | 0.71 | 0.60 |
| 0.97 | 0.67 | 0.50 | 0.45 | 0.93 | 0.89 |
| 1.00 | 0.00 | −1.00 | −1.00 | 2.00 | 2.00 |