| Literature DB >> 22844445 |
Alan J Forster1, Monica Taljaard, Carol Bennett, Carl van Walraven.
Abstract
BACKGROUND: Adverse events are poor patient outcomes caused by medical care. Their identification requires the peer-review of poor outcomes, which may be unreliable. Combining physician ratings might improve the accuracy of adverse event classification.Entities:
Mesh:
Year: 2012 PMID: 22844445 PMCID: PMC3406022 DOI: 10.1371/journal.pone.0041239
Source DB: PubMed Journal: PLoS One ISSN: 1932-6203 Impact factor: 3.240
Description of the reviewers.
| Characteristic | N = 30 |
|
| |
| Canada | 6 (20%) |
| US | 24 (80%) |
|
| 9 (4, 42) |
|
| |
| Yes | 20 (66.7%) |
| Years of experience | 3 (0.1, 30) |
| No | 10 (33.3%) |
|
| |
| Yes | 5 (16.7%) |
| No | 25 83.3%) |
|
| |
| Agree | 19 (63.3%) |
| Agree somewhat | 11 (36.7%) |
| Neutral | 0 |
| Disagree somewhat | 0 |
| Disagree | 0 |
Median (range).
Summary of the 30 individual peer-reviewers of the 319 cases.
| Reviewer number | Median score on six point ordinal scale | Number of cases rated as an adverse event | Model-based reviewer operating characteristic | |
| Sensitivity | Specificity | |||
| 1 | 1 (1,2) | 22 (7) | 0.0562 | 0.92 |
| 2 | 1 (1,2) | 43 (13) | 0.2703 | 0.9824 |
| 3 | 1 (1,3) | 51 (16) | 0.25 | 0.9181 |
| 4 | 1 (1,4) | 90 (28) | 0.5542 | 0.9532 |
| 5 | 2 (1,3) | 55 (17) | 0.3524 | 0.9832 |
| 6 | 2 (1,3) | 76 (24) | 0.3784 | 0.883 |
| 7 | 2 (1,4) | 84 (26) | 0.3992 | 0.8542 |
| 8 | 2 (1,4) | 86 (27) | 0.3129 | 0.7678 |
| 9 | 2 (1,4) | 96 (30) | 0.4952 | 0.8671 |
| 10 | 2 (1,5) | 103 (32) | 0.527 | 0.8537 |
| 11 | 2 (1,5) | 125 (39) | 0.7717 | 0.9366 |
| 12 | 2 (1,5) | 134 (42) | 0.8347 | 0.9386 |
| 13 | 2 (1,5) | 137 (43) | 0.8159 | 0.9048 |
| 14 | 2 (1,6) | 135 (42) | 0.794 | 0.8975 |
| 15 | 3 (1,4) | 129 (40) | 0.6847 | 0.838 |
| 15 | 3 (1,5) | 130 (41) | 0.6964 | 0.8424 |
| 17 | 3 (1,5) | 156 (49) | 0.7858 | 0.7676 |
| 18 | 3 (1,6) | 155 (49) | 0.8404 | 0.8207 |
| 19 | 3 (2,4) | 93 (29) | 0.5815 | 0.9593 |
| 20 | 3 (2,4) | 115 (36) | 0.4909 | 0.7523 |
| 21 | 3 (2,5) | 154 (48) | 0.7665 | 0.7627 |
| 22 | 4 (1,5) | 162 (51) | 0.8889 | 0.8217 |
| 23 | 4 (1,5) | 167 (52) | 0.875 | 0.7804 |
| 24 | 4 (1,5) | 171 (54) | 0.8485 | 0.7342 |
| 25 | 4 (2,5) | 160 (50) | 0.8484 | 0.7984 |
| 26 | 4 (2,5) | 176 (55) | 0.891 | 0.7417 |
| 27 | 4 (2,5) | 181 (57) | 0.9025 | 0.7224 |
| 28 | 4 (3,4) | 190 (60) | 0.706 | 0.4999 |
| 29 | 4 (3,5) | 204 (64) | 0.9314 | 0.613 |
| 30 | 5 (2,5) | 218 (68) | 0.9104 | 0.513 |
|
|
|
|
|
|
Physician reviewers were numbered based on their median 6-point scale score ranking.
1– Definitely due to disease, 6– definitely due to medical care.
Ordinal scale dichotomized: 1–3 ‘Due to disease’; 4–6 ‘Due to medical care’.
Area under the receiver operating characteristic curve.
Figure 1Kappa scores for 30 reviewer pairs.
This plot presents the distribution of the kappa statistics for all possible pairwise comparisons between the 30 reviewers for all 319 cases.
Posterior probability of event or non-event given the number of positive ratings.
| Total Number of Adverse Event ratings out of 30 raters | Probability that case truly is an Adverse Event |
| 0 (0%) | 0.0000 |
| 1 (3.3%) | 0.0000 |
| 2 (6.6%) | 0.0000 |
| 3 (9.9%) | 0.0000 |
| 4 (13.2%) | 0.0000 |
| 5 (16.5%) | 0.0000 |
| 6 (20.0%) | 0.0000 |
| 7 (23.3%) | 0.0000 |
| 8 (26.7%) | 0.0003 |
| 9 (30.0%) | 0.0026 |
| 10 (33.3%) | 0.0219 |
| 11 (36.7%) | 0.1612 |
| 12 (40.0%) | 0.6225 |
| 13 (43.3%) | 0.9340 |
| 14 (46.7%) | 0.9918 |
| 15 (50.0%) | 0.9990 |
| 16 (53.3%) | 0.9999 |
| 17 (56.7%) | 1.0000 |
| 18 (60.0%) | 1.0000 |
| 19 (63.3%) | 1.0000 |
| 20 (66.7%) | 1.0000 |
| 21 (70.0%) | 1.0000 |
| 22 (73.3%) | 1.0000 |
| 23 (76.7%) | 1.0000 |
| 24 (80.0%) | 1.0000 |
| 25 (83.3%) | 1.0000 |
| 26 (86.7%) | 1.0000 |
| 27 (90.0%) | 1.0000 |
| 28 (93.3%) | 1.0000 |
| 29 (96.7%) | 1.0000 |
| 30 (100%) | 1.0000 |
These estimates are from the latent class model that assumed a common sensitivity and specificity for all reviewers.
Number of adverse event ratings required for true probability of adverse event to exceed 50%, 75%, and 95%.
| Number of cases that need to be rated as an adverse event required for true probability of an adverse event | |||
| Number of Raters Per Case | ≥50% | ≥75% | ≥95% |
| 2 | 1 | 2 | Not possible |
| 3 | 2 | 2 | 3 |
| 4 | 2 | 3 | 3 |
| 5 | 2 | 3 | 4 |
| 6 | 3 | 3 | 4 |
| 7 | 3 | 4 | 5 |
| 8 | 4 | 4 | 5 |
| 9 | 4 | 5 | 5 |
| 10 | 4 | 5 | 6 |
| 11 | 5 | 5 | 6 |
| 12 | 5 | 6 | 7 |
| 13 | 6 | 6 | 7 |
| 14 | 6 | 7 | 7 |
| 15 | 6 | 7 | 8 |
This table enables the identification of the number of raters required to obtain a given certainty that a case ‘truly’ represents an adverse event. For example, if an investigator wished to be 95% certain a case represented an adverse event, then it would be necessary to have a minimum of three reviewers and all three would need to agree. With only two reviewers, once could be at best 75% certain a case represented an adverse event.