| Literature DB >> 26306281 |
Jonathan H Chen1, Russ B Altman2.
Abstract
Uncertainty and variability is pervasive in medical decision making with insufficient evidence-based medicine and inconsistent implementation where established knowledge exists. Clinical decision support constructs like order sets help distribute expertise, but are constrained by knowledge-based development. We previously produced a data-driven order recommender system to automatically generate clinical decision support content from structured electronic medical record data on >19K hospital patients. We now present the first structured validation of such automatically generated content against an objective external standard by assessing how well the generated recommendations correspond to orders referenced as appropriate in clinical practice guidelines. For example scenarios of chest pain, gastrointestinal hemorrhage, and pneumonia in hospital patients, the automated method identifies guideline reference orders with ROC AUCs (c-statistics) (0.89, 0.95, 0.83) that improve upon statistical prevalence benchmarks (0.76, 0.74, 0.73) and pre-existing human-expert authored order sets (0.81, 0.77, 0.73) (P<10(-30) in all cases). We demonstrate that data-driven, automatically generated clinical decision support content can reproduce and optimize top-down constructs like order sets while largely avoiding inappropriate and irrelevant recommendations. This will be even more important when extrapolating to more typical clinical scenarios where well-defined external standards and decision support do not exist.Entities:
Year: 2015 PMID: 26306281 PMCID: PMC4525236
Source DB: PubMed Journal: AMIA Jt Summits Transl Sci Proc
– Admission diagnoses evaluated, number of patients in training dataset, number of candidate orders referenced as appropriate in clinical practice guidelines, number of candidate orders available in pre-authored order sets, and the intersection between the latter two.
| Admission Diagnosis (ICD9) | Training Patients | Guideline Reference Orders | Order Set Items | Guideline Reference Orders in Order Set |
|---|---|---|---|---|
| GI Bleed (578) | 282 | 38 | 51 | 22 |
| Chest Pain (786.5) | 433 | 32 | 102 | 23 |
| Pneumonia (486) | 206 | 51 | 42 | 25 |
– Example top order recommendations occurring within 24 hours of admission diagnosis of Chest Pain (ICD9: 786.5), sorted by odds ratio (OR). Additional metrics include prevalence (pretest probability), positive predictive value (post-test probability), and P-value by Fisher’s exact test. Binary labels are assigned if the order exists in pre-authored order sets (order set item) or clinical practice guidelines (guideline reference order).
| Item Description | Prevalence | PPV | OR | P-Fisher | Order Set/Guideline |
|---|---|---|---|---|---|
| POC Troponin I | 16.3% | 71.4% | 14.4 | 1.6E-148 | 1 / 1 |
| EKG 12-Lead | 51.8% | 92.8% | 12.7 | 8.0E-80 | 1 / 1 |
| Nitroglycerin (Sublingual) | 1.1% | 9.2% | 11.2 | 2.3E-25 | 0 / 1 |
| Consult Cardiology | 4.6% | 28.4% | 9.6 | 8.2E-64 | 1 / 1 |
| D - Dimer (ELISA) | 1.4% | 9.7% | 9.4 | 5.2E-24 | 1 / 0 |
| Aspirin (Oral) | 24.4% | 68.4% | 7.2 | 2.4E-85 | 1 / 1 |
| CK-MB | 15.7% | 51.3% | 6.1 | 1.5E-68 | 1 / 0 |
| Troponin I | 23.8% | 62.4% | 5.6 | 3.3E-67 | 1 / 1 |
| Clopidogrel (Oral) | 5.6% | 20.6% | 4.7 | 1.5E-27 | 1 / 0 |
| Cardiac Catheterization | 2.6% | 9.9% | 4.4 | 6.7E-14 | 1 / 1 |
| Heparin Activity Level | 6.1% | 18.9% | 3.8 | 1.7E-20 | 0 / 0 |
| Lipid Panel w/Direct LDL | 8.5% | 24.5% | 3.7 | 4.6E-24 | 1 / 0 |
| NT - proBNP | 10.0% | 26.3% | 3.4 | 6.4E-23 | 1 / 1 |
| Nitroglycerin (Topical) | 2.2% | 6.2% | 3.2 | 8.1E-07 | 1 / 1 |
| Dobutamine Stress Echo | 1.2% | 3.2% | 3.0 | 6.2E-04 | 1 / 1 |
Figure 1.– Receiver operating characteristic (ROC) curves for predicting clinical practice guideline reference orders based on automated recommender methods using different score-ranking options (PPV, OR, baseline prevalence, and presence in pre-authored order sets). Area-under-curve (AUC) reported as c-statistics with 95% confidence intervals empirically estimated by sampling items with replacement 1000 times.
Figure 2.– Recommender accuracy (precision or recall) for predicting guideline reference orders as a function of the number of top K recommendations considered (up to 100) when sorting by different score-ranking options (OR, PPV, prevalence, and presence in pre-authored order sets). Data labels added for K = 10 and n, where n = Number of items available in the respective order sets.