| Literature DB >> 33709067 |
Sujay Kulshrestha1,2, Dmitriy Dligach3,4,5, Cara Joyce3,4, Richard Gonzalez1,2, Ann P O'Rourke6, Joshua M Glazer7, Anne Stey8, Jacqueline M Kruser9, Matthew M Churpek9, Majid Afshar9.
Abstract
OBJECTIVE: Trauma quality improvement programs and registries improve care and outcomes for injured patients. Designated trauma centers calculate injury scores using dedicated trauma registrars; however, many injuries arrive at nontrauma centers, leaving a substantial amount of data uncaptured. We propose automated methods to identify severe chest injury using machine learning (ML) and natural language processing (NLP) methods from the electronic health record (EHR) for quality reporting.Entities:
Keywords: interpretability; machine learning; trauma surgery
Year: 2021 PMID: 33709067 PMCID: PMC7935500 DOI: 10.1093/jamiaopen/ooab015
Source DB: PubMed Journal: JAMIA Open ISSN: 2574-2531
Patient characteristics and outcomes between severe and nonsevere injury
| Nonsevere chest injury | Severe chest injury |
| |
|---|---|---|---|
|
| 8249 | 542 | |
| Age, median (IQR) | 47 (30–65) | 42 (27–60) | <.001 |
| Sex, | <.001 | ||
| Male | 5459 (66.2) | 406 (74.9) | |
| Female | 2790 (33.8) | 136 (25.1) | |
| Race, | .026 | ||
| White | 4683 (56.8) | 282 (52.0) | |
| Black | 1922 (23.3) | 153 (28.2) | |
| Other | 1644 (19.9) | 107 (19.7) | |
| Admitting service, | <.001 | ||
| Trauma | 3992 (48.3) | 483 (89.1) | |
| Burns | 1499 (18.2) | 29 (5.4) | |
| Orthopedic surgery | 834 (10.1) | 0 (0.0) | |
| Other | 1924 (23.3) | 30 (5.5) | |
| Operative intervention, | 2101 (25.5) | 194 (35.8) | <.001 |
| OR time (mins), median (IQR) | 185 (116–270) | 167 (104–274) | .31 |
| Comorbidities, | |||
| CHF | 213 (2.6) | 12 (2.2) | .70 |
| Hypertension | 1428 (17.3) | 111 (20.5) | .068 |
| Pulmonary disease | 468 (5.7) | 74 (18.2) | <.001 |
| Diabetes | 523 (6.3) | 44 (8.1) | .12 |
| Renal disease | 210 (2.5) | 19 (3.5) | .22 |
| Liver disease | 192 (2.3) | 25 (4.6) | .001 |
| Coagulopathy | 257 (3.1) | 57 (10.5) | <.001 |
| Alcohol misuse | 592 (7.2) | 65 (12.0) | <.001 |
| Drug misuse | 441 (5.3) | 55 (10.1) | <.001 |
| Elixhauser scores, median (IQR) | |||
| Readmission score | 8 (0–21) | 13 (4–22) | .002 |
| Mortality score | 0 (−1–10) | 4 (0–13) | <.001 |
| Length of stay, median (IQR) | 2.3 (0.7–5.9) | 5.4 (1.6–13.4) | <.001 |
| Disposition, | <.001 | ||
| Home | 5831 (70.7) | 246 (45.4) | |
| Discharge to HC facility | 1822 (22.1) | 146 (26.9) | |
| AMA | 163 (2.0) | 3 (0.6) | |
| In-hospital death | 342 (4.1) | 145 (26.8) | |
| Other | 91 (1.1) | 2 (0.4) |
AMA: against medical advice; HC: healthcare; IQR: interquartile range; OR: operating room.
Other Race = American Indian, Asian, Hispanic, Multiracial, Hawaiian, Pacific Islander, Unknown.
Other Disposition = Hospice, law enforcement, unknown.
Elixhauser Scores calculated using diagnosis codes from the entire encounter.
Figure 1.Classification plots comparing (A) CNN and EN and (B) CNN and XGB models. CNN model is indicated with the solid lines in each figure. TPR = true positive rate (grey); FPR = false positive rate (black); AUC = area under curve. X-axis represents threshold at which TPR/FPR are measured.
Concordance of model predictions across test dataset
| EN ( | XGB ( | CNN ( | ||||
|---|---|---|---|---|---|---|
| Model correct | 1435 | 81.6 | 1452 | 82.6 | 1528 | 86.9 |
| All models correct | 1357 | 77.2 | 1357 | 77.2 | 1357 | 77.2 |
| Model correct, 1 or both other models wrong | 78 | 4.4 | 95 | 5.4 | 171 | 9.7 |
| Positive case | 10 | 0.6 | 5 | 0.3 | 4 | 0.2 |
| Negative case | 68 | 3.9 | 90 | 5.1 | 167 | 9.5 |
Total number in holdout test dataset = 1758.
Figure 2.Top global model explanations from ML models to predict severe chest injury. X-axis represents rescaled variable importance from (1) EN beta coefficients, (2) XGB permuted feature importance, and (3) CNN training dataset averaged LIME explanations. Y-axis represents the preferred text definition of CUIs; CUI codes omitted here for clarity. CNN LIME interpretation median r2 = 0.69 (IQR 0.46–0.93).
Figure 3.Local model interpretations generated by LIME explainer for CNN model for false negative cases (A, B) and false positive cases (C, D). X-axis represents feature weight for top ten concept unique identifiers for each case. Probability represents probability of positive case; explanation fit represents r2 of LIME classifier for the selected case.