| Literature DB >> 28708848 |
Mehdi Jamei1, Aleksandr Nisnevich1, Everett Wetchler1, Sylvia Sudat2, Eric Liu1.
Abstract
Avoidable hospital readmissions not only contribute to the high costs of healthcare in the US, but also have an impact on the quality of care for patients. Large scale adoption of Electronic Health Records (EHR) has created the opportunity to proactively identify patients with high risk of hospital readmission, and apply effective interventions to mitigate that risk. To that end, in the past, numerous machine-learning models have been employed to predict the risk of 30-day hospital readmission. However, the need for an accurate and real-time predictive model, suitable for hospital setting applications still exists. Here, using data from more than 300,000 hospital stays in California from Sutter Health's EHR system, we built and tested an artificial neural network (NN) model based on Google's TensorFlow library. Through comparison with other traditional and non-traditional models, we demonstrated that neural networks are great candidates to capture the complexity and interdependency of various data fields in EHRs. LACE, the current industry standard, showed a precision (PPV) of 0.20 in identifying high-risk patients in our database. In contrast, our NN model yielded a PPV of 0.24, which is a 20% improvement over LACE. Additionally, we discussed the predictive power of Social Determinants of Health (SDoH) data, and presented a simple cost analysis to assist hospitalists in implementing helpful and cost-effective post-discharge interventions.Entities:
Mesh:
Year: 2017 PMID: 28708848 PMCID: PMC5510858 DOI: 10.1371/journal.pone.0181173
Source DB: PubMed Journal: PLoS One ISSN: 1932-6203 Impact factor: 3.240
summary of the population under study.
| Variable | All hospital visits | Visits resulting in 30-day readmission | Visits not resulting in 30-day readmission |
|---|---|---|---|
| Home | 93.0 | 91.8 | 93.1 |
| Outpatient | 0.1 | 0.1 | 0.1 |
| Transfer | 5.1 | 5.9 | 5.0 |
| Other | 1.8 | 2.1 | 1.7 |
| Elective | 27.3 | 11.7 | 29.0 |
| Emergency | 43.2 | 58.2 | 41.6 |
| Urgent | 28.2 | 29.5 | 28.0 |
| Other | 1.4 | 0.6 | 1.4 |
| 0–44 | 29.6 | 15.1 | 31.1 |
| 45–64 | 27.2 | 30.0 | 26.9 |
| 65–84 | 31.5 | 39.0 | 30.7 |
| 85+ | 11.7 | 15.9 | 11.3 |
| 28.3 | 25.3 | 28.6 | |
| 1.0 (4.0) | 4.0 (4.0) | 1.0 (3.0) | |
| Home or self care (routine) | 70.4 | 56.2 | 71.9 |
| Home under care of home health service organization | 15.0 | 22.3 | 14.2 |
| SNF | 14.6 | 21.5 | 13.9 |
| Morning (8:00 AM–12:59 PM) | 25.9 | 19.1 | 26.7 |
| Afternoon (1:00 PM–5:59 PM) | 61.4 | 65.8 | 60.9 |
| Evening (6:00 PM–7:59 AM) | 12.6 | 15.1 | 12.4 |
| 6.5 | 8.3 | 6.3 | |
| 61.9 | 54.6 | 62.7 | |
| 17.5 | 13.8 | 17.9 | |
| Commercial | 46.1 | 36.6 | 47.1 |
| Medicare | 51.5 | 62.2 | 50.3 |
| Self-pay | 2.2 | 1.0 | 2.3 |
| Other | 0.2 | 0.2 | 0.2 |
| 9.4 | 8.8 | 9.5 | |
| 6.0 (5.0) | 10.0 (5.0) | 6.0 (6.0) | |
| 3.0 (3.0) | 4.0 (5.0) | 3.0 (3.) | |
| Single | 27.2 | 28.6 | 27.0 |
| Married/partner | 48.2 | 39.6 | 49.1 |
| Divorced/separated | 8.9 | 11.4 | 8.6 |
| Widowed | 14.8 | 19.9 | 14.3 |
| Other/unknown | 0.9 | 0.4 | 0.9 |
| In the past 3 months | 0.3 (1.0) | 0.7 (1.6) | 0.3 (0.9) |
| In the past 6 months | 0.5 (1.5) | 1.1 (2.4) | 0.5 (1.4) |
| In the past 12 months | 0.8 (2.4) | 1.7 (3.9) | 0.7 (2.1) |
| In the past 3 months | 0.3 (0.7) | 0.8 (1.3) | 0.2 (0.6) |
| In the past 6 months | 0.4 (1.1) | 1.1 (2.0) | 0.3 (0.9) |
| In the past 12 months | 0.6 (1.5) | 1.6 (3.0) | 0.5 (1.2) |
| White | 61.9 | 61.9 | 61.9 |
| Black | 11.2 | 16.3 | 10.7 |
| Other | 25.9 | 21.2 | 26.4 |
| 25.5 (15.6) | 32.0 (15.8) | 24.7 (14.9) | |
| 12.4 | 15.2 | 12.1 |
Fig 1Total number of records for each hospital under study, and their respective readmission rates.
Fig 2Data breakdown by hospital admission year.
Summary of extracted feature categories, and two sample features per category.
| Category | Count | Sample features |
|---|---|---|
| 604 | abscess, kidney_stone | |
| 287 | hcup_category_cystic_fibro, hospital_problems_count | |
| 232 | px_blood_transf, px_c_section | |
| 202 | inp_num_unique _meds, outp_med_antidotes | |
| 119 | specialty_orthopedic_surgery, specialty_hospitalist_medical | |
| 46 | length_of_stay, disch_location_home_no_service | |
| 44 | pct_married, median_household_income | |
| 39 | admission_source_transfer, admission_type_elective | |
| 26 | num_abnormal_results, tabak_very_low_albumin | |
| 19 | charlson_index, comor_chf | |
| 16 | age, if_female | |
| 11 | alcohol_no, tobacco_quit | |
| 10 | pre_12_month_inpatient, pre_6_month_inpatient | |
| 8 | bmi, pulse | |
| 4 | insurance_type_medicare, insurance_type_self-pay |
Fig 3Neural Network model architecture (Note: Layer sizes are assuming all features are used).
Fig 4Comparison of NN model performance (with retrospective validation) vs number of features.
Top most correlated features with 30-day readmission.
| Category: Feature | Linear Correlation |
|---|---|
| 0.226 | |
| 0.224 | |
| 0.215 | |
| 0.210 | |
| 0.197 | |
| 0.160 | |
| 0.157 | |
| 0.149 | |
| 0.143 | |
| 0.143 |
Comparison of the performance of our models with that of LACE, assuming a 25% intervention rate.
| Model | # Features | Precision | Recall | AUC | Training time | Evaluation time |
|---|---|---|---|---|---|---|
| 1667 | 60% | 2650 sec | 154 sec | |||
| 500 | 22% | 0.77 | 396 | 31 | ||
| 100 | 22% | 58% | 0.76 | 169 | 14 | |
| 100 | 23% | 57% | 0.77 | 669 | 43 | |
| 1667 | 17% | 41% | 0.66 | 60 | 4 | |
| 100 | 21% | 52% | 0.72 | 17 | 0.1 | |
| 4 | 21% | 49% | 0.72 | 0 | 0.2 |
*—Model parameters: neural network (as described in Methods section), random forest (1000 trees of max depth 8, with 30% of features in each tree), logistic regression (default parameters in scikit-learn package)
**—Per-fold training time was measured on a 2014 Macbook Pro with a 4-core 2.2 GHz processor and 16GB RAM. The neural network model ran on four cores, while the other models could only be run on a single core. Training was performed on 259,050 records and evaluation was performed on 64,763 records.
***—We computed the AUC for LACE by comparing the performance of LACE models at every possible threshold. However, LACE is normally used with a fixed threshold, so the given AUC overstates the performance of LACE in practice.
Performance of our model versus LACE on 2015 data when trained on data through 2014.
| Model | # Features | Precision | Recall | AUC | Training time |
|---|---|---|---|---|---|
| all | 1040 sec | ||||
| 4 | 19% | 50% | 0.71 | 0 |
*—Model parameters: neural network (as described in Methods section), random forest (1000 trees of max depth 8, with 30% of features in each tree), logistic regression (default parameters in scikit-learn package)
**—Per-fold training time was measured on a 2014 Macbook Pro with a 4-core 2.2 GHz processor and 16GB RAM. The neural network model ran on four cores, while the other models could only be run on a single core. Training was performed on 259,050 records and evaluation was performed on 64,763 records.
***—We computed the AUC for LACE by comparing the performance of LACE models at every possible threshold. However, LACE is normally used with a fixed threshold, so the given AUC overstates the performance of LACE in practice.
Fig 5Comparison of artificial neural network model with LACE in 4 different age brackets.
Fig 6Comparison of the model performance among top five Sutter Health hospitals by the number of inpatient records.
Fig 7Comparison of the neural network model’s performance among subgroups with varying medical conditions.
Comparison of performance of each feature group on the neural network model, tested by withholding one feature group at a time and measuring the impact on model AUC.
| Feature Group | Effect on AUC |
|---|---|
| + 0.010 | |
| + 0.007 | |
| + 0.006 | |
| + 0.003 | |
| + 0.003 | |
| + 0.002 | |
| + 0.001 | |
| + 0.001 | |
| + 0.001 | |
| + 0.000 | |
| + 0.000 | |
| + 0.000 | |
| – 0.001 | |
| – 0.002 | |
| – 0.005 |
Fig 8The projected saving values as a function of the intervention rate, with the example parameters given for the cost-savings analysis in the results section.