Literature DB >> 35517246

Assessing the Generalizability of a Clinical Machine Learning Model Across Multiple Emergency Departments.

Alexander J Ryu1, Santiago Romero-Brufau2, Ray Qian3, Heather A Heaton4, David M Nestler4, Shant Ayanian1, Thomas C Kingsley1.   

Abstract

Objective: To assess the generalizability of a clinical machine learning algorithm across multiple emergency departments (EDs). Patients and
Methods: We obtained data on all ED visits at our health care system's largest ED from May 5, 2018, to December 31, 2019. We also obtained data from 3 satellite EDs and 1 distant-hub ED from May 1, 2018, to December 31, 2018. A gradient-boosted machine model was trained on pooled data from the included EDs. To prevent the effect of differing training set sizes, the data were randomly downsampled to match those of our smallest ED. A second model was trained on this downsampled, pooled data. The model's performance was compared using area under the receiver operating characteristic (AUC). Finally, site-specific models were trained and tested across all the sites, and the importance of features was examined to understand the reasons for differing generalizability.
Results: The training data sets contained 1918-64,161 ED visits. The AUC for the pooled model ranged from 0.84 to 0.94 across the sites; the performance decreased slightly when Ns were downsampled to match those of our smallest ED site. When site-specific models were trained and tested across all the sites, the AUCs ranged more widely from 0.71 to 0.93. Within a single ED site, the performance of the 5 site-specific models was most variable for our largest and smallest EDs. Finally, when the importance of features was examined, several features were common to all site-specific models; however, the weight of these features differed.
Conclusion: A machine learning model for predicting hospital admission from the ED will generalize fairly well within the health care system but will still have significant differences in AUC performance across sites because of site-specific factors.
© 2022 The Authors.

Entities:  

Keywords:  ANOVA, analysis of variance; AUC, area under the receiver-operator characteristic; ED, emergency department; ESI, emergency severity index; GBM, gradient-boosted machine; ML, machine learning

Year:  2022        PMID: 35517246      PMCID: PMC9062323          DOI: 10.1016/j.mayocpiqo.2022.03.003

Source DB:  PubMed          Journal:  Mayo Clin Proc Innov Qual Outcomes        ISSN: 2542-4548


Many health care systems are becoming interested in operationalizing machine learning (ML) predictive models to guide resource allocation., The use of ML to predict which emergency department (ED) patients may ultimately require hospital admission has the potential to reduce ED overcrowding and improve hospital efficiency., Existing literature has reported that regression analysis, and ML approaches7, 8, 9 can be suitable for this task, achieving excellent performance. Specifically, regression models, tree-based models, and neural network models performed well when structured data inputs were used, whereas neural networks displayed better performance when unstructured features, such as clinical notes, were included. Modern health care systems often include a network of hospitals, with lower-capacity community or critical-access hospitals serving more remote geographies while referring patients with complications to the network’s larger secondary or tertiary care centers. With these structures becoming increasingly common, it will be important to understand how an algorithm for predicting admissions generalizes across different sites in the health care system, particularly given the context of mounting concerns regarding the generalizability of health care ML algorithms. A better understanding of this issue will likely help guide model training and implementation for many health care systems looking to implement these types of algorithms. In particular, we hypothesized that issues with model generalizability could arise from site-specific or model-specific factors. Site-specific factors contributing to model generalizability could include differing patient characteristics or ED practice patterns, which may also be related to specific capabilities of a given ED. Model-specific factors may include random errors or differences in model performance relating to the volume of training data available, which may be of particular concern for small-volume, rural EDs. Here, we assessed the generalizability of a model that predicts the likelihood of hospital admission of ED patients. We elected to use a gradient-boosted machine (GBM) model for this task, given that we anticipated using structured data features for the ease of model implementation in practice. Neural network models were considered, but we suspected that performance gains would be minimal over GBM, given the lack of unstructured features, and there would be an increase in the complexity of model training and tuning. We hypothesized that a model training strategy that uses pooled data across ED sites with downsampling training data set sizes would lead to optimal model performance across the sites. Additionally, we suspected that a model trained at a referral center might generalize poorly to a small, rural ED and vice versa. Finally, we hypothesized that there would be some common clinical features that are found to be important across all predictive models.

Patients and Methods

Data Collection

We obtained data on all ED patients presenting to our health system’s largest ED (Rochester, Minnesota) from May 5, 2018, to December 31, 2019. We also obtained data on all ED patients from 3 regional satellite EDs of various sizes (Eau Claire, Wisconsin; Austin, Minnesota; and New Prague, Minnesota) and 1 distant-hub site (Phoenix, Arizona) from May 1, 2018, to December 31, 2018. Our study was deemed exempt and granted a waiver of consent by the Mayo Clinic Institutional Review Board (#21-006291). The data were stored in our organization’s electronic health record data warehouse and extracted using SAP Web Intelligence. Data on the hospital’s surrounding communities were collected from public US Census Bureau reports.

Feature Selection

We captured and engineered a total of 47 model features pertaining to real-time clinical information about the patients (age and initial vital signs), the chief symptom, any ED protocols activated, and the mode and timing of arrival at the ED. Chief symptom labels were coded by our ED and, thus, did not have issues with accommodating free text. Emergency department protocol activation refers to our ED’s use of several standardized documentation protocols for certain high-acuity situations such as ST-elevation myocardial infarction or trauma. We prioritized using features that would be available early in a patient’s ED course, likely within the first 15 minutes, to ensure that our predictions are maximally forward looking. Of note, all EDs included in our study shared a high degree of data harmonization because of pre-existing data infrastructure work. However, local practices—and, thus, patterns in the data—may differ from site to site. All features included in the model, along with data type, are listed in the Supplemental Appendix (available online at http://www.mcpiqojournal.org).

Pooled Model

We then used the XGBoost package with a random 70/15/15 train/validation/test split of all the sites’ data combined to train a model and subsequently test it against a 15% random test sample of each site’s data. The area under the receiver operating characteristic (AUC) and its SD, using jackknife estimation, were calculated for each model test. For the jackknife estimation of SD, we calculated the AUC for each of n test sets, where n corresponded to the size of a given test set, with each test set consisting of n−1 observations, with each observation being excluded once. We then calculated the SD of this AUC distribution. Analysis of variance (ANOVA) was performed to assess any significance of differences between the sites.

Pooled Model With Downsampled N

Because the number of visits sampled at each site differed by up to 30 times, we sought to remove this factor as a potential contributor to differing model performance by site. Therefore, each site’s data were randomly sampled down to the size of our smallest ED site. The train/validation/test split was repeated, and a second model was trained on this downsampled, pooled data and again tested across each site. Furthermore, ANOVA was repeated.

Site-Specific Models With Cross-Site Testing

To further understand the potential reasons for the differing model performance, we then trained a model on each ED’s downsampled data and tested each of those models across the sites to understand whether admission decisions were consistently easier or more difficult to predict at certain sites. Furthermore, ANOVA was performed for each of the 5 site-specific models to compare their AUC results across the sites.

Importance of Features

Finally, we examined the top 10 features that the model deemed most important across the 5 site-specific models and the pooled model to further understand how the model was making predictions at each site.

Results

The characteristics of the 5 EDs studied are summarized in Table 1, along with basic demographic information about their surrounding communities. Specifically, in the US component of our health care system, there are 3 major EDs, in Rochester, Minnesota; Phoenix, Arizona; and Jacksonville, Florida, with the Rochester site being the largest. Surrounding the Rochester site is a network of smaller hospitals ranging from secondary referral sites to critical-access hospitals. As described in Table 1, we attempted to capture 1 ED of each of our institutional size designations.
Table 1

Characteristics of the Different Emergency Department Sitesa

CharacteristicRochester, MinnesotaPhoenix, ArizonaEau Claire, WisconsinAustin, MinnesotaNew Prague, Minnesota
ED visits/y78,00044,00034,00018,0007,000
ED capabilities76 beds, level 1 trauma center, inpatient psychiatry, ED observation unit27 beds, level 4 trauma capabilities27 beds, level 2 trauma center17 beds, level 4 trauma center4 beds, level 4 trauma center, critical-access hospital
City population115,5571,680,99268,18725,1147,899
City % White82%66%91%93%97%
City median household income (2019)$73,106$57,459$55,477$48,127$77,949

ED, emergency department.

Characteristics of the Different Emergency Department Sitesa ED, emergency department. Table 2 presents the characteristics of patient samples from each site. For context, we included information about the fraction of patients admitted, ED visit duration, and emergency severity index (ESI; lower numbers indicate greater patient acuity) as well as the number of visits used for the training, validation, and testing of the model per site.
Table 2

Patient Sample Characteristics by Emergency Department Sitea

CharacteristicRochester, MinnesotaPhoenix, ArizonaEau Claire, WisconsinAustin, MinnesotaNew Prague, Minnesota
Training N64,16118,23312,5066,5581,919
Validation N12,3223,9072,6801,405411
Test N12,3223,9072,6801,405411
Patients admitted (%)35%49%27%18%21%
Median ED visit duration (h)4.43.73.32.83.0
Patients ESI 1-3 (%)80%88%82%67%75%
Mean patient age (SD)55.9±20.758.6±19.752.4±21.452.5±22.354.0±21.8

ED, emergency department; ESI, emergency severity index; SD, standard deviation.

Patient Sample Characteristics by Emergency Department Sitea ED, emergency department; ESI, emergency severity index; SD, standard deviation. Table 3 presents the results of AUC testing across the sites of models trained on pooled data before and after the downsampling of data set Ns. The downsampled models matched each site’s data to the same N as that of New Prague, Minnesota, before pooling the data. The AUC performance decreased slightly, and its SD increased at each site in which the downsampling of the data occurred. The same was true when the models on the pooled samples were tested. The differences in AUC by site were found to be significant, with P<.01, determined using ANOVA. New Prague, Minnesota, was not included in ANOVA for the second model because its AUC was identical to that of Austin, Minnesota.
Table 3

Performance of Pooled Model Before and After N Downsamplinga

Test sites listed to the rightRochester, MinnesotaPhoenix, ArizonaEau Claire, WisconsinAustin, MinnesotaNew Prague, MinnesotaAll sites combined
Pooled, not downsampledAUC ± SD0.89±.00002b0.84±.0001b0.94±.0001b0.92±.0003b0.86±.001b0.88±.00002
Pooled, downsampledAUC ± SD0.87±.0007c0.83±.0008c0.92±.0006c0.86±.0008c0.86±.0010.87±.0001

AUC, area under the receiver-operator characteristic; SD, standard deviation.

Indicates P<.001 by analysis of variance when tested across the means with matching symbols.

Indicates P<.001 by analysis of variance when tested across the means with matching symbols.

Performance of Pooled Model Before and After N Downsamplinga AUC, area under the receiver-operator characteristic; SD, standard deviation. Indicates P<.001 by analysis of variance when tested across the means with matching symbols. Indicates P<.001 by analysis of variance when tested across the means with matching symbols. Table 4 presents the results of the testing of site-specific models, each trained using training sets of downsampled Ns, across all the sites and pooled test data. All AUCs were significantly different from each other. When the models were tested at the same site at which they were trained, the performance was highest at Eau Claire, Wisconsin (0.93), and lowest at New Prague, Minnesota (0.79), suggesting that the features included in our model were particularly well suited to patients and clinical practices in Eau Claire, Wisconsin, and less so in New Prague, Minnesota. For 4 of the 5 site-specific models, the AUC performance was the best at Eau Claire, Wisconsin. Interestingly, the AUC was higher in Eau Claire, Wisconsin, than that at “home” sites where the models were trained. The lowest AUC was noted for a model trained at our tertiary referral center, Rochester, Minnesota, and was tested at our critical-access hospital, New Prague, Minnesota. A model trained on data from New Prague, Minnesota, similarly performed poorly when tested on data from Rochester, Minnesota, suggesting that the patients and practices at these sites were most dissimilar.
Table 4

Performance of Site-specific Models With Downsampled Nsa

Training sites listed below; test sites listed to the rightRochester, MinnesotaPhoenix, ArizonaEau Claire, WisconsinAustin, MinnesotaNew Prague, MinnesotaAll sites combined
Rochester, MinnesotaAUC ± SD0.85±.0009b0.78±.001b0.89±.0007b0.84±.0009b0.71±.001b0.84±.0002
Phoenix, ArizonaAUC ± SD0.77±.0009c0.81±.0008c0.90±.0007c0.85±.0008c0.81±.0010.82±.0002
Eau Claire, WisconsinAUC ± SD0.79±.0009d0.81±.0009d0.93±.0005d0.88±.0007d0.75±.001d0.84±.0002
Austin, MinnesotaAUC ± SD0.75±.0009e0.79±.0009e0.91±.0006e0.88±.0007e0.79±.0010.83±.0002
New Prague, MinnesotaAUC ± SD0.74±.0009f0.8±.0009f0.86±.0007f0.88±.0008f0.82±.001f0.83±.0002

AUC, area under the receiver-operator characteristic; SD, standard deviation.

Indicates P<0.01 by analysis of variance when tested across the means with matching symbols.

Indicates P<0.01 by analysis of variance when tested across the means with matching symbols.

Indicates P<0.01 by analysis of variance when tested across the means with matching symbols.

Indicates P<0.01 by analysis of variance when tested across the means with matching symbols.

Indicates P<0.01 by analysis of variance when tested across the means with matching symbols.

Performance of Site-specific Models With Downsampled Nsa AUC, area under the receiver-operator characteristic; SD, standard deviation. Indicates P<0.01 by analysis of variance when tested across the means with matching symbols. Indicates P<0.01 by analysis of variance when tested across the means with matching symbols. Indicates P<0.01 by analysis of variance when tested across the means with matching symbols. Indicates P<0.01 by analysis of variance when tested across the means with matching symbols. Indicates P<0.01 by analysis of variance when tested across the means with matching symbols. Table 5 presents the top 10 important features for each model, along with the associated feature weight. Specifically, we reported the average gain across all splits in which a feature was used. There appeared to be substantial overlap in features that the model determined to be of high importance, but the weights of these features varied substantially across the sites. The ESI score, arrival by an ambulance, patient’s age, and having had an electrocardiogram were among the top 10 features at all the sites. In the pooled model, having been a patient in Eau Claire, Wisconsin, was the only site-specific indicator (there was 1 indicator feature per site) among the top 10 features.
Table 5

Top 10 Important Features for Each Modela

Data aggregationRochester, MinnesotaPhoenix, ArizonaEau Claire, WisconsinAustin, MinnesotaNew Prague, MinnesotaPooled model
Feature 1 (weight)ESI321 (171)ESI (27)Weight (108)Had EKG? (56)Had EKG? (87)Had EKG? (141)
Feature 2 (weight)Ambulance (115)Had EKG? (13)Had EKG? (53)ESI (26)Ambulance (35)ESI321 (56)
Feature 3 (weight)Weight (65)Fever (12)Wheelchair (25)Oxygen saturation (15)ESI (33)ESI (52)
Feature 4 (weight)ESI (25)Altered mental status (10)ESI (25)Age (14)Wheelchair (25)Weight (43)
Feature 5 (weight)Had EKG? (19)Wheelchair (7)Ambulance (23)Suicidal (11)Oxygen saturation (22)Ambulance (38)
Feature 6 (weight)Wheelchair (17)Age (7)Suicidal (22)Weight (11)Temperature (21)Wheelchair (25)
Feature 7 (weight)Chest pain (17)Chest pain (5)Age (20)Ambulance (10)Age (16)Eau Claire (23)
Feature 8 (weight)From outside hospital (14)Ambulance (5)From outside hospital (18)Pulse (8)Weight (14)Abdominal pain (20)
Feature 9 (weight)Age (13)Temperature (5)Respiratory rate (16)Respiratory rate (7)Diastolic blood pressure (13)Age (20)
Feature 10 (weight)Resuscitation status (12)Weakness (5)Diastolic blood pressure (10)Abdominal pain (6)Respiratory rate (13)Chest pain (18)

EKG, electrocardiogram; ESI, emergency severity index.

Top 10 Important Features for Each Modela EKG, electrocardiogram; ESI, emergency severity index.

Discussion

This article presents a strategy for assessing the generalizability of a clinical artificial intelligence model across multiple ED sites. Although previous literature has reported the suitability of GBMs for similar clinical prediction tasks, few, if any, studies have proposed methods for investigating the generalizability of these models across multiple health care sites.7, 8, 9, We used GBMs to train a model predicting the likelihood of a patient’s admission from the ED early in their ED course. We trained a model on pooled data from 5 diverse ED sites and tested the model’s performance at each of the sites, noting high, but different, AUC performance across the sites. We then downsampled the data set Ns across the sites to correct for any effect of site overrepresentation on differing AUCs and found small but significant decreases in the model’s performance. This investigation revealed significant differences in the AUC performance when examined using ANOVA, even after downsampling the data set Ns; the AUCs for our downsampled, pooled model ranged from 0.83 to 0.92 when tested across the sites, which is likely attributable to a combination of patient-specific and clinical practice (ie, site-specific) factors. We also found that after the downsampling of Ns, the AUC performance decreased by 0.01-0.06 per site. Importantly, the decrease in the model performance by site was not proportional to the degree of downsampling that occurred. Additionally, the pooled, downsampled model performed better than site-specific models at all the sites when compared with how these performed at their home sites, with the exception of Eau Claire, Wisconsin. This suggests that patient data from the other sites aided the pooled model in detecting patterns that were generalizable, thus raising the performance for all the sites. Overall, this suggests that training a model on site-pooled data, over as long a time period as possible, will lead to the best model performance; however, even with significant downsampling, from pooled training N=103,377 to downsampled N=9595, the performance decreases only modestly. We then trained site-specific models using the downsampled data sets for each of our 5 ED sites and tested those across all 5 ED sites. This exercise suggested that at 1 site in particular (Eau Claire, Wisconsin), admission decisions were easier to predict than at any other site, which is likely because of a combination of patient factors and practice patterns, although the reasons for this were not explicitly captured in our study. Additionally, this suggested that a model trained at a tertiary referral ED would perform particularly poorly at a critical-access ED and vice versa. This seems like an expected finding, given the differences in patient complexity and ED capabilities. Upon examining the top 10 important features at each site, we noted that there was significant overlap of all the models, including the ESI score, arrival by an ambulance, patient age, and having had an electrocardiogram. This suggests that these factors are indeed highly relevant clinical features for our prediction, rather than site-specific outliers. The weight of the top 10 features, however, differed significantly between the site-specific models. Comparing our results with those of the existing literature, our model’s AUC results fit into the higher end of reported AUCs, supporting the suitability of GBMs for this task.7, 8, 9, There also appear to be limited studies that examined model generalizability, particularly using ML models. One study examined the generalizability of a logistic regression model across disparate hospitals and retrained the model for each hospital. Other similar studies focused on a single ED,,,,15, 16, 17, 18, 19 used curated survey data,, or grouped multiple EDs together for model training. Our study was limited in that only 1 health care system’s EDs were represented, although the EDs were substantially variable in size and geography. All the EDs included shared a high degree of data harmonization, which facilitated model testing. Finally, our algorithm generated predictions at only 1 time point in the patients’ ED course, thereby neglecting situations in which major changes occurring in a patient’s ED course significantly alter their probability of admission. With respect to model deployment, our health care system’s largest ED is currently piloting an initiative in which hospital medicine physicians collaborate with emergency medicine physicians to help facilitate patient triage. The leadership of this initiative is interested in using this model to more quickly identify patients for interventions to expedite both admission and discharge. For other health care systems looking to implement similar models, our data appear to suggest that the best strategy for achieving high performance across multiple sites is to train a model on pooled data, without downsampling, that contain features identifying which patients came from which site, as was done in our study. This approach contributes to not only a high-performing model but also the ease of maintenance, with only 1 model to oversee, to which additional training data can be easily added over time.

Conclusion

Overall, this study provides a strategy for systematically assessing the generalizability of a clinical ML model across multiple ED sites and provides the estimates of AUC differences across multiple scenarios. We determined that optimal GBM model performance is achieved when trained on multisite, pooled data; however, even with this strategy, the model will perform differently at some sites, which was not explained by the random errors or differences in the training set N. Instead, the differences in the model performance were likely due to site-specific factors. When the importance of the features was examined for our GBM models, the ESI score, arrival by an ambulance, patient’s age, and having had an electrocardiogram appeared to be important clinical predictors across all the sites, although the importance of the features differed.
  4 in total

1.  Prediction of admission in pediatric emergency department with deep neural networks and triage textual data.

Authors:  Bruno P Roquette; Hitoshi Nagano; Ernesto C Marujo; Alexandre C Maiorano
Journal:  Neural Netw       Date:  2020-03-18

2.  Early prediction of hospital admission for emergency department patients: a comparison between patients younger or older than 70 years.

Authors:  Jacinta A Lucke; Jelle de Gelder; Fleur Clarijs; Christian Heringhaus; Anton J M de Craen; Anne J Fogteloo; Gerard J Blauw; Bas de Groot; Simon P Mooijaart
Journal:  Emerg Med J       Date:  2017-08-16       Impact factor: 2.740

3.  Prediction of Emergency Department Hospital Admission Based on Natural Language Processing and Neural Networks.

Authors:  Xingyu Zhang; Joyce Kim; Rachel E Patzer; Stephen R Pitts; Aaron Patzer; Justin D Schrager
Journal:  Methods Inf Med       Date:  2017-08-16       Impact factor: 2.176

4.  Text mining approach to predict hospital admissions using early medical records from the emergency department.

Authors:  Filipe R Lucini; Flavio S Fogliatto; Giovani J C da Silveira; Jeruza L Neyeloff; Michel J Anzanello; Ricardo S Kuchenbecker; Beatriz D Schaan
Journal:  Int J Med Inform       Date:  2017-01-05       Impact factor: 4.046

  4 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.