Literature DB >> 34448870

Development and Assessment of an Interpretable Machine Learning Triage Tool for Estimating Mortality After Emergency Admissions.

Feng Xie1, Marcus Eng Hock Ong1,2, Johannes Nathaniel Min Hui Liew1, Kenneth Boon Kiat Tan2, Andrew Fu Wah Ho1,2, Gayathri Devi Nadarajan2, Lian Leng Low1,3, Yu Heng Kwan1,4, Benjamin Alan Goldstein1,5, David Bruce Matchar1,6, Bibhas Chakraborty1,5,7, Nan Liu1,8,9.   

Abstract

Importance: Triage in the emergency department (ED) is a complex clinical judgment based on the tacit understanding of the patient's likelihood of survival, availability of medical resources, and local practices. Although a scoring tool could be valuable in risk stratification, currently available scores have demonstrated limitations.
Objectives: To develop an interpretable machine learning tool based on a parsimonious list of variables available at ED triage; provide a simple, early, and accurate estimate of patients' risk of death; and evaluate the tool's predictive accuracy compared with several established clinical scores. Design, Setting, and Participants: This single-site, retrospective cohort study assessed all ED patients between January 1, 2009, and December 31, 2016, who were subsequently admitted to a tertiary hospital in Singapore. The Score for Emergency Risk Prediction (SERP) tool was derived using a machine learning framework. To estimate mortality outcomes after emergency admissions, SERP was compared with several triage systems, including Patient Acuity Category Scale, Modified Early Warning Score, National Early Warning Score, Cardiac Arrest Risk Triage, Rapid Acute Physiology Score, and Rapid Emergency Medicine Score. The initial analyses were completed in October 2020, and additional analyses were conducted in May 2021. Main Outcomes and Measures: Three SERP scores, namely SERP-2d, SERP-7d, and SERP-30d, were developed using the primary outcomes of interest of 2-, 7-, and 30-day mortality, respectively. Secondary outcomes included 3-day mortality and inpatient mortality. The SERP's predictive power was measured using the area under the curve in the receiver operating characteristic analysis.
Results: The study included 224 666 ED episodes in the model training cohort (mean [SD] patient age, 63.60 [16.90] years; 113 426 [50.5%] female), 56 167 episodes in the validation cohort (mean [SD] patient age, 63.58 [16.87] years; 28 427 [50.6%] female), and 42 676 episodes in the testing cohort (mean [SD] patient age, 64.85 [16.80] years; 21 556 [50.5%] female). The mortality rates in the training cohort were 0.8% at 2 days, 2.2% at 7 days, and 5.9% at 30 days. In the testing cohort, the areas under the curve of SERP-30d were 0.821 (95% CI, 0.796-0.847) for 2-day mortality, 0.826 (95% CI, 0.811-0.841) for 7-day mortality, and 0.823 (95% CI, 0.814-0.832) for 30-day mortality and outperformed several benchmark scores. Conclusions and Relevance: In this retrospective cohort study, SERP had better prediction performance than existing triage scores while maintaining easy implementation and ease of ascertainment in the ED. It has the potential to be widely applied and validated in different circumstances and health care settings.

Entities:  

Mesh:

Year:  2021        PMID: 34448870      PMCID: PMC8397930          DOI: 10.1001/jamanetworkopen.2021.18467

Source DB:  PubMed          Journal:  JAMA Netw Open        ISSN: 2574-3805


Introduction

Triage in the emergency department (ED) for admission and appropriate level of hospital care is a complex clinical judgment based on the tacit understanding of the patient’s likely short-term course, availability of medical resources, and local practices.[1,2] Besides triage categories, early warning scores are also used to identify patients at risk of having adverse events. One such example is the Cardiac Risk Assessment Triage (CART) score,[3] which calculates a score based on a patient’s vital signs, indicating their risk for cardiac arrest, subsequent transfer to the intensive care unit, and mortality.[4] To date, few studies[5,6,7,8] have investigated variables of short-term and long-term mortality among the general ED population, using the limited data available at the point of triage. Most ED-specific scores are targeted toward specific conditions, such as the quick Sepsis-Related Organ Failure Assessment for infection and sepsis,[5,6] CART for cardiac conditions, or Predicting Mortality in the Emergency Department for elderly populations.[7,8] Several general purpose scores have been adopted by the ED, such as the Modified Early Warning Score (MEWS) and Acute Physiology and Chronic Health Evaluation (APACHE) II score. However, MEWS has only moderate predictive capabilities, with an area under the curve (AUC) of 0.71,[9] and APACHE II requires laboratory variables unavailable at the point of triage.[10] In the fast-paced ED environment, a scoring tool needs to be accurate and straightforward. To address the need for a risk tool appropriate to the ED workflow, we developed the Score for Emergency Risk Prediction (SERP) using a general-purpose machine learning–based scoring framework named AutoScore.[11] The resulting tool was compared in a test set to the current triage system used in Singapore, the Patient Acuity Category Scale (PACS),[12] and several established early warning or triage scores.

Methods

Study Design and Setting

We performed a retrospective cohort study of patients seen in the ED of Singapore General Hospital (SGH). Singapore is a city-state in Southeast Asia with a rapidly aging society[13]; currently, approximately 1 in 5 Singaporeans are 60 years or older.[14] The SGH is the largest and oldest public tertiary hospital in Singapore. The SGH ED receives more than 120 000 visits and has 36 000 inpatient admissions annually. The electronic health record (EHR) data were obtained from Singapore Health Services and analyzed. This study was approved by Singapore Health Services’ Centralized Institutional Review Board, and a waiver of consent was granted for EHR data collection and analysis because of the retrospective nature of the study. All data were deidentified. This study followed the Strengthening the Reporting of Observational Studies in Epidemiology (STROBE) reporting guideline.[15]

Study Population

All patients visiting the SGH ED from January 1, 2009, until December 31, 2016, who were subsequently admitted, were included. We denote these included episodes as emergency admissions. Patients younger than 21 years were excluded. We also excluded noncitizen patients who might not have complete medical records. These admission episodes from January 1, 2009, to December 31, 2015, were randomly split into 2 nonoverlapping cohorts: a training cohort (80%) and a validation cohort (20%). The admission episodes in 2016 were assigned to the testing cohort. This sequential testing design was chosen to be more consistent with future application scenarios and evaluate whether population shift would influence the model’s performance.

Outcome

The primary outcomes used to develop and test the tool were 2-, 7-, and 30-day mortality, defined as deaths within 2, 7, and 30 days after emergency admission, respectively. Three SERP scores, namely SERP-2d, SERP-7d, and SERP-30d, were developed using the corresponding primary outcome. We also tested the performance of those clinical scores on the secondary outcomes, including inpatient mortality, defined as deaths in the hospital, and 3-day mortality, defined as deaths within 72 hours after the time of admission. Death records were obtained from the national death registry and were matched to specific patients in the EHR.

Data Collection and Candidate Variables

We extracted data from the hospital’s EHR through the SingHealth Electronic Health Intelligence System. Patient details were deidentified, complying with Health Insurance Portability and Accountability Act regulations. Comorbidities were obtained from hospital diagnosis and discharge records in the preceding 5 years before patients’ index emergency admissions. They were extracted from the International Classification of Diseases, Ninth Revision (ICD-9) and International Statistical Classification of Diseases and Related Health Problems, Tenth Revision (ICD-10),[16] globally used diagnostic tools for epidemiology and clinical purposes. We preselected candidate variables available in the ED before hospital admission to ensure that SERP was clinically useful and valid for early risk stratification of patients in the ED. Candidate variables included demographic characteristics, administrative variables, medical history in the preceding year, vital signs, and comorbidities. The list of candidate variables is given in the eTable 1 in the Supplement. Comorbidity variables were defined according to the Charlson Comorbidity Index. We used the algorithms developed and updated by Quan et al[17] for the linkage between Charlson Comorbidity Index and ICD-9/ICD-10 codes.

Statistical Analysis

The data were analyzed using R software, version 3.5.3 (R Foundation for Statistical Computing). The initial analyses were completed in October 2020, and additional analyses were conducted in May 2021. Baseline characteristics of the study population were analyzed on all 3 cohorts (training, validation, and testing). In the descriptive summaries, numbers (percentages) were reported for categorical variables. For continuous variables, means (SDs) were reported. During the analysis, the value for a vital sign would be considered as an outlier and set to missing if it were beyond the plausible physiologic ranges based on clinical knowledge. For example, any value of vital signs below 0, heart rate above 300/min, respiration rate above 50/min, systolic blood pressure above 300 mm Hg, diastolic blood pressure above 180 mm Hg, or oxygen saturation as measured by pulse oximetry above 100% was deemed an outlier. Subsequently, all missing values were imputed using the median value of the training cohort. We implemented the AutoScore,[11] a machine learning–based clinical score generation algorithm, to derive the SERP scoring models. AutoScore combines machine learning and logistic regression, integrates multiple modules of data manipulation, and automates the development of parsimonious sparse-score risk models for predefined outcomes. In addition, it enables users to build interpretable clinical scores quickly and seamlessly, which can be easily implemented and validated in clinical practice. The training cohort was used for the generation of the tentative SERP models using AutoScore framework. The validation cohort was used to evaluate multiple candidate SERP models for parameter tuning and model selection. Then, we calculated the performance metrics of the final SERP model based on the testing cohort. Finally, we used the primary outcomes for model derivation and applied primary and secondary outcomes for model evaluation. The implementation details and methodologic descriptions are elaborated in eFigure 1 and the eMethods in the Supplement. After model derivation, the predictive performance of the final SERP scores was reported based on the testing cohort, where bootstrapped samples were applied to calculate 95% CIs. Each of the SERP breakdowns was allocated a score that reflected the magnitude of disturbance to each variable. The individual scores were then summed to derive the aggregated SERP score for risk stratification of outcomes. The predictive power of SERP was measured using the AUC in the receiver operating characteristic (ROC) analysis. Sensitivity, specificity, positive predictive value, and negative predictive value were calculated under the optimal threshold, defined as the point nearest to the upper-left corner of the ROC curve. The metrics calculated under different thresholds were also compared to evaluate predictive performance. By using the same testing cohort, we compared the 3 SERP scores with PACS, MEWS,[18] the National Early Warning Score (NEWS),[19] CART,[3] the Rapid Acute Physiology Score (RAPS),[20] and the Rapid Emergency Medicine Score (REMS),[21] in estimating multiple mortality outcomes in this study.

Results

Baseline Characteristics of the Study Cohort

Between January 1, 2009, and December 31, 2015, a total of 280 833 individual admission episodes were assessed, including 224 666 in the training cohort (mean [SD] patient age, 63.60 [16.90] years; 113 426 [50.5%] female) and 56 167 in the validation cohort (mean [SD] patient age, 63.58 [16.87] years; 28 427 [50.6%] female). In addition, 42 676 admission episodes in the year 2016 were included in the testing cohort (mean [SD] patient age, 64.85 [16.80] years; 21 556 [50.5%] female) (Figure 1). The mortality rates observed in the training cohort were 0.8% at 2 days, 2.2% at 7 days, and 5.9% at 30 days. The ethnic compositions were similar to the population norm (74.3% for Chinese, 12.9% for Malay, 10.0% for Indian, and 2.8% for others). A total of 39 548 episodes (17.6%) were triaged as PACS 1, and 128 644 episodes (57.3%) were triaged as PACS 2. Table 1 indicates that patient characteristics in the training and validation cohorts were similar in terms of age, sex, racial and ethnic compositions, and other characteristics. Compared with those in the training and validation cohorts, patients in the testing cohort were slightly older and had a higher risk of triage to PACS 1, with more people having comorbidities of dementia, diabetes, and kidney diseases. The patients in the testing cohort also had marginally lower mortality rates and higher numbers of emergency admissions or operations in the past year. This difference likely reflects the population shift and improvements in health care over time.
Figure 1.

Flow of the Study Cohort Formation

Table 1.

Baseline Characteristics of the Study Cohort

CharacteristicTraining (n = 224 666)Validation (n = 56 167)Testing (n = 42 676)
Age, mean (SD), y63.60 (16.90)63.58 (16.87)64.85 (16.80)
Sex
Male111 240 (49.5)27 740 (49.4)21 120 (49.5)
Female113 426 (50.5)28 427 (50.6)21 556 (50.5)
Race/ethnicity
Chinese167 004 (74.3)41 765 (74.4)31 441 (73.7)
Indian22 403 (10.0)5592 (10.0)4440 (10.4)
Malay29 040 (12.9)7213 (12.8)5465 (12.8)
Other6219 (2.8)1597 (2.8)1330 (3.1)
PACS triage categories
P139 548 (17.6)9823 (17.5)9913 (23.2)
P2128 644 (57.3)32 058 (57.1)22 885 (53.6)
P3 and P456 474 (25.1)14 286 (25.4)9878 (23.1)
Shift time
8 am to 4 pm113 758 (50.6)28 461 (50.7)21 870 (51.2)
4 pm to midnight84 503 (37.6)21 050 (37.5)15 907 (37.3)
Midnight to 8 am26 405 (11.8)6656 (11.9)4899 (11.5)
Day of week
Friday31 553 (14.0)7893 (14.1)5839 (13.7)
Monday37 703 (16.8)9581 (17.1)7139 (16.7)
Weekend57 785 (25.7)14 283 (25.4)10 901 (25.5)
Midweek97 625 (43.5)24 410 (43.5)18 797 (44.0)
Vital signs, mean (SD)
Pulse, /min81.57 (16.41)81.62 (16.37)85.95 (18.36)
Respiration, /min17.80 (1.57)17.80 (1.59)18.23 (2.04)
Spo2, %98.12 (2.84)98.12 (2.70)97.34 (4.18)
Blood pressure, mm Hg
Diastolic70.99 (13.23)71.01 (13.20)72.36 (13.95)
Systolic133.77 (24.49)133.80 (24.58)137.73 (27.87)
Comorbidities
Myocardial infarction14 927 (6.6)3801 (6.8)2841 (6.7)
Congestive heart failure28 511 (12.7)7136 (12.7)4897 (11.5)
Peripheral vascular disease14 531 (6.5)3539 (6.3)2541 (6.0)
Stroke32 993 (14.7)8062 (14.4)5062 (11.9)
Dementia6901 (3.1)1699 (3.0)1515 (3.6)
Chronic pulmonary disease24 275 (10.8)6138 (10.9)3912 (9.2)
Rheumatoid disease3341 (1.5)881 (1.6)615 (1.4)
Peptic ulcer disease9879 (4.4)2505 (4.5)1362 (3.2)
Diabetes
None145 889 (64.9)36 457 (64.9)27 204 (63.7)
Diabetes without chronic complications24 268 (10.8)6064 (10.8)1247 (2.9)
Diabetes with complications54 509 (24.3)13 646 (24.3)14 225 (33.3)
Hemiplegia or paraplegia14 545 (6.5)3609 (6.4)1880 (4.4)
Kidney disease49 884 (22.2)12 483 (22.2)10 377 (24.3)
Cancer
None185 121 (82.4)46 251 (82.3)35 374 (82.9)
Local tumor, leukemia, and lymphoma20 838 (9.3)5136 (9.1)3613 (8.5)
Metastatic solid tumor18 707 (8.3)4780 (8.5)3689 (8.6)
Liver disease
None209 865 (93.4)52 562 (93.6)39 704 (93.0)
Mild liver disease11 112 (4.9)2676 (4.8)2156 (5.1)
Severe liver disease3689 (1.6)929 (1.7)816 (1.9)
Health care use, mean (SD)
Emergency admissions in the past year1.05 (2.35)1.05 (2.35)1.12 (2.51)
Operations in the past year0.20 (0.72)0.20 (0.72)0.28 (0.94)
ICU admissions in the past year0.03 (0.26)0.02 (0.26)0.03 (0.29)
HD admissions in the past year0.10 (0.51)0.10 (0.51)0.08 (0.45)
Mortality-related outcomes
2 d1801 (0.8)449 (0.8)295 (0.7)
3 d2464 (1.1)622 (1.1)416 (1.0)
7 d4888 (2.2)1241 (2.2)779 (1.8)
14 d8040 (3.6)2009 (3.6)1349 (3.2)
Inpatient8616 (3.8)2151 (3.8)1515 (3.6)
30 d13 244 (5.9)3285 (5.8)2310 (5.4)

Abbreviations: HD, high-dependency; ICU, intensive care unit; PACS, Patient Acuity Category Scale; Spo2, oxygen saturation as measured by pulse oximetry.

Data are presented as number (percentage) of patients unless otherwise indicated.

Abbreviations: HD, high-dependency; ICU, intensive care unit; PACS, Patient Acuity Category Scale; Spo2, oxygen saturation as measured by pulse oximetry. Data are presented as number (percentage) of patients unless otherwise indicated.

Selected Variables and SERP Score

AutoScore was used to select the most discriminative variables from all 26 candidate variables (eTable 1 in the Supplement). Parsimony plots (ie, model performance vs complexity) based on the validation set were used for determining the choice of variables (eFigure 2 in the Supplement). We chose 6 variables as the parsimonious choice for SERP-2d and SERP-30d, whereas SERP-7d with 5 variables achieved a good balance in the parsimony plot. Five variables were chosen by all 3 SERP scores, including age, heart rate, respiration rate, diastolic blood pressure, and systolic blood pressure. These selected variables highlighted the importance of vital signs in risk-triaging patients in emergency settings. As seen from eFigure 2 in the Supplement, when more variables were added to the scoring model, the performance was not markedly improved. The SERP scores derived based on primary outcomes (2-, 7-, and 30-day mortality) were tabulated in Table 2. All 3 scores summed from their included variables ranged from 0 to approximately 60. We used the testing cohort to evaluate the performance of the SERP scores. eFigure 3 in the Supplement depicts the distribution of episodes at different score intervals, which had near-normal distribution. For SERP-2d and SERP-30d, most patients had a risk score between 16 and 24, and few patients had scores under 9 or above 40. As seen in eFigure 4 in the Supplement, the observed mortality rate increased as our risk scores increased in the testing cohort. In terms of different components of SERP, when age was younger than 30 years, its corresponding risk (quantified as points) was the lowest; when age was older than 80 years, the risk was the highest. Likewise, when a reported diastolic blood pressure was between 50 and 94 mm Hg, the corresponding risk was the lowest, and when it was lower than 49 mm Hg, the risk was the highest. Thus, SERP scores had varying points for each component according to the outcomes of interest.
Table 2.

Three Versions of the SERP Derived From the Primary Outcomes

VariableSERP scores
SERP-2dSERP-7dSERP-30d
Age, y
<30000
30-499108
50-79131714
≥80172119
Heart rate, /min
<60321
60-69000
70-94342
95-109686
≥11010129
Respiration rate, /min
<1611108
16-19000
≥20766
Blood pressure, mm Hg
Systolic
<10010128
100-114465
115-149112
≥150000
Diastolic
<50543
50-94000
≥95122
Spo2, %
<907NANA
90-945NANA
≥950NANA
Cancer history
NoneNANA0
Local tumor, leukemia, and lymphomaNANA6
Metastatic solid tumorNANA14

Abbreviations: NA, not applicable; SERP, Score for Emergency Risk Prediction; Spo2, oxygen saturation as measured by pulse oximetry.

Abbreviations: NA, not applicable; SERP, Score for Emergency Risk Prediction; Spo2, oxygen saturation as measured by pulse oximetry.

Performance Evaluation

The performance of the SERP scores and other clinical scores as assessed by ROC analysis in the testing cohort are reported in Table 3 and Figure 2. SERP had promising discriminatory capability in estimating all mortality-related outcomes. The SERP-30d achieved the best performance for short-term and long-term mortality prediction, with an AUC of 0.821 (95% CI, 0.796-0.847) for 2-day mortality, an AUC of 0.826 (95% CI, 0.811-0.841) for 7-day mortality, an AUC of 0.823 (95% CI, 0.814-0.832) for 30-day mortality, and an AUC of 0.810 (95% CI, 0.799-0.821) for inpatient mortality. eTables 2 and 3 in the Supplement summarize the predictive performance of the SERP scores and their comparators on 30- and 2-day mortality risk estimation, respectively.
Table 3.

Comparison of AUC Values Achieved by Different Triage Scores on the Testing Cohort

ScoreAUC value (95% CI) by mortality
2 d3 d7 dInpatient30 d
SERP-2d0.821 (0.796-0.847)0.815 (0.793-0.837)0.798 (0.781-0.814)0.769 (0.757-0.781)0.754 (0.744-0.765)
SERP-7d0.810 (0.783-0.837)0.805 (0.783-0.828)0.793 (0.776-0.809)0.765 (0.753-0.777)0.754 (0.744-0.764)
SERP-30d0.821 (0.796-0.847)0.824 (0.804-0.845)0.826 (0.811-0.841)0.810 (0.799-0.821)0.823 (0.814-0.832)
CART0.779 (0.751-0.807)0.769 (0.745-0.793)0.738 (0.720-0.756)0.704 (0.691-0.717)0.700 (0.689-0.711)
PACS0.796 (0.775-0.817)0.778 (0.758-0.797)0.750 (0.735-0.765)0.703 (0.691-0.715)0.680 (0.670-0.690)
MEWS0.763 (0.734-0.792)0.750 (0.725-0.774)0.721 (0.702-0.739)0.680 (0.667-0.694)0.663 (0.652-0.674)
NEWS0.803 (0.774-0.832)0.792 (0.767-0.817)0.773 (0.755-0.791)0.734 (0.720-0.747)0.711 (0.700-0.723)
RAPS0.683 (0.652-0.715)0.674 (0.647-0.700)0.633 (0.613-0.653)0.594 (0.580-0.608)0.580 (0.568-0.591)
REMS0.729 (0.701-0.758)0.723 (0.698-0.748)0.693 (0.674-0.712)0.669 (0.656-0.682)0.659 (0.648-0.670)

Abbreviations: AUC, area under the curve; CART, Cardiac Arrest Risk Triage; MEWS, Modified Early Warning Score; NEWS, National Early Warning Score; PACS, Patient Acuity Category Scale; RAPS, Rapid Acute Physiology Score; REMS, Rapid Emergency Medicine Score; SERP, Score for Emergency Risk Prediction.

Figure 2.

Receiver Operating Characteristic Curves of Score for Emergency Risk Prediction (SERP) Scores and Other Benchmark Clinical Scores for 2- and 7-Day Mortality

AUC indicates area under the curve; CART, Cardiac Arrest Risk Triage; MEWS, Modified Early Warning Score; NEWS, National Early Warning Score; PACS, Singapore-based Patient Acuity Category Scale; RAPS, Rapid Acute Physiology Score; REMS, Rapid Emergency Medicine Score.

Abbreviations: AUC, area under the curve; CART, Cardiac Arrest Risk Triage; MEWS, Modified Early Warning Score; NEWS, National Early Warning Score; PACS, Patient Acuity Category Scale; RAPS, Rapid Acute Physiology Score; REMS, Rapid Emergency Medicine Score; SERP, Score for Emergency Risk Prediction.

Receiver Operating Characteristic Curves of Score for Emergency Risk Prediction (SERP) Scores and Other Benchmark Clinical Scores for 2- and 7-Day Mortality

AUC indicates area under the curve; CART, Cardiac Arrest Risk Triage; MEWS, Modified Early Warning Score; NEWS, National Early Warning Score; PACS, Singapore-based Patient Acuity Category Scale; RAPS, Rapid Acute Physiology Score; REMS, Rapid Emergency Medicine Score.

Discussion

This cohort study developed parsimonious and point-based SERP scores based on 2-, 7-, and 30-day mortality for risk-stratifying patients after emergency admissions. SERP scores were more accurate in identifying patients who died during short- or long-term care than other point-based clinical tools (ie, PACS, NEWS, MEWS, CART, RAPS, and REMS). A previous study[22] developed a model for inpatient mortality using variables including basic demographic, administrative, and clinical information acquired in the ED. Despite the model showing good discriminative performance, the need to use a computer with 19 variables to calculate a score limited its applicability and interpretability. Instead, SERP is an additive, point-based triage tool, making it simple, quick to calculate, transparent, and interpretable. Moreover, SERP has the advantage of easy implementation, enabling its wide application in different real-world circumstances. Among the 3 SERP scores, SERP-30d achieved satisfactory performance on short-term (eg, 2- or 3-day) mortality and relatively long-term (eg, 30-day) mortality risk estimation. Several possible reasons exist for SERP-30d to excel. The 2-day mortality rate in our cohort was as low as 0.8%. Thus, SERP-2d was developed based on highly imbalanced data, for which the abundance of samples from the majority class (survival group) could overwhelm the minority class in predictive modeling. As a comparison, 30-day mortality contained all 2- or 7-day mortality cases and was more prevalent at a rate of 5.9%, making the SERP-30d score more reliable and accurate. Our results reaffirmed the value of 30-day mortality as an essential indicator for the ED.[23,24,25] Besides vital signs, the SERP-30d score included comorbidities, the importance of which has been demonstrated in several studies.[26,27,28,29] For example, Chu et al[26] reported the contribution of patient comorbidity to short- and long-term mortality. Among all 3 SERP scores, age was selected as a key variable through a data-driven process, which aligns with the evidence on the vital role of age among ED patients.[27,28,29] The SERP scores could provide an objective measure during ED triage to estimate a patient’s mortality risk. Although physicians can generally ascertain the severity of a patient’s acute condition and the threat to life, their decisions are often subjective and depend on an individual’s experience and knowledge. In a study[30] of elderly patients, although physicians could estimate the 30-day mortality during the consultation, they missed 4 of every 5 deaths, with a sensitivity of 20% only. Like the Emergency Severity Index,[31] some triage scores may achieve better performance in risk estimation but require some subjective variables. Some recent studies[32,33,34] highlighted the role of data-driven, objective clinical decision tools to help physicians rethink and reassess the triage process in the ED. Because our SERP scores only comprise objective elements, they can be easily computed by trained medical assistants or integrated into an existing hospital EHR, without the need for professional medical personnel. Therefore, one can rapidly estimate a patient’s risk of death without adversely affecting ED workloads, which is important in the fast-paced ED environment and other heterogenous emergency care systems run by generalists rather than emergency medicine specialists. Given our tool’s purpose as an adjunct to clinical acumen during the consultation, such a risk stratification tool would conceivably be used when a physician plans to admit a patient and considers the level of service that might be appropriate for that individual. Ultimately, the most important unanswered question is whether SERP can improve outcomes in actual clinical practice. To address this, prospective studies are needed to validate its real-world predictive capabilities and determine appropriate thresholds to stratify the ED population into various risk categories. In addition, given the strength of SERP as a simple and interpretable scoring tool, further assessments could be performed to evaluate the more intangible aspects of score implementation.[35,36] Such measures would include determining SERP’s long-term sustainability, overall cost-effectiveness, and physician-perceived acceptability. These future assessments might lend credence to SERP as an effective and accurate tool for decision-making within the ED.

Strengths and Limitations

This study has several strengths. First, machine learning–based variable selection by AutoScore[11] can efficiently filter out redundant information to achieve a sparse solution. Sanchez-Pinto et al[37] also suggested that variable selection plays an essential role in reducing the complexity of prediction models without compromising their accuracy, especially when facing a large number of candidate features extracted from EHRs.[38] Likewise, Liu et al[39] demonstrated that more variables did not necessarily lead to better prediction of adverse cardiac events. The second strength of SERP is the size of the data set that was used to derive the risk scores. This data set is one of the largest used to generate a point-based triage model, with a cohort of more than 300 000 emergency admissions during 8 years, obtained from a large tertiary hospital. Third, the SERP scores consistently performed well in the testing cohort, even with changes in patient characteristics, outcome prevalence, and clinical practices amid the continuously evolving clinical environment.[40] This study also has several limitations. First, the data set used in this study was based on EHR data of routinely collected variables. Thus, some variables, such as socioeconomic status, were not used in SERP score development. Second, because this was a single-center study at a tertiary hospital, the performance of SERP scores may vary in different settings. Third, our ED cohort accounted for only ED admissions, which might influence score generalizability when applying the scores to a general ED population.

Conclusions

SERP is a parsimonious and point-based scoring tool for triaging patients in the ED. In this cohort study, SERP performed better in comparison with existing triage scores and has the advantage of easy implementation and ease of ascertainment at ED presentation. SERP scores have the potential to be widely used and validated in different circumstances and health care settings. Following the clinical application of SERP in ED triage, more tailored scores can be derived in various clinical areas through the machine learning–based AutoScore framework in the future.
  37 in total

1.  Reliability and validity of a new five-level triage instrument.

Authors:  R C Wuerz; L W Milne; D R Eitel; D Travers; N Gilboy
Journal:  Acad Emerg Med       Date:  2000-03       Impact factor: 3.451

2.  Validation of a modified Early Warning Score in medical admissions.

Authors:  C P Subbe; M Kruger; P Rutherford; L Gemmel
Journal:  QJM       Date:  2001-10

3.  Prediction of mortality among emergency medical admissions.

Authors:  S Goodacre; J Turner; J Nicholl
Journal:  Emerg Med J       Date:  2006-05       Impact factor: 2.740

4.  Comparison of the Emergency Severity Index versus the Patient Acuity Category Scale in an emergency setting.

Authors:  Ru Ying Fong; Wee Sern Sim Glen; Ahmad Khairil Mohamed Jamil; Wilson Wai San Tam; Yanika Kowitlawakul
Journal:  Int Emerg Nurs       Date:  2018-06-07       Impact factor: 2.142

5.  Comparison of variable selection methods for clinical predictive modeling.

Authors:  L Nelson Sanchez-Pinto; Laura Ruth Venable; John Fahrenbach; Matthew M Churpek
Journal:  Int J Med Inform       Date:  2018-05-21       Impact factor: 4.046

Review 6.  Clinical Decision Support Systems for Triage in the Emergency Department using Intelligent Systems: a Review.

Authors:  Marta Fernandes; Susana M Vieira; Francisca Leite; Carlos Palos; Stan Finkelstein; João M C Sousa
Journal:  Artif Intell Med       Date:  2019-11-17       Impact factor: 5.326

7.  Comparison of Prediction Model Performance Updating Protocols: Using a Data-Driven Testing Procedure to Guide Updating.

Authors:  Sharon E Davis; Robert A Greevy; Thomas A Lasko; Colin G Walsh; Michael E Matheny
Journal:  AMIA Annu Symp Proc       Date:  2020-03-04

8.  Autonomic nervous system activity as risk predictor in the medical emergency department: a prospective cohort study.

Authors:  Christian Eick; Konstantinos D Rizas; Christine S Meyer-Zürn; Patrick Groga-Bada; Wolfgang Hamm; Florian Kreth; Dietrich Overkamp; Peter Weyrich; Meinrad Gawaz; Axel Bauer
Journal:  Crit Care Med       Date:  2015-05       Impact factor: 7.598

9.  Automated feature selection of predictors in electronic medical records data.

Authors:  Jessica Gronsbell; Jessica Minnier; Sheng Yu; Katherine Liao; Tianxi Cai
Journal:  Biometrics       Date:  2019-04-02       Impact factor: 2.571

10.  Prospective Validation of a Checklist to Predict Short-term Death in Older Patients After Emergency Department Admission in Australia and Ireland.

Authors:  Magnolia Cardona; Michael O'Sullivan; Ebony T Lewis; Robin M Turner; Frances Garden; Hatem Alkhouri; Stephen Asha; John Mackenzie; Margaret Perkins; Sam Suri; Anna Holdgate; Luis Winoto; David C W Chang; Blanca Gallego-Luxan; Sally McCarthy; Ken Hillman; Dorothy Breen
Journal:  Acad Emerg Med       Date:  2018-12-14       Impact factor: 3.451

View more
  7 in total

1.  An external validation study of the Score for Emergency Risk Prediction (SERP), an interpretable machine learning-based triage score for the emergency department.

Authors:  Jae Yong Yu; Feng Xie; Liu Nan; Sunyoung Yoon; Marcus Eng Hock Ong; Yih Yng Ng; Won Chul Cha
Journal:  Sci Rep       Date:  2022-10-19       Impact factor: 4.996

2.  Development and validation of an interpretable prehospital return of spontaneous circulation (P-ROSC) score for patients with out-of-hospital cardiac arrest using machine learning: A retrospective study.

Authors:  Nan Liu; Mingxuan Liu; Xinru Chen; Yilin Ning; Jin Wee Lee; Fahad Javaid Siddiqui; Seyed Ehsan Saffari; Andrew Fu Wah Ho; Sang Do Shin; Matthew Huei-Ming Ma; Hideharu Tanaka; Marcus Eng Hock Ong
Journal:  EClinicalMedicine       Date:  2022-05-06

3.  A Prehospital Triage System to Detect Traumatic Intracranial Hemorrhage Using Machine Learning Algorithms.

Authors:  Daisu Abe; Motoki Inaji; Takeshi Hase; Shota Takahashi; Ryosuke Sakai; Fuga Ayabe; Yoji Tanaka; Yasuhiro Otomo; Taketoshi Maehara
Journal:  JAMA Netw Open       Date:  2022-06-01

4.  Development and validation of an interpretable machine learning scoring tool for estimating time to emergency readmissions.

Authors:  Feng Xie; Nan Liu; Linxuan Yan; Yilin Ning; Ka Keat Lim; Changlin Gong; Yu Heng Kwan; Andrew Fu Wah Ho; Lian Leng Low; Bibhas Chakraborty; Marcus Eng Hock Ong
Journal:  EClinicalMedicine       Date:  2022-03-06

5.  Implementation of prediction models in the emergency department from an implementation science perspective-Determinants, outcomes and real-world impact: A scoping review protocol.

Authors:  Sze Ling Chan; Jin Wee Lee; Marcus Eng Hock Ong; Fahad Javaid Siddiqui; Nicholas Graves; Andrew Fu Wah Ho; Nan Liu
Journal:  PLoS One       Date:  2022-05-12       Impact factor: 3.240

6.  External validation of the Survival After ROSC in Cardiac Arrest (SARICA) score for predicting survival after return of spontaneous circulation using multinational pan-asian cohorts.

Authors:  Maehanyi Frances Rajendram; Faraz Zarisfi; Feng Xie; Nur Shahidah; Pin Pin Pek; Jun Wei Yeo; Benjamin Yong-Qiang Tan; Matthew Ma; Sang Do Shin; Hideharu Tanaka; Marcus Eng Hock Ong; Nan Liu; Andrew Fu Wah Ho
Journal:  Front Med (Lausanne)       Date:  2022-09-08

7.  Leveraging Large-Scale Electronic Health Records and Interpretable Machine Learning for Clinical Decision Making at the Emergency Department: Protocol for System Development and Validation.

Authors:  Nan Liu; Feng Xie; Fahad Javaid Siddiqui; Andrew Fu Wah Ho; Bibhas Chakraborty; Gayathri Devi Nadarajan; Kenneth Boon Kiat Tan; Marcus Eng Hock Ong
Journal:  JMIR Res Protoc       Date:  2022-03-25
  7 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.