Literature DB >> 33965640

Learning from past respiratory failure patients to triage COVID-19 patient ventilator needs: A multi-institutional study.

Harris Carmichael1, Jean Coquet2, Ran Sun2, Shengtian Sang2, Danielle Groat3, Steven M Asch4, Joseph Bledsoe5, Ithan D Peltan6, Jason R Jacobs3, Tina Hernandez-Boussard7.   

Abstract

BACKGROUND: Unlike well-established diseases that base clinical care on randomized trials, past experiences, and training, prognosis in COVID19 relies on a weaker foundation. Knowledge from other respiratory failure diseases may inform clinical decisions in this novel disease. The objective was to predict 48-hour invasive mechanical ventilation (IMV) within 48 h in patients hospitalized with COVID-19 using COVID-like diseases (CLD).
METHODS: This retrospective multicenter study trained machine learning (ML) models on patients hospitalized with CLD to predict IMV within 48 h in COVID-19 patients. CLD patients were identified using diagnosis codes for bacterial pneumonia, viral pneumonia, influenza, unspecified pneumonia and acute respiratory distress syndrome (ARDS), 2008-2019. A total of 16 cohorts were constructed, including any combinations of the four diseases plus an exploratory ARDS cohort, to determine the most appropriate cohort to use. Candidate predictors included demographic and clinical parameters that were previously associated with poor COVID-19 outcomes. Model development included the implementation of logistic regression and three ensemble tree-based algorithms: decision tree, AdaBoost, and XGBoost. Models were validated in hospitalized COVID-19 patients at two healthcare systems, March 2020-July 2020. ML models were trained on CLD patients at Stanford Hospital Alliance (SHA). Models were validated on hospitalized COVID-19 patients at both SHA and Intermountain Healthcare.
RESULTS: CLD training data were obtained from SHA (n = 14,030), and validation data included 444 adult COVID-19 hospitalized patients from SHA (n = 185) and Intermountain (n = 259). XGBoost was the top-performing ML model, and among the 16 CLD training cohorts, the best model achieved an area under curve (AUC) of 0.883 in the validation set. In COVID-19 patients, the prediction models exhibited moderate discrimination performance, with the best models achieving an AUC of 0.77 at SHA and 0.65 at Intermountain. The model trained on all pneumonia and influenza cohorts had the best overall performance (SHA: positive predictive value (PPV) 0.29, negative predictive value (NPV) 0.97, positive likelihood ratio (PLR) 10.7; Intermountain: PPV, 0.23, NPV 0.97, PLR 10.3). We identified important factors associated with IMV that are not traditionally considered for respiratory diseases.
CONCLUSIONS: The performance of prediction models derived from CLD for 48-hour IMV in patients hospitalized with COVID-19 demonstrate high specificity and can be used as a triage tool at point of care. Novel predictors of IMV identified in COVID-19 are often overlooked in clinical practice. Lessons learned from our approach may assist other research institutes seeking to build artificial intelligence technologies for novel or rare diseases with limited data for training and validation.
Copyright © 2021 Elsevier Inc. All rights reserved.

Entities:  

Keywords:  COVID-19; Invasive mechanical ventilation; Machine learning; Respiratory failure; Triage tool

Mesh:

Year:  2021        PMID: 33965640      PMCID: PMC8159260          DOI: 10.1016/j.jbi.2021.103802

Source DB:  PubMed          Journal:  J Biomed Inform        ISSN: 1532-0464            Impact factor:   8.000


Introduction

SARS-CoV-2 has infected over 30 million people worldwide and its disease syndrome, COVID-19, is responsible for more than 940,000 deaths as of September 17, 2020 [1]. With limited disease-specific clinical experience or data, the medical community is substituting evidence and clinical experience from similar diseases to guide treatment and monitoring practices in COVID-19 [2]. Although experience is accumulating, it remains difficult to accurately identify those patients that need close attention and distinguish those that can safely be monitored in lower acuity settings. In the early and middle stages of COVID-19, patients present with respiratory symptoms that are clinically indistinguishable from hundreds of other upper respiratory infections in terms of symptoms and severity [3], [4], [5]. Critical differences appear later in the disease course when patients present in severe hypoxemia and can rapidly deteriorate late in the symptomatic timeline, requiring advanced oxygenation support. An important unanswered question in this pandemic is how to determine which patients will need advanced respiratory support and which will not. The ability to accurately predict invasive mechanical ventilation (IMV) in COVID-19 would allow health systems to appropriately utilize limited resources to more closely monitor high risk patients and could be used to target enrollment for trials of therapies intended to prevent IMV. Barriers to accurate, unbiased prediction of COVID-19 disease trajectory include lack of readily available patient data to train and test models in the early outbreak of the disease and no public datasets available to test and validate models outside of a single healthcare setting. In this study, we hypothesized that machine learning (ML) prediction models trained on patients with respiratory infections similar to COVID-19 could accurately identify hospitalized COVID-19 patients who are at risk of IMV within 48 h. In addition, we share lessons learned from the process of implementing this ML framework and the importance of finding the most representative data at the beginning of the pandemic crisis. The framework can be generalized to benefit any healthcare facility facing limited disease-specific data, as was the case with COVID-19.

Material and methods

Study design

This multicenter retrospective study included patients from two hospitals belonging to the Stanford Healthcare Alliance (SHA) in California using Epic medical record systems and 22 Intermountain Healthcare hospitals in Utah and Idaho using Cerner medical record systems. We evaluated model performance in predicting IMV among COVID-19 hospitalized patients using models developed in COVD-like disease (CLD) patients. The framework of the study is described elsewhere [6]. The Transparent Reporting of a Multivariable Prediction Model for Individual Prognosis or Diagnosis (TRI-POD) checklist was followed [7].

Patient cohorts

COVID-like disease patients: The CLD cohort identified from SHA included patients hospitalized between January 1, 2008 and June 30, 2019 with an acute respiratory infection (unspecified pneumonia, bacterial pneumonia, viral pneumonia, or influenza) or ARDS on the basis of international classification of diseases, ninth and tenth revision (ICD-9-CM and ICD-10-CM) diagnosis codes (eTable 1). Fig. 1 a presents the cohort selection process of the CLD patients. Patients were excluded if they were younger than 18 years old or had Do Not Intubate (DNI) status. We then divided patients into five disease groups based on ICD code using the following hierarchy: ARDS > influenza > viral pneumonia > bacterial pneumonia > unspecified pneumonia. Patients with no laboratory test results during their hospitalizations due to immediate intubation after admission or other unknown reasons were excluded from the training set.
Fig. 1

Cohort selection diagram for (A) COVID-like patients at Stanford Hospital Alliance (SHA) and (B) COVID-19 patients at SHA and Intermountain Healthcare.

Cohort selection diagram for (A) COVID-like patients at Stanford Hospital Alliance (SHA) and (B) COVID-19 patients at SHA and Intermountain Healthcare. COVID-19 patients: At SHA and Intermountain, hospitalized adult patients with COVID-19 were identified with a positive SARS-CoV-2 polymerase chain reaction result or, at SHA, a diagnosis code for COVID-19 (eTable 1 in the Supplement). Patients included were admitted from March 9 to July 25, 2020 at SHA and from March 16 to May 24, 2020 at Intermountain. Fig. 1b describes selection of COVID-19 cohorts from the two institutions.

Data collection

Patient demographic and clinical information were captured from electronic health records (EHRs), including baseline demographics, comorbidities, co-infections, symptoms, and laboratory results during hospitalization (Table 1 ). These variables were identified from the literature that were associated with COVID-19 disease severity [8], [9], [10], [11]. Relevant comorbidities were extracted up to three years prior to hospitalization, including cardiovascular disease, diabetes, cancer, hypertension, chronic respiratory disease, respiratory failure, kidney disease, Alzheimer’s disease, and cirrhosis. Symptoms from 15 days before admission were also extracted for each patient at SHA and symptoms from time of COVID testing were obtained from all Intermountain patients. We identified patients’ coinfection with other respiratory pathogens using both ICD codes and the results of respiratory pathogen panels.
Table 1

Features included in model development.

Demographics and clinical characteristicsLaboratory findingsCo-infection*

Age at admission, years

Gender

Ever smoke (all life before IP)

Comorbidity present (3 years before admission)

Cancer

Chronic Respiratory Disease

Cardiovascular Disease

Hypertension

Type 2 Diabetes

Respiratory Failure

Kidney Disease

Alzheimer’s Disease

Cirrhosis

Symptoms*

Cough

Dyspnea

Tachypnea, respiratory rate > 20

Hypoxemia, oxygen saturation ≤ 90%

Rhinorrhea

Nose congestion

Fever, temperature > 37C/98.6F

Sputum

Pharyngitis (sore throat)

Headache

Fatigue

Conjunctivitis

Diarrhea

Anosmia

Myalgias

White blood cell count, K/uL

Lymphocyte count, K/uL

ALT, U/L

Creatinine, mg/dL

Lactate dehydrogenase, U/L

Creatine kinase, U/L

High-sensitivity cardiac troponin I, ng/mL

D-dimer, ng/mL

Prothrombin time, seconds

Serum ferritin, ng/mL

IL-6, pg/mL

Procalcitonin, ng/mL

Lactate dehydrogenase, U/L

Platelet count, K/uL

C-reactive protein, mg/dL

Total bilirubin, mg/dL

Blood urea nitrogen, mg/dL

Albumin, g/dL

O2 Saturation, mm/Hg

Fraction of inspired oxygen, FiO2, %

Aspartate Aminotransferase, U/L

Sodium (Na+), mmol/L

Potassium (K+), mmol/L

Adenovirus

Chlamydia pneumoniae

Coronavirus

Influenza A

Influenza B

Metapneumovirus

Mycoplasma pneumonia

Parainfluenza 1

Parainfluenza 2

Parainfluenza 3

Parainfluenza 4

Rhinovirus/Enterovirus

Respiratory syncytial virus (RSV)

within 15 days before admission.

Features included in model development. Age at admission, years Gender Ever smoke (all life before IP) Comorbidity present (3 years before admission) Cancer Chronic Respiratory Disease Cardiovascular Disease Hypertension Type 2 Diabetes Respiratory Failure Kidney Disease Alzheimer’s Disease Cirrhosis Symptoms* Cough Dyspnea Tachypnea, respiratory rate > 20 Hypoxemia, oxygen saturation ≤ 90% Rhinorrhea Nose congestion Fever, temperature > 37C/98.6F Sputum Pharyngitis (sore throat) Headache Fatigue Conjunctivitis Diarrhea Anosmia Myalgias White blood cell count, K/uL Lymphocyte count, K/uL ALT, U/L Creatinine, mg/dL Lactate dehydrogenase, U/L Creatine kinase, U/L High-sensitivity cardiac troponin I, ng/mL D-dimer, ng/mL Prothrombin time, seconds Serum ferritin, ng/mL IL-6, pg/mL Procalcitonin, ng/mL Lactate dehydrogenase, U/L Platelet count, K/uL C-reactive protein, mg/dL Total bilirubin, mg/dL Blood urea nitrogen, mg/dL Albumin, g/dL O2 Saturation, mm/Hg Fraction of inspired oxygen, FiO2, % Aspartate Aminotransferase, U/L Sodium (Na+), mmol/L Potassium (K+), mmol/L Adenovirus Chlamydia pneumoniae Coronavirus Influenza A Influenza B Metapneumovirus Mycoplasma pneumonia Parainfluenza 1 Parainfluenza 2 Parainfluenza 3 Parainfluenza 4 Rhinovirus/Enterovirus Respiratory syncytial virus (RSV) within 15 days before admission.

Outcome

The study outcome was the risk of IMV within 48 h. Daily risk scores were generated for each patient based on the most recent data available, providing a rolling window of risk scores throughout the patient’s hospital stay. CPT codes for endotracheal intubation (31500) or ventilator management (94002–94005) were used to identify IMV at SHA; respiratory documentation abstracted from the electronic data warehouse was used to identify IMV initiation at Intermountain. A sensitivity analysis was performed for prediction of IMV at 24 and 72 h. The non-IMV group included patients who were discharged alive and died without IMV.

Machine learning (ML) algorithms

Cohort Generation. We explored the performance of multiple models trained on 16 training cohorts including 15 cohorts with different combinations of influenza, bacterial pneumonia, viral pneumonia and unspecified pneumonia plus a separate exploratory cohort of patients who developed ARDS during their hospital stay. Initially, many consensus-driven treatment recommendations for COVID were modeled on ARDS algorithms [12]. As ARDS is a syndrome with many possible etiologies, including non-infectious etiologies, it was modeled separately. Details of cohort generation have been previously been published [6]. Model Development. We developed a rolling window of predictions; each hospital day, a risk score for IMV in the upcoming 48 h was derived for each patient using the most recent laboratory results that were available up until two weeks, with one value per variable. To handle the data imbalance challenge, undersampling in conjunction with an oversampling strategy was used to balance samples between groups. Random undersampling was applied to trim the number of patients in the majority class while oversampling strategies were performed to generate synthetical samples of the minority class. Oversampling approaches considered included synthetic minority oversampling technique (SMOTE), borderline SMOTE, and SVM SMOTE. For each of the 16 training cohorts, 70% of data were randomly selected as training set and 30% were used as a validation set. Several machine learning models that have been frequently applied to predict similar outcomes and with good interpretability were selected for model training, including logistic regression and decision tree, AdaBoost and XGBoost. We used the three decision tree based algorithms because they have been previously applied to predict clinical events in patients with respiratory diseases based on EHR data [13], [14], [15]. To determine the best hyperparameters for each ML algorithm, a grid search and 10-fold cross-validation were used on the training data, and hyperparameters of the model with the best AUC were selected. Using the selected hyperparameters, we tested the model in the remaining unseen 30% of the data. The training procedure stopped once the performance on the validation dataset did not improve after 20 training iterations. Sixteen ML models were derived based on different combinations of the five CLD cohorts. The model training processes are presented in eFigure 1. Model Performance. The performance of the trained models were quantified using the area under the receiver operator characteristic curve (AUC). We further validated the models on the COVID-19 cohorts. Additional evaluation criteria considered included accuracy (ACC), positive predictive value (PPV), negative predictive value (NPV), sensitivity, specificity, and positive likelihood ratio (PLR). A threshold of 0.5 was selected and values greater than 0.5 indicated a positive prediction (i.e. IMV would occur in the next 48 h). We further calculated Brier scores to determine the accuracy of the models’ probabilistic predictions on the two testing sites. SHapley Additive exPlanations (SHAP) values were calculated for model interpretation [16].

Sensitivity analysis

Sensitivity analyses measured the prediction model’s performance of IMV during alternative follow-up windows (24 h and 72 h).

Statistical analysis

We used descriptive statistics to compare the demographic and clinical characteristics of the CLD and COVID-19 combined cohorts from SHA and Intermountain. We further compared patient symptoms and laboratory test results between IMV and non-IMV COVID-19 patients. Independent t-test or Mann-Whitney U test were used for continuous features, and Pearson Chi-square test or Fisher’s exact test was applied for categorical features wherever appropriate. Statistical significance for primary analysis was set at p-value < 0.05. All tests were 2-tailed. ML models were implemented with Python scikit-learn and XGBoost packages and analyses were performed in R version 3.5.2.

Results

We included a total of 14,030 patients with CLDs and 444 patients in the COVID-19 cohort (Table 2 ). Patients who had no laboratory results (n = 2,695, 19.2%) were excluded. COVID-19 patients were significantly younger (53.4 ± 17.2 vs. 63.7 ± 18.9, p < 0.001) with fewer patients with comorbidities (47.7% vs. 64.9%, p < 0.001) than the CLD patients. Minorities were disproportionately represented among COVID-19 patients, who were also less likely to have public insurance (56.5% vs 63.9%, p < 0.001) or have a smoking history (21.4% vs 40.4%, p < 0.001).
Table 2

Baseline characteristics of patients hospitalized in the COVID-Like Disease (CLD) cohort and the COVID-19 cohort.

VariablesCLDCOVID-19
p-value*
SHAIntermountainTotal
Total, N14,030185259444
Age (years), Mean ± SD63.7 ± 18.953.9 ± 17.853.0 ± 16.553.4 ± 17.2<0.001
GenderMale7673 (54.7)92 (49.7)145 (56.0)237 (53.4)0.582
Race, N (%)White7990 (56.9)42 (22.7)179 (69.1)221 (49.8)<0.001
Asian2213 (15.8)24 (13.0)<1026 (5.9)
Black867 (6.2)<10<1015 (3.4)
Other2612 (18.6)97 (52.4)72 (27.8)169 (38.1)
EthnicityHispanic2023 (14.4)94 (50.8)98 (37.8)192 (43.2)<0.001
Insurance, N(%)Private3309 (23.6)38 (20.5)123 (47.5)161 (36.3)< 0.001
Public8976 (63.9)130 (70.2)121 (46.7)251 (56.5)
Other333 (2.4)17 (9.2)14 (5.4)31 (7.0)
Ever smoke, N (%)5675 (40.4)55 (29.7)40 (15.4)95 (21.4)< 0.001
One or more comorbidity, N(%)9108 (64.9)64 (34.6)148 (57.1)212 (47.7)<0.001
IMV, N (%)3775 (26.9)25 (13.5)35 (13.5)60 (13.5)<0.001
Time to IMV (days), Mean ± SD4.1 ± 7.63.2 ± 4.32.7 ± 2.22.9 ± 3.20.001
Inpatient mortality, N (%)1925 (13.7)<10†11 (4.3)<0.001

Note. *p value was calculated comparing patients with CLDs to the COVID-19 combined cohort; Percentages may not add up to 100 due to missing values.

Baseline characteristics of patients hospitalized in the COVID-Like Disease (CLD) cohort and the COVID-19 cohort. Note. *p value was calculated comparing patients with CLDs to the COVID-19 combined cohort; Percentages may not add up to 100 due to missing values. Among patients with CLDs, 26.9% received IMV beginning an average of 4.1 ± 7.6 days after hospital admission, and in the COVID-19 cohort 13.5% received IMV beginning 2.9 ± 3.2 days from admission (p = 0.001). COVID-19 patients requiring IMV had lower platelets and lymphocyte counts, higher lactate dehydrogenase (LDH) and c-reactive protein (CRP), and higher fraction of inspired oxygen (Table 3 ).
Table 3

Clinical characteristics of hospitalized COVID-19 patients, stratified by Invasive Mechanical Ventilation (IMV).

Variablenon-IMV
IMV
p-value
SHAIntermountainCombinedSHAIntermountainCombined
Patients (N)137224361253560
Age57.8 ± 15.253.3 ± 16.655.0 ± 16.156.8 ± 20.053.1 ± 16.454.6 ± 18.00.861
Gender, male67 (48.9)122 (54.46)189 (52.3)11(44.0)23 (65.7)34 (56.7)0.535
Ever smoke31 (22.6)33 (14.73)64 (17.7)12(48.0)<1019 (31.7)0.012
1 + comorbidities41 (29.9)124 (55.36)165 (45.7)10(40.0)24 (68.57)35 (58.3)0.070
Co-Infections (Yes/No)10 (7.3)<1011 (3.0)<100(0.0)<100.435



Symptoms at admission
Cough105 (76.6)182 (81.3)287 (79.5)17 (68.0)28 (80.0)45 (75.0)0.429
Dyspnea104 (75.9)188 (83.9)292 (80.0)20 (80.0)32 (91.4)52 (86.7)0.284
Fever97 (70.8)133 (59.4)230 (63.7)16 (64.0)27 (77.1)43 (71.7)0.232
Fatigue62 (45.3)67 (29.9)129 (35.7)10 (40.0)13 (37.1)23 (38.3)0.698
Myalgias52 (38.0)83 (37.1)135 (37.4)10 (40.0)<1019 (31.7)0.394
Headache28 (20.4)46 (20.5)74 (20.5)<10<1011 (18.3)0.699
Diarrhea24 (17.5)79 (35.3)103 (28.5)<1016 (45.7)23 (38.3)0.125
Tachypnea50 (36.5)33 (13.5)83 (23.0)11 (44.0)<1019 (31.7)0.146
Pharyngitis22 (16.1)66 (29.5)88 (24.4)<1013 (37.1)20 (33.3)0.141
Hypoxemia34 (24.8)25 (10.3)59 (16.3)15 (60.0)<1023 (38.3)<0.001
Sputum25 (18.2)44 (19.6)69 (19.1)<10<1011 (18.3)0.887
Anosmia20 (14.6)26 (11.6)46 (12.7)<10<10<100.816
Rhinorrhea<1060 (26.8)64 (17.7)<1013 (37.1)14 (23.3)0.301
Nose congestion<100<10<100 (0.0)<100.599



Laboratory Values, Mean ± SD
White blood cell count, K/uL7.0 ± 2.56.6 ± 2.56.7 ± 2.58.3 ± 4.16.7 ± 5.87.4 ± 5.10.095
Lymphocyte count, K/uL1.3 ± 0.71.3 ± 0.81.3 ± 0.80.9 ± 0.40.9 ± 0.50.9 ± 0.4<0.001
Platelet count, K/uL262.7 ± 97.8243.5 ± 93.1250.8 ± 94.9197.0 ± 70.8192.2 ± 96.6194.4 ± 85.7<0.001
Alanine aminotransferase, U/L50.4 ± 41.254.3 ± 67.952.7 ± 58.835.7 ± 23.751.6 ± 55.344.2 ± 43.70.308
Aspartate aminotransferase, U/L54.1 ± 44.657.4 ± 51.756.1 ± 49.143.9 ± 18.673.2 ± 81.659.6 ± 61.20.634
Total bilirubin, mg/dL0.6 ± 1.80.6 ± 0.30.6 ± 1.10.6 ± 0.20.7 (0.4)0.6 ± 0.30.841
Albumin, g/dL3.5 ± 0.53.5 ± 0.43.5 ± 0.53.3 ± 0.63.6 ± 0.53.5 ± 0.50.453
Blood urea nitrogen, mg/dL15.8 ± 11.914.3 ± 8.914.8 ± 10.122.4 ± 17.818.3 ± 11.520.2 ± 14.7<0.001
Troponin I, ng/mL0.08 ± 0.10.45 ± 4.10.42 ± 3.890.12 ± 0.170.03 ± 0.10.05 ± 0.10.671
D-dimer, ng/mL1233.4 ± 699.11398 ± 20181374.7 ± 1893.11285.5 ± 890.91308 ± 7161303.8 ± 743.50.883
Lactate dehydrogenase, LDH, U/L363.7 ± 113.7359.9 ± 132.6361.6 ± 124.6460.7 ± 150.2540.9 ± 230.4499.4 ± 193.0<0.001
Serum ferritin, ng/mL896.6 ± 905.6851.8 ± 1074870.3 ± 1007.91289.5 ± 896.61315 ± 11221300.1 ± 994.60.031
Procalcitonin, ng/mL0.2 ± 0.30.4 ± 1.40.3 ± 1.10.6 ± 0.70.6 ± 1.40.6 ± 1.10.112
C-reactive protein, mg/dL8.3 ± 6.09.0 ± 6.88.8 ± 6.617.9 ± 11.713.5 ± 9.115.6 ± 10.4<0.001
Creatine kinase, U/L188.0 ± 226.9481.2 ± 1660322.6 ± 1136.2219.1 ± 258.2192.0 ± 229.0209.6 ± 248.90.660
Sodium (Na+), mmol/L136.7 ± 3.2137.0 ± 2.9136.9 ± 3.0137.7 ± 3.0135.7 ± 3.5136.6 ± 3.20.546
Potassium (K+), mmol/L4.0 ± 0.33.8 ± 0.43.9 ± 0.34.0 ± 0.73.8 ± 0.53.9 ± 0.60.884
Creatinine, mg/dL1,0 ± 1.31.0 ± 1.21.0 ± 1.21.3 ± 1.81.0 ± 0.41.1 ± 1.30.363
Prothrombin time, (s)14.0 ± 2.415.1 ± 4.514.3 ± 3.214.4 ± 1.715.8 ± 6.214.8 ± 3.40.545
Oxygen saturation, SpO2, %95.7 ± 1.993.9 ± 2.294.6 ± 2.195.6 ± 2.392.9 ± 2.894.2 ± 2.60.189
Fraction of inspired oxygen, %0.29 ± 0.10.56 ± 0.30.4 ± 0.20.6 ± 0.20.6 ± 0.30.6 ± 0.2<0.001
Clinical characteristics of hospitalized COVID-19 patients, stratified by Invasive Mechanical Ventilation (IMV).

Model performance

Across all patients, we retrieved a total of 79,233 patient days, of which 2841 were positive intubation days. The XGBoost model outperformed other ML models in the training data and was used for all subsequent analyses; the performance metrics for the XGBoost model in the training data are presented in eTable 3. The model that included all pneumonia and influenza cohorts, without ARDS, performed best across all metrics at both validation sites. Table 4 present the performance of the best performing model at SHA and Intermountain. In the COVID-19 validation cohorts, the AUC was 0.77 at SHA and 0.65, at Intermountain (Fig. 2 ), and the statistics as a triage tool for impending IMV remained robust with excellent specificity and PLR (Table 4). While other models had higher AUCs, the model containing the four pneumonia and influenza cohorts had better overall performance metrics, such as PPV, NPV and ACC (eTables 4 and 5). The sensitivities were 0.15 at SHA and 0.10 at Intermountain, while specificity was 0.99 at both sites. The Brier score was 0.047 at SHA and 0.033 at Intermountain. Only 1.8% and 1.2% of total patient days were predicted as positive at SHA and IM respectively. The PPV in COVID-19 patients was 0.29 at SHA and 0.23 at Intermountain while NPV was 0.97 at both institutions. PLR were 10.7 and 10.0 at SHA and Intermountain, respectively.
Table 4

Performance of the best performing modele at Stanford Hospital Alliance (SHA) Intermountain Healthcare for COVID-like disease and COVID-19 hospitalized patients.

Validation datasetsAUC*ACC aPPVbSensitivityF-scoreSpecificityNPVcPLRd
COVID-like disease patients0.8710.8290.7310.4910.5880.9400.8498.183
COVID-19 patients at SHA0.7720.9570.2860.1500.1970.9860.97010.714
COVID-19 patients at Intermountain0.6480.9650.2350.1000.1400.990.9749.999

AUC: Area under the ROC Curve; aACC: accuracy; bPPV: positive predictive value; cNPV: negative predictive value. dPLR: positive likelihood ratio; eselected model with all four diseases: viral pneumonia, bacterial pneumonia, unspecified pneumonia, and influenza.

Fig. 2

Receiver Operating Characteristic Curves for trained Model on COVID-19 patients at (A) Stanford Hospital Alliance and (B) Intermountain Healthcare.

Performance of the best performing modele at Stanford Hospital Alliance (SHA) Intermountain Healthcare for COVID-like disease and COVID-19 hospitalized patients. AUC: Area under the ROC Curve; aACC: accuracy; bPPV: positive predictive value; cNPV: negative predictive value. dPLR: positive likelihood ratio; eselected model with all four diseases: viral pneumonia, bacterial pneumonia, unspecified pneumonia, and influenza. Receiver Operating Characteristic Curves for trained Model on COVID-19 patients at (A) Stanford Hospital Alliance and (B) Intermountain Healthcare.

Algorithm variable importance

The top ten most important variables for the IMV prediction were: D-dimer, serum ferritin, LDH, fraction of inspired oxygen, total bilirubin, SPO2, CRP, prothrombin time, sodium (Na+), and creatine kinase. The full ranked list of variables is available in eFigure 2 in the Supplement. We used the trained model that predicted IMV within 48 h to test different prediction lead times, specifically IMV at 24 and 72 h. IMV prediction at 24 h had higher AUCs, yet lower PPVs compared to the 48-hour model at both validation sites. Predicting IMV at 72 h produced AUCs lower than the 48-hour models yet higher PPVs, with the highest PPV reaching 0.381 in the SHC validation set. (eTable 6)

Discussion

Faced with a novel disease and high - even overwhelming - patient volumes, clinicians and hospitals need to understand how well tools and care patterns designed for apparently similar respiratory syndromes can function for COVID-19 care. We leveraged data from other respiratory syndrome patients (COVID-19 like patients) to train and test a ML model to predict IMV within 48 h and measured its performance when applied to hospitalized COVID-19 patients from two independent healthcare centers with different EHR systems. Our framework leverages the abundance of retrospective respiratory failure patients to provide a testing bed for discovery and innovation. Based on our model’s excellent clinical performance in PLR, the best utility would be in highlighting those patients at highest risk of IMV within 48 h out of all hospitalized COVID patients with a moderate risk of respiratory failure. Under anticipated surges of hospitalized patients based on local community incidence, this tool may help clinicians and hospital systems with limited specialized resources focus limited expertise on COVID-19 patients at highest risk of deterioration. Given the relatively low rates of positive results from the model (13.5%) there is a permissive tolerance of false positives that may receive closer monitoring but not require intubation. Excellent NPV and specificity allow confidence in standard of care monitoring for those patients with negative results from the model. Such a tool could also help enrich therapeutic trials with patients more likely to benefit from particular therapies, thereby improving study power and efficiency. We believe lessons learned from the COVID experience can serve others interested in designing and implementing prognostic models in novel or rare diseases, where training data may be limited. The model performed well at SHA but less optimal in Intermountain Differences from the two facilities were significant and include, but are not limited to, EMR systems, EHR source data definitions and data acquisition systems, geographic area covered, patient demographics (Tables 2 and 3), provider practice habits, equipment, and dedicated tertiary training facility vs collection of community hospitals and tertiary referral center. This variation may explain some of the differences in AUC but also make it likely that these results are robust against even greater variation than already observed and reported in this manuscript. Importantly, while AUC's did vary, the tool continued to maintain utility at the included sites. The AUC and F-score derived in both the training and validation data sets are comparable to other validated inpatient prediction models such as APACHE II and IV [17] and current COVID prediction models [18]. However, many predictive models for COVID-19 report only model accuracy and fail to report the additional information needed to guide clinical decisions, such as sensitivity and specificity. In order for any predictive model to be useful, the clinical utility of the model must be a priority and an understanding of how such a model can be used at point of care is essential. In this study the top performing model was based on performance across all metrics, not only the AUC. We therefore selected the model trained on the cohort that included all pneumonia and influenza patients, which did not have the highest AUC in the validation sets but had higher PPVs and PLR. In addition, our study builds on the existing literature by predicting IMV within a 48-hour rolling window, updating individual patient inputs daily, and was externally validated using data from a multihospital health system employing a different EHR system. Hospitals experiencing a surge of COVID-19 patients will have limited experienced hospital clinicians or infectious disease providers, which emphasizes the need for a tool to help triage patients and identify those most likely to clinically deteriorate. ML algorithms, such as our model, are ideal triage tools that may sort patients for higher priority to be reviewed by these experts. In consonance with emerging evidence on the association of inflammation and COVID-19 outcomes [19], our model identified inflammatory and coagulation markers (e.g., D-dimer, ferritin, LDH, CRP, etc.) as features more predictive than factors such as age, degree of hypoxic respiratory failure, and other markers of organ failure. The fact that our model leverages potentially underappreciated laboratory features, not included in many prior predictive models, for hospitalized respiratory syndromes indicates that these results could identify impending decline in patients otherwise overlooked. Additionally, excellent PLR—positive results dramatically increase probability of IMV—at both institutions and indicate our model can appropriately identify patients for close monitoring of imminent IMV. Although PPV was low, this largely reflects the low daily intubation rate (1.2% at our external validation site, Intermountain). A “positive” risk score nevertheless indicated a substantial increase in individual risk as assessed by the PLR. Lastly, the rolling daily risk assessments are also an important feature of our predictive model and is particularly important due to the rapid deterioration observed in COVID-19 patients. This model would be well suited for automatic screening to target COVID patients for closer monitoring on admission and during daily rounds. Despite the disease’s novelty, the literature is already flooded with prediction models for COVID-19, including risk, diagnostic, and prognostic models [14], [18], [20], [21], [22]. Our results are similar to previous work attempting to predict clinical COVID-19 patient clinical deterioration [23]. While some of our predictive variables overlap with previous models, our model identifies emerging inflammatory markers previously overlooked. Given the pressing issues of the pandemic, some of these models were under-developed and potentially biased, resulting in retraction of reports [24]. Our framework and transparency in reporting of the details of the training population, model architecture, and hyperparameter tuning follows emerging standards for reporting ML models in healthcare [7], [25]. We have made our algorithms and code publicly available, and hope others will test and build on our work. This study has limitations that may affect the interpretation of results. First, there was a relatively small number of COVID-19 patients receiving IMV in our population. This could negatively impact the stability of model's performance estimates. Second, we selected features that were associated with COVID-19 outcomes in existing literature. As our knowledge of the disease progresses, additional features are emerging that may also contribute to the risk of IMV, which were not included in the model. However, this study lays the framework for additional studies to validate emerging risk predictors. Finally, clinical practice variation, differences in institutional protocols, case volumes, and availability of clinical resources could all affect the predictive accuracy of our model. COVID-19 patient surges may substantially alter the threshold for initiating IMV, which may limit the utility of our model in regions with extreme patient volumes. As part of a future investigation, an interesting strategy would be to identify the best threshold for both institutes using a subset of the COVID-19 cohort to calibrate a local threshold. Due to the small number of patients with COVID in some cohort groups, our study was limited in the scope of machine learning methods to train the model. We selected several machine learning models that have been successfully applied in other studies and with good interpretability in terms of feature significance, which is essential when implementing such predictive models in clinical settings. However, with the increase in the number of COVID-19 patients and the possibility of performing a crossover study, a future investigation would be to apply several deep learning methods, such as recurrent neural network (RNN), which would allow us to take into consideration the temporal sequence between hospitalization days. These deep learning methods may have the potential to outperform our current model in terms of accuracy and should be considered in future studies with larger sample sizes Integrating prospective and retrospective outcomes from regions with varied patient volumes and demographics would strengthen our model and we encourage the reporting of this data to improve our models.

Conclusion

Medical knowledge often progresses by analogizing from pathophysiologically related but distinct disease entities. To address the lack of COVID specific data, we used the abundant historical data from related conditions to address prognostication for a novel epidemic infection. The commonality between COVID-19 and other common respiratory failure diseases can help provide guidance to providers treating COVID-19 patients during this novel pandemic where clinical experience and evidence remain sparse. When validated on COVID-19 patients from two healthcare systems, our best-performing model demonstrated the capacity to meaningfully inform the clinical suspicion that a COVID-19 patient will need IMV. Future efforts to increase model sensitivity and further bolster positive predictive value will improve the model’s clinical utility.

Lessons learned

In the beginning of the pandemic, many institutes were challenged with a limited number of cases for model training and limited knowledge about the novel COVID-19 disease. The quick surge of COVID-19 patients overflows health facilities and a decision support tool was in great need for triaging patients. Many rushed to produce models with limited training data [20] that led to potentially biased models [24]. This work proposes an alternative approach that addresses the needs of many institutions to support decision making in the void of robust data sets. Such a framework could support the rapid dissemination of prognostic models at the next pandemic or for rare diseases. In the haste of developing AI-based COVID decision tools, most published models have fallen short of providing safe and effective guidance [24]. In a continuously updated systematic review of COVID-19 prediction models, most models lack transparency and are at high risk of bias [20]. External validation is not common and performance drops dramatically when these models are tested on a different dataset, which can make these models irrelevant for deployment at the point of care outside the training data site [26]. The model reported in this study also suffers from generalizability to some extent, however there are important lessons that can be learned from this endeavor. Data extraction procedures vary greatly across systems. To ensure generalizability, one must empirically validate each feature used in the model, including missingness, distribution, time captured, and units captured. For example, oxygen flow was an essential feature of this model, yet captured very differently at the two sites. For one site, conversion of values for FiO2 were performed; the values recorded included either the liters of oxygen per minutes delivered or the actual FiO2 values, we converted the oxygen delivered into FiO2 when necessary. While the other site has this feature recorded cleaner and more complete. In this work, one institute created a panel to obtain many features for each COVID patient hospitalized. In contrast, features at the other hospital had to be scraped from flowcharts. Generalizability was also affected by the differences in patient populations, where there were important differences in race, ethnicity, smoking status, and pre-existing conditions, which might also be a reflection of data capture. However, given that we are one of the few COVID models that have actually performed cross-site validation, the model is likely more generalizable than most. This work demonstrated the feasibility of cross-site validation in a very short period of time and suggests that clinical decision tools need to prove robustness against a wide range of generalizability challenges. In addition to model development, we provide guidance on the implementation of our framework and multi-institute validation. First, it is important to identify ‘like-cohorts’ that share outcome characteristics and clinical manifestations with the disease of interest. These like-cohorts must be common with sufficient sample sizes for training, testing, and validating the ML models. Second, a system should be in place to extract cohorts and variables from the EHRs and, importantly, identify new variables that might be unique to the disease of interest and stored in flowsheets or as unstructured data in clinical narrative text. Third, at the onset of the pandemic, COVID-19 treatment was unsure and guidelines did not exist, large practice variation was seen across the globe and needs to be considered when implementing clinical decision support tools. Therefore, cross-validation across care settings is necessary. During cross-validation, every variable used in the model must be thoroughly investigated, including data capture, storage, missingness and validity. Finally, geographical variation must be considered, as the outbreak of COVID-19 in New York City is greatly different from outbreaks in other cities across the globe, again highlighting the need for cross validation. The lessons learned for clinical model application were that models trained on data of a “patient-like me” cohort (i.e., COVID-like diseases) can have good accuracy and present clinical utility by identifying inpatients at high risk of decomposition. Importantly, we also learned that laboratory factors associated with inflammation (i.e., D-dimer, ferritin, LDH, CRP), that have not been previously appreciated as predictors for patient deterioration in respiratory illnesses, were highly influential in our model. The framework provides an opportunity to quickly identify unsuspected risk factors associated with disease outcomes at the onset of future pandemics.

Funding xxx

IDP was supported by a career development award (K23GM129661) from the .

Ethics approval and consent to participate

This study received approval from the Institutional Review Board (IRB) of both participating healthcare systems.

Consent for publication

Not applicable.

Availability of data and materials

All code and trained ML models are publicly available via GitHub as well as a de-identified sample dataset [URL to be published with manuscript].

Authors' contributions

THB takes responsibility for the integrity of the data and the accuracy of the data analysis. Concept and design: THB Statistical analysis: JC, RS, and HLC Collection of data: JC, RS, DG, JJ, IDP Interpretation of data: All authors Drafting of the manuscript: HLC, JC, RS, SS, THB Critical revision of the manuscript: All authors Administrative, technical, or material support: THB, IDP Study supervision: THB, IDP

Declaration of Competing Interest

The authors declare the following financial interests/personal relationships which may be considered as potential competing interests: Ithan Peltan has received research support to their institution from Janssen Pharmaceuticals, Immunexpress Inc., and Regeneron Pharmaceuticals. The other authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.
  22 in total

1.  Community Prevalence of SARS-CoV-2 Among Patients With Influenzalike Illnesses Presenting to a Los Angeles Medical Center in March 2020.

Authors:  Brad Spellberg; Meredith Haddix; Rebecca Lee; Susan Butler-Wu; Paul Holtom; Hal Yee; Prabhu Gounder
Journal:  JAMA       Date:  2020-05-19       Impact factor: 56.272

2.  What Other Countries Can Learn From Italy During the COVID-19 Pandemic.

Authors:  Stefania Boccia; Walter Ricciardi; John P A Ioannidis
Journal:  JAMA Intern Med       Date:  2020-07-01       Impact factor: 21.873

3.  Length-of-Stay Prediction for Pediatric Patients With Respiratory Diseases Using Decision Tree Methods.

Authors:  Fei Ma; Limin Yu; Lishan Ye; David D Yao; Weifen Zhuang
Journal:  IEEE J Biomed Health Inform       Date:  2020-02-24       Impact factor: 5.772

4.  Development and Validation of a Clinical Risk Score to Predict the Occurrence of Critical Illness in Hospitalized Patients With COVID-19.

Authors:  Wenhua Liang; Hengrui Liang; Limin Ou; Binfeng Chen; Ailan Chen; Caichen Li; Yimin Li; Weijie Guan; Ling Sang; Jiatao Lu; Yuanda Xu; Guoqiang Chen; Haiyan Guo; Jun Guo; Zisheng Chen; Yi Zhao; Shiyue Li; Nuofu Zhang; Nanshan Zhong; Jianxing He
Journal:  JAMA Intern Med       Date:  2020-08-01       Impact factor: 21.873

5.  Comparison of acute physiology and chronic health evaluation II (APACHE II) and acute physiology and chronic health evaluation IV (APACHE IV) severity of illness scoring systems, in a multidisciplinary ICU.

Authors:  Yeldho Eason Varghese; M S Kalaiselvan; M K Renuka; A S Arunkumar
Journal:  J Anaesthesiol Clin Pharmacol       Date:  2017 Apr-Jun

6.  Bias at warp speed: how AI may contribute to the disparities gap in the time of COVID-19.

Authors:  Eliane Röösli; Brian Rice; Tina Hernandez-Boussard
Journal:  J Am Med Inform Assoc       Date:  2021-01-15       Impact factor: 4.497

7.  Learning From Past Respiratory Infections to Predict COVID-19 Outcomes: Retrospective Study.

Authors:  Shengtian Sang; Ran Sun; Jean Coquet; Harris Carmichael; Tina Seto; Tina Hernandez-Boussard
Journal:  J Med Internet Res       Date:  2021-02-22       Impact factor: 5.428

8.  MINIMAR (MINimum Information for Medical AI Reporting): Developing reporting standards for artificial intelligence in health care.

Authors:  Tina Hernandez-Boussard; Selen Bozkurt; John P A Ioannidis; Nigam H Shah
Journal:  J Am Med Inform Assoc       Date:  2020-12-09       Impact factor: 4.497

9.  Clinical course and risk factors for mortality of adult inpatients with COVID-19 in Wuhan, China: a retrospective cohort study.

Authors:  Fei Zhou; Ting Yu; Ronghui Du; Guohui Fan; Ying Liu; Zhibo Liu; Jie Xiang; Yeming Wang; Bin Song; Xiaoying Gu; Lulu Guan; Yuan Wei; Hui Li; Xudong Wu; Jiuyang Xu; Shengjin Tu; Yi Zhang; Hua Chen; Bin Cao
Journal:  Lancet       Date:  2020-03-11       Impact factor: 79.321

10.  Development and validation of a model for individualized prediction of hospitalization risk in 4,536 patients with COVID-19.

Authors:  Lara Jehi; Xinge Ji; Alex Milinovich; Serpil Erzurum; Amy Merlino; Steve Gordon; James B Young; Michael W Kattan
Journal:  PLoS One       Date:  2020-08-11       Impact factor: 3.240

View more
  3 in total

1.  Preparing for the next pandemic via transfer learning from existing diseases with hierarchical multi-modal BERT: a study on COVID-19 outcome prediction.

Authors:  Khushbu Agarwal; Sutanay Choudhury; Sindhu Tipirneni; Pritam Mukherjee; Colby Ham; Suzanne Tamang; Matthew Baker; Siyi Tang; Veysel Kocaman; Olivier Gevaert; Robert Rallo; Chandan K Reddy
Journal:  Sci Rep       Date:  2022-06-24       Impact factor: 4.996

2.  Novel informatics approaches to COVID-19 Research: From methods to applications.

Authors:  Hua Xu; David L Buckeridge; Fei Wang; Peter Tarczy-Hornoch
Journal:  J Biomed Inform       Date:  2022-02-16       Impact factor: 8.000

3.  Early identification of patients admitted to hospital for covid-19 at risk of clinical deterioration: model development and multisite external validation study.

Authors:  Fahad Kamran; Shengpu Tang; Erkin Otles; Dustin S McEvoy; Sameh N Saleh; Jen Gong; Benjamin Y Li; Sayon Dutta; Xinran Liu; Richard J Medford; Thomas S Valley; Lauren R West; Karandeep Singh; Seth Blumberg; John P Donnelly; Erica S Shenoy; John Z Ayanian; Brahmajee K Nallamothu; Michael W Sjoding; Jenna Wiens
Journal:  BMJ       Date:  2022-02-17
  3 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.