Literature DB >> 34583416

Development of Prediction Models for Unplanned Hospital Readmission within 30 Days Based on Common Data Model: A Feasibility Study.

Sooyoung Yoo1, Jinwook Choi2,3, Borim Ryu1,2, Seok Kim1.   

Abstract

BACKGROUND: Unplanned hospital readmission after discharge reflects low satisfaction and reliability in care and the possibility of potential medical accidents, and is thus indicative of the quality of patient care and the appropriateness of discharge plans.
OBJECTIVES: The purpose of this study was to develop and validate prediction models for all-cause unplanned hospital readmissions within 30 days of discharge, based on a common data model (CDM), which can be applied to multiple institutions for efficient readmission management.
METHODS: Retrospective patient-level prediction models were developed based on clinical data of two tertiary general university hospitals converted into a CDM developed by Observational Medical Outcomes Partnership. Machine learning classification models based on the LASSO logistic regression model, decision tree, AdaBoost, random forest, and gradient boosting machine (GBM) were developed and tested by manipulating a set of CDM variables. An internal 10-fold cross-validation was performed on the target data of the model. To examine its transportability, the model was externally validated. Verification indicators helped evaluate the model performance based on the values of area under the curve (AUC).
RESULTS: Based on the time interval for outcome prediction, it was confirmed that the prediction model targeting the variables obtained within 30 days of discharge was the most efficient (AUC of 82.75). The external validation showed that the model is transferable, with the combination of various clinical covariates. Above all, the prediction model based on the GBM showed the highest AUC performance of 84.14 ± 0.015 for the Seoul National University Hospital cohort, yielding in 78.33 in external validation.
CONCLUSIONS: This study showed that readmission prediction models developed using machine-learning techniques and CDM can be a useful tool to compare two hospitals in terms of patient-data features. The Author(s). This is an open access article published by Thieme under the terms of the Creative Commons Attribution-NonDerivative-NonCommercial License, permitting copying and reproduction so long as the original work is given appropriate credit. Contents may not be used for commercial purposes, or adapted, remixed, transformed or built upon. (https://creativecommons.org/licenses/by-nc-nd/4.0/).

Entities:  

Mesh:

Year:  2021        PMID: 34583416      PMCID: PMC8714301          DOI: 10.1055/s-0041-1735166

Source DB:  PubMed          Journal:  Methods Inf Med        ISSN: 0026-1270            Impact factor:   2.176


Background and Significance

While unplanned hospital readmissions are frequent and costly, they are potentially avoidable. 1 Readmission to a hospital shortly after discharge refers to the case where a patient requires retreatment within a short period of time after receiving a particular treatment. 2 Receiving unplanned re-care is based on the premise that there were qualitative problems with the first treatment. The factors behind hospital readmission could be numerous, complex, and interrelated. 3 In 2012, the United States began imposing penalties on hospitals (1% of the hospital's base Medicare inpatient payments 2 4 ) with high 30-day readmission rates for heart failure, acute myocardial infarction, and pneumonia; fines of about $280 million were imposed on 2,213 hospitals in that year. Later, additional penalties were added for chronic lung disease and coronary artery bypass transplantation, and the penalties were increased to 2 and 3%, respectively. Policymakers and medical institutions have proposed several programs to reduce readmission and improve accessibility. 5 Several related studies have investigated patient-related risk factors for hospital readmissions. 6 7 8 9 10 11 12 13 14 Robert and Tamer 15 conducted a predictive study of readmission within 30 days using the LACE index (includes length of stay, acute admission status, Charlson comorbidity index [CCI], and emergency department visits in the past year) and HOSPITAL score (includes hemoglobin at discharge, discharge from oncology service, sodium level at discharge, procedure during index hospitalization, index hospitalization type, number of admissions in the past year, and length of stay) on patients discharged from the Memorial Medical Center from October 2015 to March 2016, with an area under the curve (AUC) value of 0.75 for the HOSPITAL score and an AUC value of 0.58 for the LACE index. 15 Miller et al evaluated the ability to independently predict hospital readmissions within 30 days and compared occupational capabilities with the LACE index to develop predictive tools to identify patients at high risk of unplanned readmissions. 16 Shameer et al demonstrated the potential biomarker sets for readmission probability. 17 Among 1,068 target patients, 178 were readmitted within 30 days (readmission rate 16.66%), and electronic medical record (EMR) data (including diagnostic codes, drugs, laboratory measurements, surgical procedures, and vitals) were extracted and used in multistage modeling using the Naive Bayes algorithm. As a result of their study, compared with the existing readmission prediction model, the EMR-wide prediction model achieved an AUC of 0.78, which was found to be an effective application of data-based machine learning. Although many studies for predicting hospital readmission have been conducted in this manner, the developed prediction models are difficult to apply to other institutions because they were made based on data from a specific single institution. For example, to apply a machine-learning prediction algorithm to data from other hospitals, it is necessary to process data in the form of input values appropriate to the algorithm; in some cases, the program needs to be modified according to the data features of other institutions. Health care data are collected and stored for many purposes, including (1) to directly facilitate research as a form of survey or registry data, or (2) to record the conduct of health care (usually called electronic health record [EHR]), or (3) to manage payments for health care such as claims data. All three are routinely used for clinical research (the latter two as secondary use data), and all three types have their unique content formatting and encoding. 18 The reuse of EHR data for research is a relatively new field and, so far, there has been a lack of awareness regarding the code setting engineering issue. 19 A common data standard can alleviate this need by omitting the extraction step and allowing standardized analytics to be executed on the data in its native environment, that is, the analytics come to the data instead of the other way around. 18 Within the last decade, several common data models (CDMs) have been collaboratively developed for clinical research data. These include the Sentinel CDM, 20 the National Patient-Centered Clinical Research Network CDM, 21 the Health Care Systems Research Network (formerly known as the HMO Research Network) Virtual Data Warehouse, 22 the Observational Medical Outcomes Partnership (OMOP) CDM, 23 and the Clinical Data Interchange Standards Consortium Study Data Tabulation Model. 24 According to the literature, 25 OMOP CDM performs the best, ranking highest in a majority of the evaluation criteria when compared with the other CDM models for EHR-based longitudinal registries based on content coverage, integrity, flexibility, simplicity, integration, and implementability. To maintain and expand OMOP CDM, Observational Health Data Sciences and Informatics (OHDSI) is currently developing open-source tools for data quality and characterization, medical product safety surveillance, comparative effectiveness, quality of care, and patient-level predictive modeling. 25 26 27 28 29 30 Seoul National University Bundang Hospital (SNUBH) and Seoul National University Hospital (SNUH) EHRs have been transformed into OMOP CDM for longitudinal retrospective observation research. Once a database has been converted to OMOP CDM, it can be analyzed using OMOP CDM-based tools. To the best of our knowledge, there has been no study until now that uses CDM data in the development of 30-day unplanned hospital readmission prediction models. In particular, the contribution of this study is to identify effective variables for hospital readmission within 30 days. As a multicenter study, we utilized data from two hospitals to carry out verification with each other and evaluate which hospital's data and models using specific variables are the most suitable models for the topic of readmission prediction. Notably, although many hospital readmission prediction models have been studied, it is difficult to apply the developed model directly to data from other hospitals, and there is a paucity of research exploring clinical variables for patient readmission within 30 days.

Objectives

In this study, we aimed to develop and validate a patient-level prediction model for all-cause 30-day hospital readmission by applying machine-learning methods, with EMR-based clinical data converted to CDM. To the best of our knowledge, none of the studies using OMOP CDM data have explored patient-level prediction models for hospital readmission within 30 days. In this study, clinical variable combinations obtained during the period that can be effective in predicting readmissions within 30 days were considered. Therefore, the research hypothesis of this study is that the predictive model developed using CDM is easy to apply to other institutions and can explore clinical variables that might influence hospital readmission within 30 days. The CDM prediction models showed external validity in terms of model performance as well as transportability.

Methods

Study Data Description

We conducted an observational cohort study using CDM-converted EHR data from two tertiary general university hospitals: SNUH and SNUBH. These study sites have converted EHR data over a 15-year period of more than two million patients, including patient demographics, diagnosis, drug exposures, test orders and their results, surgeries, family histories, and past medical histories, into OMOP CDM. SNUH acts as the parent hospital for SNUBH and thus both use the same EHR system. Evidently, SNUH is considerably larger than SNUBH in terms of size. The parent hospital is located in the center of Seoul, and the child hospital is located in Gyeonggi-do, near Seoul, an hour away from the parent hospital. Patients who visited the hospitals between January 1, 2017 and December 31, 2018 were included in the study. We excluded individuals who died during hospitalization and who visited a hospital for clinical trial. We organized study cohorts for patients living in Seoul or Gyeonggi Province in Korea. Figs. 1 and 2 show the study cohort design in this study.
Fig. 1

SNUBH cohort design of the study. SNUBH, Seoul National University Bundang Hospital.

Fig. 2

SNUH cohort design of the study. SNUH, Seoul National University Hospital.

SNUBH cohort design of the study. SNUBH, Seoul National University Bundang Hospital. SNUH cohort design of the study. SNUH, Seoul National University Hospital. Patients' clinical features were extracted from all diagnosis records prior to the end of the readmission interval, such as gender, age at visit, diagnosis history, medications, and some calculated indices such as CCI (Romano adaptation). All clinical events in OMOP CDM are expressed as concepts, which represent the semantic notion of each event. Concepts are coded as individual concept codes, higher-level concept codes, and groups of higher-level codes based on the level of standard terminologies. Each Standard Concept has a unique domain assignment, which defines which table they are recorded in. For example, signs, symptoms, and diagnosis concepts are in the Condition Domain and are recorded in “CONDITION_CONCEPT_ID” of the “CONDITION_OCCURRENCE CDM” table, which are mainly mapped into SNOMED CT standard terminologies. We used the Feature Extraction package developed by OHDSI as an open-source CDM tool to create features for a cohort. Below is a list of different EHR-driven features with the descriptions of data used in this study. The names of OMOP CDM data tables are written in capital letters. Please refer to the OHDSI GitHub Web site for more details regarding OMOP CDM specifications. 31 Diagnosis and medication information : Records of a person suggesting the presence of a disease or medical condition stated as a diagnosis, a sign, or a symptom, which is observed by a clinician, are contained in the “CONDITION_OCCURRENCE” table. The data table “DRUG_EXPOSURE” in CDM database captures records about exposure to a drug ingested or otherwise introduced into the body. Drugs include prescription and over-the-counter medicines. Individual diagnosis and medication records are extracted from “CONDITION_OCCURRENCE” and “DRUG_EXPOSURE” data tables in the CDM database. The table “CONDITION_ERA” includes data for the span of time the patient is assumed to have had a given condition. This data table contains chronological periods of “CONDITION_OCCURRENCE.” In addition, “CONDITION_GROUP_ERA” is composed of a higher hierarchy concept for the given covariate per CONDITON_ERA table. For instance, “Pneumonia” may be a parent of the different subtypes of “pneumonia” that are found in condition era; thus, a person with different diagnosis subtypes of pneumonia will be counted only once in the “pneumonia” group, while without this grouping, this person would contribute a count to each pneumonia subtype the person was diagnosed with. Likewise, “DRUG_ERA” is defined as a span of time when the patient is assumed to be exposed to a particular active ingredient. Notably, “DRUG_ERA” is not the same as a “DRUG_EXPOSURE”: exposures are individual records corresponding to the source when the drug was delivered to the patient, while successive periods of “DRUG_EXPOSURE” are combined under certain rules to produce continuous drug eras. These diagnoses and medication records are mainly mapped into SNOMED CT and RxNorm standard terminologies. For surgery and clinical examination tests : Patient's surgical record was derived from the “PROCEDURE” table in OMOP CDM. In the case of patient clinical examination test data, test order and its result value of a record were extracted from two tables (“MEASUREMENT” and “OBSERVATION” data tables in the database) according to the type of each test. Visit records : The “VISIT_OCCURRENCE” table includes information about a patient's EHR, either as inpatient, outpatient, or emergency department visits. The number of visits was counted based on the patient's visit type and used as a variable. Other demographic information and clinical scores such as patient gender, age group, and location were extracted from the “PERSON” table. Clinical index scores, such as CCI, diabetes complications severity index, and CHA2DS2-VASc score for estimating the risk of stroke of each individual patient, were also used as model variables. We considered the time boundaries of each feature as follows: (1) long-term covariate combination contains variables acquired during the 365 days prior to discharge date; (2) medium-term setting contains variables included during the 180 days prior to discharge date; and (3) short-term covariate combination contains variables included during the 30 days prior to discharge date.

Unplanned Hospital Readmission

The primary outcome was 30-day unplanned hospital readmission. This was defined as a hospitalization through outpatient visit or emergency room visit in a study period, except planned schedule. We referred to the quality measure of Hospital-Wide All-Cause Unplanned Readmission (HWR) from Centers for Medicare and Medicaid Services (CMS). According to the HWR measure, CMS classifies planned readmissions as planned disease or treatment groups, including chemotherapy, organ transplant, rehabilitation, and other planned treatments or surgeries. We defined scheduled admissions first, and the remaining admissions were assumed to be unscheduled visits. Among the confirmed hospital admissions during the study period, approximately 20% were unplanned. Figure 3 shows the timeline of patient visits during the 30-day hospital revisit period.
Fig. 3

Patient visit timeline based on readmission definition.

Patient visit timeline based on readmission definition.

Model Development and Clinical Covariate Combinations

A patient-level prediction model was iteratively developed and validated using SNUBH and SNUH CDM data. Two types of experiments were performed in this study. Through the first experiment, our intent was to find the effective variable time for predicting readmission within 30 days. To predict rehospitalization within 30 days, clinical variables that occurred 365 days before discharge date (long-term covariate span), 180 days before discharge (medium-term covariate span), and entering predictive models for variables before 30 days (short-term covariate span) were explored in the experiment. In the second experiment, we tested which variables are effective for rehospitalization within 30 days. For diagnoses and drug variables 30 days prior to the discharge date, we designed variables that differ in levels of granularity in each data. Once the prediction model was developed using SNUH data, external evaluation was applied against the SNUBH data. In the opposite case, a model developed with data from SNUBH was evaluated externally using data from SNUH (see Supplementary Fig. S1 , available in the online version only). Variables such as the patient's demographic information and clinical index scores, diagnosis, medications, visit frequency records, surgeries, and clinical examination test records were included. Table 1 summarizes the differences in the combination settings for each variable. For convenience, we describe each variable combination with a name: DM-SC (“diagnosis, medication, surgeries, clinical exam”) and DM (“diagnosis, medication”).
Table 1

Differences in combination settings for each variable

CategoryDM-SCDM 1DM 2DM 3
DemographicsGender, age group, index monthGender, age group, index month, demographics time in cohortGender, age group, index month, demographics time in cohortGender, age group, index month, demographics time in cohort
Clinical index scoreCharlson index, DCSI, Chads2, Chads2VascCharlson index, DCSI, Chads2, Chads2VascCharlson index, DCSI, Chads2VascCharlson index, Chads2Vasc
DiagnosisCondition occurrence, distinct condition countCondition occurrence, distinct condition countCondition era, distinct condition countCondition group era, distinct condition count
MedicationDrug exposure, drug era, distinct ingredient countDrug exposure, distinct ingredient countDrug era, distinct ingredient countDrug group era, distinct ingredient count
Visit recordsTotal count, visit types countVisit types countVisit types countVisit types count
SurgeriesProcedureDistinct procedure countDistinct procedure countDistinct procedure count
Clinical examination testObservationDistinct observation countDistinct observation countDistinct observation count
MeasurementDistinct measurement countDistinct measurement countDistinct measurement count

Abbreviations: DCSI, diabetes complications severity index; DM, diagnosis, medication; DM-SC, diagnosis, medication, surgeries, clinical exam.

Abbreviations: DCSI, diabetes complications severity index; DM, diagnosis, medication; DM-SC, diagnosis, medication, surgeries, clinical exam. DM-SC covariate combination includes all records of a patient's diagnosis, medications, test orders and their results, procedure, or surgical history. Meanwhile, DM combinations 1, 2, and 3 mainly concern patient diagnoses and medication information with different conceptual levels. The lowest granularity reflects individual diagnostic information for each patient visit date (DM 1). Rather, DM 2 allows aggregation of chronic conditions that require frequent ongoing care, instead of treating each diagnostic information as an independent event. With higher granularity of concepts in DM 3, the aggregations of chronic individual diagnostic eras are “grouped” together under the ancestor hierarchy of the condition concept found in the CONDITON_ERA table in CDM. In addition, to examine the differences in the characteristics of readmitted patients, the experiment was conducted by dividing the patients into three groups: all age group, patients 65 years of age and older, and children and adolescents under 18 years of age. We developed machine-learning–based models based on LASSO logistic regression (LR), decision tree (DT), random forest (RF), adaptive boosting (AdaBoost), and gradient boosting machine (GBM) to predict 30-day readmissions and compared their performance. These machine-learning methods were considered owing to their following characteristics. LR is a common and basic algorithm, which is widely used in disease risk prediction and epidemiology. 32 DT is also used in different areas of medical decision making. 33 RF, as an ensemble algorithm of trees, applies a bootstrap algorithm to extract multiple samples from the training set randomly, and trains the samples with the weak classifier. 34 An AdaBoost classifier is a meta-estimator that begins by fitting a classifier on the original dataset and then fits additional copies of the classifier on the same dataset but where the weights of incorrectly classified instances are adjusted such that subsequent classifiers focus more on difficult cases. 35 GBM is a distributed and high-performance gradient lifting framework based on a DT algorithm designed for fast computational time, especially with very large datasets. 36 Patient-level prediction R package developed by the OHDSI was used to train and test the models. Furthermore, 10-fold cross-validation was primarily used for internal validation. To compare the performances of models applied to the prediction of readmission, we used the area under the receiver operating characteristic curve (AUROC) as the primary evaluation criterion. All the models were externally validated to examine their portability.

Ethical Considerations

The study was performed in compliance with the World Medical Association Declaration of Helsinki and Ethical Principles for Medical Research Involving Human Subjects. This study was approved by the SNUBH Institutional Review Board (X-1908–559–901) and SNUH Institutional Review Board (E-2002–002–1097).

Results

Overall, 106,304 index hospitalizations were included in our SNUH study cohort, of which 32,242 resulted in a 30-day readmission ( Table 2 ). Individuals had a mean age of 46.8 years, and slightly more than half were males. The average length of stay was 2.5 days in the entire cohort and 2.9 days in the readmitted cohort. In the SNUBH cohort, 93,553 index hospitalizations were included in our study cohort, of which 26,320 resulted in a 30-day readmission ( Table 3 ). Individuals had a mean age of 46.3 years, and slightly more than half were females. The average length of stay was 2.3 days in the entire cohort and 2.8 days in the readmitted cohort. Individuals in the cohort with a 30-day readmission had markedly different socioeconomic and clinical characteristics compared with those not readmitted.
Table 2

Basic characteristics of SNUH data per visit type

CharacteristicEntire cohortDerived cohortp -Value
ReadmittedNot readmitted
Age, y, mean (SD)46.8 (27.5)49.2 (25.8)45.1 (28.4)
Gender Male, n (%) 50.451.151.10.001
Female, n (%) 49.648.948.9
Age at hospital visit10 under18.514.520.3
10s5.85.95.7
20s5.665.4
30s6.77.36.5
40s7.88.57.5
50s12.514.311.8
60s17.518.517
70s17.116.817.2
80s7.77.27.8
90s0.80.70.9
Season at time of dischargeSpring24.924.826.3
Summer26.326.826.8
Fall24.324.924.0
Winter24.523.425.0
Admission weekdayMonday17.617.517.7
Tuesday15.615.315.7
Wednesday15.415.215.5
Thursday15.115.514.9
Friday11.913.411.3
Saturday9.19.78.9
Sunday15.213.416.0
Average length of stay, mean (SD)2.5 (4.4)2.9 (4.9)2.4 (4.2)
Charlson comorbidity index, mean0.210.380.18

Abbreviations: SD, standard deviation; SNUH, Seoul National University Hospital.

Table 3

Basic characteristics of SNUBH data per visit type

CharacteristicEntire cohortDerived cohortp -Value
ReadmittedNot readmitted
Age, y, mean (SD)46.3 (27.7)49.2 (25.8)45.1 (28.4)
Gender Male, n (%) 49.648.849.90.003
Female, n (%) 50.451.250.1
Age at hospital visit10 under18.813.320.9
10s5.45.35.4
20s4.84.74.9
30s7.47.67.3
40s10.412.69.6
50s13.214.512.6
60s14.815.914.3
70s15.616.415.2
80s8.58.58.5
90s1.11.21.1
Season at time of dischargeSpring25.726.325.5
Summer26.426.826.3
Fall24.224.923.9
Winter23.722.024.4
Admission weekdayMonday16.716.716.7
Tuesday14.815.014.7
Wednesday15.215.415.1
Thursday14.714.714.7
Friday14.015.013.5
Saturday10.310.210.4
Sunday14.313.014.8
Average length of stay, mean (SD)2.3 (4.3)2.8 (4.8)2.1 (4.1)
Charlson comorbidity index, mean0.280.390.26

Abbreviations: SD, standard deviation; SNUBH, Seoul National University Bundang Hospital.

Abbreviations: SD, standard deviation; SNUH, Seoul National University Hospital. Abbreviations: SD, standard deviation; SNUBH, Seoul National University Bundang Hospital. To predict an outcome occurrence considering time boundaries of features, we developed models with data from two hospitals and compared the performance ( Table 4 ). In this experiment, no external verification was performed, only the comparison of the results of the two hospital models for the time variable was performed. Here, each model was developed using SNUH data (mother hospital) and SNUBH data (child hospital).
Table 4

Overall performance on covariate time-boundary settings

ModelLong-term (−365 days)Medium-term (−180 days)Short-term (−30 days)
Train/test SNUHTrain/test SNUBHTrain/test SNUHTrain/test SNUBHTrain/test SNUHTrain/test SNUBH
LASSO logistic regression70.8565.8470.9868.6480.4776.62
Decision tree72.0359.3572.0265.6580.9474.4
Random forest 74.02 65.0574.1370.55 82.75 78.08
AdaBoost73.0364.8573.1667.1081.0175.64
Gradient boosting machine71.11 67.49 75.44 71.07 82.52 79.75

Abbreviations: SNUBH, Seoul National University Bundang Hospital; SNUH, Seoul National University Hospital.

Note: The highest performance in each column is marked in bold.

Abbreviations: SNUBH, Seoul National University Bundang Hospital; SNUH, Seoul National University Hospital. Note: The highest performance in each column is marked in bold. Overall, the performance of the model developed with the data of the parent hospital (SNUH) was high. The model with the best performance in common was GBM, and the LASSO LR and DT showed low performance. Through this experiment, we observed that the effective variable time boundary for readmission within 30 days is to use the clinical variable for short-term covariate span (30 days before discharge). This result was based on variables such as data from surgery, procedures, and clinical examination result, as well as diagnostic records and medication data. The model performance for the short-term variables was the best. It can thus be interpreted that to predict readmission within 30 days, the model developed with data for 30 days based on the discharge date makes the most accurate prediction. The AUROC curve in Fig. 4 shows the results for the best performance described in Table 6 , SNUH-developed model and short-term covariate combination. It can be seen in Fig. 4 that GBM showed the best performance with an AUROC of 84.14, while LASSO LR achieved an AUROC of 80.14.
Fig. 4

Results of the developed model based on SNUH test data with short-term covariates. SNUH, Seoul National University Hospital.

Table 6

Overall performance on SNUH-data-trained classification models

ModelDM 1DM 2DM 3
Train/testSNUHValidationSNUBHTrain/testSNUHValidationSNUBHTrain/testSNUHValidationSNUBH
LASSO logistic regression77.0472.1279.0773.7080.1776.17
Decision tree75.8771.6680.7674.5681.0173.05
Random forest79.3374.9782.36 77.34 82.2477.66
AdaBoost77.1173.3681.0876.4981.2978.14
Gradient boosting machine 80.90 75.94 83.90 76.71 84.14 78.33

Abbreviations: DM, diagnosis, medication; SNUBH, Seoul National University Bundang Hospital; SNUH, Seoul National University Hospital.

Note: The highest performance in each column is marked in bold.

Results of the developed model based on SNUH test data with short-term covariates. SNUH, Seoul National University Hospital. Hence, in our second experiment, we developed models for variables prior to 30 days before discharge and performed external validation of each hospital model by using data such as diagnoses and drugs, which are relatively unlikely to have gaps in term mapping (DM covariates setting). Table 5 shows the results of the SNUBH data-trained model, with external validation (AUROC, with short-term covariates), and Table 6 shows the results of the SNUH data-trained model. In DM 3, the variables were calculated with Group Era (higher concept of period variable) to include the higher concept, while DM 1 contains only individual diagnoses and medications.
Table 5

Overall performance on SNUBH-data-trained classification models

ModelDM 1DM 2DM 3
Train/testSNUBHValidationSNUHTrain/testSNUBHValidationSNUHTrain/testSNUBHValidationSNUH
LASSO logistic regression65.8462.4468.6465.2276.6273.38
Decision tree64.0359.3565.6562.4474.4070.37
Random forest65.05 65.66 70.55 67.87 78.0875.10
AdaBoost64.8561.1367.1064.3975.6473.93
Gradient boosting machine 67.49 63.75 71.07 65.73 79.75 75.34

Abbreviations: DM, diagnosis, medication; SNUBH, Seoul National University Bundang Hospital; SNUH, Seoul National University Hospital.

Note: The highest performance in each column is marked in bold.

Abbreviations: DM, diagnosis, medication; SNUBH, Seoul National University Bundang Hospital; SNUH, Seoul National University Hospital. Note: The highest performance in each column is marked in bold. Abbreviations: DM, diagnosis, medication; SNUBH, Seoul National University Bundang Hospital; SNUH, Seoul National University Hospital. Note: The highest performance in each column is marked in bold. We categorized the entire experimental cohort data into three groups: all age group, patients 65 years of age and older, and children and adolescents under 18 years of age, based on the age at the time of the patient visit. When comparing the performance of the prediction model based on the age of the patient, it was confirmed that the prediction performance was lowest in the group of children and adolescents under 18 years of age ( Fig. 5 ). From this result, it can be assumed that the predictive model we developed was made suitable for predicting elderly patients.
Fig. 5

The result of the GBM model and the AdaBoost model by patient age group. AdaBoost, adaptive boosting; GBM, gradient boosting machine.

The result of the GBM model and the AdaBoost model by patient age group. AdaBoost, adaptive boosting; GBM, gradient boosting machine.

Discussion

In this study, we demonstrated how prediction tools can be integrated generically into two different clinical settings and provide an exemplary use case for predicting 30-day hospital readmission. We compared the performance of models developed using data from two hospitals and compared the prediction performance of readmissions. Conventional patient readmission risk assessments are performed using a variety of assessment tools ranging from multidisciplinary patient interviews to simple screening tools with fewer variables. 8 10 37 38 39 Many of the previously developed readmission prediction models, including the readmission prediction scores mentioned above, account for most of the models based on statistical calculations or LR analysis using several clinical variables. With the development of machine-learning technology, attempts to introduce new machine-learning techniques in predicting readmission are increasing. 12 13 40 41 However, there are insufficient proven cases applicable to actual clinical environments. From this viewpoint, the development of a CDM-based prediction tool has an advantage. The prediction model using CDM is easily transferable and can be applied to various institutions that have CDM data. Because each data element constituting the CDM data is mapped to standard terminologies, it is possible to interpret the common meaning of each institutional model. By comparing the results of the prediction models applied to the two hospitals based on OMOP CDM in this study, we corroborate that the type of patient visit had the greatest influence on the prediction of hospital readmission within 30 days. In the performance comparison experiment of the model made from the data of the parent hospital and the child hospital, there may be various causes such as differences in patient composition of each hospital or differences in term mapping in the process of converting source data to CDM data. In our first experiment, the results were based on variables such as data from surgery, procedures, and clinical examination result, as well as diagnostic records and medication data (DM-SC covariates). The reason for the difference in the performance of the two hospital models is probably due to the distribution of the source data or the mapping of terms that are different for each domain of the CDM. Among them, the prediction performance was the best in the short-term covariates input as patient variables within 30 days from the discharge date. It can thus be interpreted that to predict readmission within 30 days, the model developed with data for 30 days based on the discharge date makes the most accurate prediction. We confirmed how to combine the semantic units of variable data used when constructing the readmission prediction model that shows good results. Data from only two hospitals were used for this study, and there is a limitation in not being able to use data from several hospitals that have OMOP CDM. However, it can be applied to other hospitals as well. Furthermore, with more sophisticated data processing, such as adopting deep learning-based techniques, the model can be expected to perform better. We hope to derive and apply more features of clinical entities or deep learning techniques in future studies.

Conclusions

In this study, predictive models were developed based on CDM that could explore clinical variables to predict hospital readmission within 30 days. As a result, in the model targeting the 30-day prediction, when the data 30 days before the discharge date were used, the prediction performance was the best. In addition, it was confirmed that making a predictive model by creating a variable with data on a high-level concept yields a better performance. The CDM prediction models showed external validity in terms of the model performance as well as transportability. The model developed in this study can be expanded and be used by clinicians in the field.
  29 in total

Review 1.  Clinical code set engineering for reusing EHR data for research: A review.

Authors:  Richard Williams; Evangelos Kontopantelis; Iain Buchan; Niels Peek
Journal:  J Biomed Inform       Date:  2017-04-22       Impact factor: 6.317

2.  Predicting all-cause readmissions using electronic health record data from the entire hospitalization: Model development and comparison.

Authors:  Oanh Kieu Nguyen; Anil N Makam; Christopher Clark; Song Zhang; Bin Xie; Ferdinand Velasco; Ruben Amarasingham; Ethan A Halm
Journal:  J Hosp Med       Date:  2016-02-29       Impact factor: 2.960

3.  Predicting non-elective hospital readmissions: a multi-site study. Department of Veterans Affairs Cooperative Study Group on Primary Care and Readmissions.

Authors:  D M Smith; A Giobbie-Hurder; M Weinberger; E Z Oddone; W G Henderson; D A Asch; C M Ashton; J R Feussner; P Ginier; J M Huey; D M Hynes; L Loo; C E Mengel
Journal:  J Clin Epidemiol       Date:  2000-11       Impact factor: 6.437

4.  PREDICTIVE MODELING OF HOSPITAL READMISSION RATES USING ELECTRONIC MEDICAL RECORD-WIDE MACHINE LEARNING: A CASE-STUDY USING MOUNT SINAI HEART FAILURE COHORT.

Authors:  Khader Shameer; Kipp W Johnson; Alexandre Yahi; Riccardo Miotto; L I Li; Doran Ricks; Jebakumar Jebakaran; Patricia Kovatch; Partho P Sengupta; Sengupta Gelijns; Alan Moskovitz; Bruce Darrow; David L David; Andrew Kasarskis; Nicholas P Tatonetti; Sean Pinney; Joel T Dudley
Journal:  Pac Symp Biocomput       Date:  2017

5.  Evaluating common data models for use with a longitudinal community registry.

Authors:  Maryam Garza; Guilherme Del Fiol; Jessica Tenenbaum; Anita Walden; Meredith Nahm Zozus
Journal:  J Biomed Inform       Date:  2016-10-29       Impact factor: 6.317

6.  Potentially Preventable 30-Day Hospital Readmissions at a Children's Hospital.

Authors:  Sara L Toomey; Alon Peltz; Samuel Loren; Michaela Tracy; Kathryn Williams; Linda Pengeroth; Allison Ste Marie; Sarah Onorato; Mark A Schuster
Journal:  Pediatrics       Date:  2016-08       Impact factor: 7.124

Review 7.  Utility of models to predict 28-day or 30-day unplanned hospital readmissions: an updated systematic review.

Authors:  Huaqiong Zhou; Phillip R Della; Pamela Roberts; Louise Goh; Satvinder S Dhaliwal
Journal:  BMJ Open       Date:  2016-06-27       Impact factor: 2.692

8.  The HOSPITAL score and LACE index as predictors of 30 day readmission in a retrospective study at a university-affiliated community hospital.

Authors:  Robert Robinson; Tamer Hudali
Journal:  PeerJ       Date:  2017-03-29       Impact factor: 2.984

9.  A public-private partnership develops and externally validates a 30-day hospital readmission risk prediction model.

Authors:  Shahid A Choudhry; Jing Li; Darcy Davis; Cole Erdmann; Rishi Sikka; Bharat Sutariya
Journal:  Online J Public Health Inform       Date:  2013-07-01

10.  The HOSPITAL score as a predictor of 30 day readmission in a retrospective study at a university affiliated community hospital.

Authors:  Robert Robinson
Journal:  PeerJ       Date:  2016-09-08       Impact factor: 2.984

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.