Literature DB >> 32548236

Cohort discovery and risk stratification for Alzheimer's disease: an electronic health record-based approach.

Donna Tjandra1, Raymond Q Migrino2,3, Bruno Giordani4, Jenna Wiens1.   

Abstract

BACKGROUND: We sought to leverage data routinely collected in electronic health records (EHRs), with the goal of developing patient risk stratification tools for predicting risk of developing Alzheimer's disease (AD).
METHOD: Using EHR data from the University of Michigan (UM) hospitals and consensus-based diagnoses from the Michigan Alzheimer's Disease Research Center, we developed and validated a cohort discovery tool for identifying patients with AD. Applied to all UM patients, these labels were used to train an EHR-based machine learning model for predicting AD onset within 10 years.
RESULTS: Applied to a test cohort of 1697 UM patients, the model achieved an area under the receiver operating characteristics curve of 0.70 (95% confidence interval = 0.63-0.77). Important predictive factors included cardiovascular factors and laboratory blood testing.
CONCLUSION: Routinely collected EHR data can be used to predict AD onset with modest accuracy. Mining routinely collected data could shed light on early indicators of AD appearance and progression.
© 2020 The Authors. Alzheimer's & Dementia: Translational Research & Clinical Interventions published by Wiley Periodicals, Inc. on behalf of Alzheimer's Association.

Entities:  

Keywords:  cohort discovery; early prediction; electronic health record; machine learning

Year:  2020        PMID: 32548236      PMCID: PMC7293993          DOI: 10.1002/trc2.12035

Source DB:  PubMed          Journal:  Alzheimers Dement (N Y)        ISSN: 2352-8737


INTRODUCTION

Alzheimer's disease (AD), the most common form of dementia, affects approximately 5.8 million Americans, and that number is expected to more than double by 2050.1 The physiological changes in the brain associated with AD, including amyloid beta (Aβ) and tau buildup, are currently suspected to take place at least 20 years before symptom onset. Earlier identification of at‐risk individuals could lead to earlier and more effective treatment. Predictive modeling for AD risk has focused on AD‐specific biomarkers such as cerebrospinal fluid (CSF), neuropsychological test scores, and complex medical imaging. , , , , , , , , , , , , , , These are not routinely collected in clinical care, and thus apply to only a subset of individuals for whom these data are available. Importantly, because collection of these biomarkers can be invasive or involve significant cost/logistics, they are rarely obtained during the pre‐clinical stage, limiting current predictive ability of these biomarkers to short‐term horizons (eg, 2–4 years). , , , , , , , In contrast, we aimed to leverage existing databases of routinely collected electronic health record (EHR) data to develop predictive models for AD that can identify at‐risk individuals up to a decade in advance. EHRs often contain decades of longitudinal clinical data (eg, medications and comorbidities) for thousands of patients. However, these data have been largely underused in studying pre‐clinical signs of AD progression. , , , The ability to automatically identify patients with AD using EHR data would increase the feasibility of downstream computational analyses on large‐scale datasets, without requiring labor‐intensive chart review. To this end, we first developed and validated a cohort discovery tool that can be applied to EHR data for automatic classification of AD individuals. Second, we applied this tool to a large cohort of patients and used machine learning techniques to develop and validate a model for estimating patient risk of developing AD within a 10‐year prediction horizon. Applied more broadly, such an approach could help in identifying risk factors that arise well in advance of clinical symptoms.

METHODS

We describe the inclusion/exclusion criteria that were applied to two datasets to obtain our study cohorts, one for building the cohort discovery tool and another for building the predictive model.

Study cohorts

Our analyses relied on two study cohorts: (1) the cohort discovery tool–cohort and (2) the risk stratification model–cohort. These cohorts were extracted from the Michigan Alzheimer's Disease Research Center (Michigan‐ADRC) and the University of Michigan's Research Data Warehouse (RDW). The Michigan‐ADRC, which focuses on memory and aging research, contains data for 789 participants from ∼2005 to 2019. All participants received a consensus‐based clinical diagnosis using the National Alzheimer's Coordinating Center Uniform Dataset criteria. , The RDW contains records of patient encounters (defined as inpatient and outpatient visits) with Michigan Medicine for more than 4 million patients dating from ∼2000 to 2019. These data consist of all clinical data associated with the encounter (eg, medications). This study was approved by the Institutional Review Board at the University of Michigan.

RESEARCH IN CONTEXT

Systematic review: We searched the literature for reports on predictive modeling and cohort discovery in Alzheimer's disease (AD). Previous research has analyzed data not routinely collected in clinical care, has focused on relatively short prediction horizons (eg, 3 years), or is limited in the scope of electronic health record (EHR) data considered. Interpretation: We developed and validated an EHR‐based cohort discovery tool for AD patients. This tool facilitates analyses of EHR data without requiring manual chart review. Using this tool, we developed and validated an EHR‐based model for predicting AD onset up to 10 years in advance. Covariates associated with the outcome align in part with the AD literature. Novel associations included forms of health‐care use and urine tests. Such findings can be used to stimulate hypothesis generation and/or aid in longitudinal study recruitment. Future directions: Associations identified by our model require further investigation. Model performance could be improved with additional longitudinal data and the inclusion of censored individuals. The first cohort, the cohort discovery tool‐cohort, included all Michigan‐ADRC participants with at least one RDW encounter at or after the age of 65 years. Only this age group was considered, because most cases of AD occur in that population. Our second cohort, the risk stratification model‐cohort, included patients with at least one RDW encounter between the ages of 68 and 72 years who had at least 10 years of follow‐up or who converted to AD within 10 years. This age range allowed for a relatively large study population. We excluded patients with an AD diagnosis before 68 years. Here, AD refers to probable AD, because AD cannot be officially diagnosed until after death and because this diagnosis was commonly used throughout this period.

Cohort discovery tool

Using diagnoses provided by the Michigan‐ADRC, we investigated the accuracy of different EHR‐based rules for identifying AD patients in RDW. Each rule aimed to identify RDW encounters associated with patients with an AD diagnosis and was based on EHR variables related to AD: diagnosis codes for AD, medications for AD, procedure codes for psychological/cognitive testing, and procedure codes involving moderate to high complexity medical decision making (details in Appendix S1 in supporting information). For example, one rule labeled RDW encounters with a current or previous AD diagnosis code and a prescription for an AD‐associated medication as AD . We also evaluated an existing tool from the Phenotype Knowledge Base (PheKB), which labeled patients with at least five encounters with a dementia diagnosis code or prescription for an AD‐associated medication as AD. Applied to a set of encounters in RDW for a patient, the first encounter that met the EHR‐based criteria was labeled as “AD” by the cohort discovery rule. Because AD is currently irreversible, we labeled all subsequent encounters as “AD.” The labels produced by each EHR‐based rule were compared to the Michigan‐ADRC diagnoses at the patient level. Michigan‐ADRC participants are followed longitudinally, and thus may have multiple timestamped diagnoses (eg, cognitively normal, mild cognitive impairment, AD). As ground truth, we labeled the 6 months preceding the first AD diagnosis from the Michigan‐ADRC and anytime thereafter as AD. Prior work has shown that clinical diagnoses of AD have good diagnostic accuracy to histopathology‐confirmed AD. If a patient was never diagnosed with AD, then we considered them “not AD” until 6 months after their last Michigan‐ADRC diagnosis. Using these time frames as ground truth, comparisons to the corresponding RDW encounters were made as follows (Figure 1). Only those whose RDW and ground truth time windows overlapped were included during evaluation. If at least one AD‐diagnosed RDW encounter was within the Michigan‐ADRC‐defined AD window, the patient was considered to have been correctly identified by the EHR‐based rule (true positive). We defined false positives as those with at least one AD‐diagnosed RDW encounter but no Michigan‐ADRC diagnosis for AD within the Michigan‐ADRC‐defined AD time window. True negatives were defined as those not identified by the EHR‐based rule and who never received a Michigan‐ADRC diagnosis for AD, while a false negative had a Michigan‐ADRC diagnosis for AD, but was missed by the EHR‐based rule.
FIGURE 1

Comparing Michigan Alzheimer's Disease Research Center (MIchigan‐ADRC) and Michigan Medicine's Research Data Warehouse (RDW) encounters for a sample patient. Each row represents a timeline for the respective dataset, and encounters are indicated with squares. Shading along the Michigan‐ADRC timeline indicates consensus‐based diagnoses. A true positive is counted if at least one identified Alzheimer's disease (AD) RDW encounter overlaps with the Michigan‐ADRC defined AD window (eg, the encounters in the blue circles)

Comparing Michigan Alzheimer's Disease Research Center (MIchigan‐ADRC) and Michigan Medicine's Research Data Warehouse (RDW) encounters for a sample patient. Each row represents a timeline for the respective dataset, and encounters are indicated with squares. Shading along the Michigan‐ADRC timeline indicates consensus‐based diagnoses. A true positive is counted if at least one identified Alzheimer's disease (AD) RDW encounter overlaps with the Michigan‐ADRC defined AD window (eg, the encounters in the blue circles) Results were summarized by the true positive rate (sensitivity), false positive rate (specificity), positive predictive value (PPV), and F1 score (F1). We measured a population‐adjusted PPV, since the Michigan‐ADRC dataset is enriched compared to the general population (details in Appendix S2 in supporting information). When evaluating EHR‐based rules against each other, we prioritized maximizing the F1 score to balance the population‐adjusted PPV and sensitivity. In the case of ties, we considered the adjusted PPV, unadjusted PPV, specificity, and sensitivity, in that order. Given the rule with highest F1 score, we evaluated when patients received the diagnosis within RDW relative to the Michigan‐ADRC, by measuring the time from the first AD Michigan‐ADRC diagnosis to the first AD‐labeled encounter in RDW. We also examined our ability to identify AD at the encounter level. Using the ground truth labels outlined earlier, a confusion matrix was constructed to show the number of encounters (AD/not AD) that were correctly and incorrectly identified by the EHR‐based rule. Results are reported as the median with an empirical 95% confidence interval (CI), over 1000 bootstrapped samples. Statistical significance relative to the best rule was determined by whether the upper bound of the 95% CI for the F1 score was below the lower bound F1 score of the best rule.

Predictive model

In the following sections, we frame the problem of predicting AD over a 10‐year horizon using EHR‐extracted data. We describe feature engineering, including which EHR components were used, and model training. We then describe model evaluation in terms of predictive performance and influential features.

Outcome

To control for the effect of age on risk of developing AD, we aligned patients in our risk stratification cohort (Section 2.1) based on their earliest visit between 68 and 72 years. Patients were labeled according to the cohort discovery tool (Section 2.2) as converting to AD within 10 years or not. The date of conversion was defined as the date of the first encounter meeting the cohort discovery tool's criteria. Patients were labeled positive if they converted within 10 years of alignment and negative otherwise.

Variable extraction

Given the “alignment visit", each patient was represented by a high‐dimensional feature vector summarizing all encounters in the 1000 days prior to alignment. A look‐back period of 1000 days was chosen based on the median length of available history. We extracted data pertaining to diagnoses (ICD9 [International Classification of Diseases, Ninth Revision] codes), procedures (CPT [current procedural terminology] codes), medications (medication name, ingredient name, and VA [Veterans Affairs] class code), laboratory results (LOINC [Logical Observation Identifiers Names and Codes] and result values), vital sign measurements (eg, temperature), health‐care utilization (eg, encounter types), and demographic information (eg, race). Features were categorized as “time‐invariant” or “time‐dependent.” Time‐invariant features were patient characteristics that do not change over time (eg, race), and time‐dependent features were those associated with a specific encounter or timestamp (eg, diagnoses). Data were pre‐processed with FIDDLE (Flexible Data‐Driven Pipeline), using a time window of 250 days, a pre‐ and post‐filter threshold of 0.0003, and a frequency threshold of 1.0. Feature vectors for each patient were constructed by concatenating their time‐invariant and time‐dependent data corresponding to the 1,000 days prior to alignment.

Model training

Data were split using an 80%–20% training–test random stratified split. Using the training data, we performed model selection. Minimizing the L2‐regularized hinge loss, we trained a linear‐support vector machine to predict AD onset for patients aligned between 68 and 72 years over a 10‐year horizon. The amount of regularization was tuned using five‐fold cross‐validation on the training set, sweeping C = (0.001‐1000) on a logarithmic scale. Analyses were performed in Python 3.6 using SciKitlearn.

Model evaluation

Overall performance of our predictive model was measured using the area under the receiver operating characteristics curve (AUROC) and a confusion matrix measuring sensitivity, specificity, PPV, and accuracy based on a threshold at the 65th percentile on the held‐out test set. We measured model calibration using the Brier score (details in Appendix S3 in supporting information). Additionally, we examined the model's ability to classify AD converters among patients with memory impairments, reporting the AUROC and confusion matrix (details and results in Appendix S9 in supporting information). We report all model evaluation results as empirical 95% confidence intervals generated using 1000 bootstrapped samples unless otherwise stated. We also assessed the model's ability to predict over the 10‐year horizon by examining the number of correctly predicted converters with respect to their time to conversion (time between alignment and first AD diagnosis). Because the model outputs a continuous risk score, we classified patients as “high risk" if their risk score was above the 65th percentile and as low risk otherwise. We examined five non‐overlapping conversion windows, reporting the sensitivity for each. Beyond model performance, we examined which categories of EHR information (eg, diagnoses vs procedures) were the most informative for prediction by comparing the AUROCs on models trained with different subsets of features (eg, training only on diagnosis features or training only on procedural features). We also analyzed the model's most important features using permutation importance, in which any decrease in AUROC was measured by randomly permuting all patient values within a feature or group of correlated features (R≥|0.7|). The most important features were identified as those with the largest drop in AUROC, taken as the median over 100 permutations and whose lower bound on an empirical 95% confidence interval was above zero.

RESULTS

In the following sections, we identify the best EHR‐based rule for cohort discovery. We then summarize performance of the predictive model in terms of AUROC, calibration, and learned risk factors. From 789 Michigan‐ADRC volunteers, 624 (79%) were 65 years and older and had encounters with Michigan Medicine (details in Appendix S4 in supporting information); 24.8% of the 624 volunteers converted to AD. Among several cohort discovery rules (Figure 2), the one that best identified AD patients included those with a diagnosis code for AD (Table S1 in supporting information; median F1‐score = 0.73 [95% CI = 0.68‐0.78], median adjusted PPV = 0.77 [95% CI = 0.71‐0.82], median sensitivity = 0.70 [95% CI = 0.65‐0.74]). The PheKB tool performed significantly worse in terms of median F1‐score = 0.55 (95% CI = 0.48‐0.62, P < .05) and median sensitivity = 0.45 (95% CI = 0.31‐0.51, P < .05).
FIGURE 2

Cohort discovery results. Comparison of results from cohort discovery tools which tested a single electronic health record (EHR) component, were previously published, or whose median F1 score was >0.5. Each color corresponds to the identification tool indicated in the figure legend. Complexity in medical decisions was measured by the amount and variety of patient data examined by a physician, patient risk, and treatment options. A “*” in the figure legend denotes criteria whose F1 score was significantly worse than the best cohort discovery tool

Cohort discovery results. Comparison of results from cohort discovery tools which tested a single electronic health record (EHR) component, were previously published, or whose median F1 score was >0.5. Each color corresponds to the identification tool indicated in the figure legend. Complexity in medical decisions was measured by the amount and variety of patient data examined by a physician, patient risk, and treatment options. A “*” in the figure legend denotes criteria whose F1 score was significantly worse than the best cohort discovery tool Among the true positives identified by our best rule, the first RDW diagnosis occurred on average 177 days before (95% CI = 278 before‐68 days after) the first AD Michigan‐ADRC diagnosis. At the encounter level, this rule yielded a median PPV of 0.59 (95% CI = 0.56‐0.63) and a median sensitivity of 0.82 (95% CI = 0.72‐0.83; details in Appendix S5 in supporting information). Applying the cohort‐discovery rule with the highest F1‐score to RDW (Figure 3) yielded a study population of 8474 patients, of which 4.14% converted to AD within 10 years from alignment (Table 1). FIDDLE extracted 268 time‐invariant features and 3963 time‐dependent features per time window across four time windows (feature breakdown in Appendix S6 in supporting information). The training and test sets consisted of 6777 and 1697 patients, respectively.
FIGURE 3

Applying inclusion/exclusion criteria. We begin with all patients in Michigan Medicine's Research Data Warehouse (RDW). Numbers in each box correspond to the number of patients included/excluded

TABLE 1

Select characteristics of study cohort

Patient demographicsRDW, N = 8,474
Number of encounters per patient pre‐alignment (IQR)11 (4‐25)
Number of encounters per patient post‐alignment (IQR)84 (36‐172)
Female (%)54.94

Obtained from the inclusion/exclusion criteria in Figure 3.

Abbreviations: AD, Alzheimer's disease; IQR, interquartile range; RDW, Michigan Medicine's Research Data Warehouse.

Applying inclusion/exclusion criteria. We begin with all patients in Michigan Medicine's Research Data Warehouse (RDW). Numbers in each box correspond to the number of patients included/excluded Select characteristics of study cohort Obtained from the inclusion/exclusion criteria in Figure 3. Abbreviations: AD, Alzheimer's disease; IQR, interquartile range; RDW, Michigan Medicine's Research Data Warehouse. On the test set, we achieved an AUROC of 0.70 (95% CI = 0.63‐0.77; Figure S2a in supporting information) and a Brier score of 0.028 (95% CI = 0.025‐0.029; Figure S1 in supporting information). Thresholding at the 65th percentile, we achieved a sensitivity of 0.62 (95% CI = 0.60‐0.63), a specificity of 0.66 (95% CI = 0.65‐0.66), and a PPV of 0.07 (95% CI = 0.05‐0.09), for an overall accuracy of 0.66 (95% CI = 0.65‐0.66; Table S5 in supporting information). The model predicted AD onset over long and short prediction horizons with high sensitivity (Figure S3 in supporting information), though performance generally decreased as the prediction horizon increased: 87% patients who converted within 2.5 years of alignment were correctly identified, while the model correctly identified only 53% of those who converted within 8.4 to 10 years of alignment. The distribution of time to conversion was left skewed, with most patients converting >6 years post‐alignment. Overall, data on laboratory test results, procedures, and health‐care utilization had the most predictive power (Figure 4a, Figure S2b). Predicting using laboratory test results alone was able to achieve an AUROC of 0.62 (95% CI = 0.55‐0.69). However, the best performance was achieved when all categories of EHR data were combined. Using longitudinal data from all previous encounters up to 1000 days prior to alignment also improved performance, compared to when data from only the encounter of alignment was used AUROC = 0.54 (95% CI = 0.47‐0.61; Figure 4b). The top 10 important features pertained to health‐care utilization, procedures involving laboratory blood testing, and cardiovascular risk factors (Figure 4c, Table 2), with the median drop in AUROC between 0.002 and 0.040.
FIGURE 4

Comparison of electronic health record (EHR) data contributions. A, Analysis of individual EHR data fields. Comparison of model performance when trained with specific fields of EHR data. In this experiment, all data up to 1000 days prior to alignment were used. Error bars represent 95% confidence intervals. B, Analysis of longitudinal data. Comparison of model performance when trained on information from all encounters up to 1000 days prior to alignment versus training on information from up to 500 days before alignment and information from alignment only. In this experiment, data from all EHR components were used. Error bars represent 95% confidence intervals. The black dashed line represents the receiver operating characteristic curve for random predictions. C, Analysis of individual features. Broad categories in which the features from Table 2 can fall. Number correspond to those found in Table 2

TABLE 2

Important features

Feature groupDescriptionDrop in AUROC (95% CI)
1. Age between 59 and 68

Maximum age between 59 and 68

Age between 59 and 68

0.0400 (0.0251‐0.0675)
2. Visit type – outpatient between 250 and 500 days before alignment

Patient has an outpatient visit

Time between visits is in (0, 2] days

0.0180 (0.0060‐0.0360)
3. Age between 71 and 72

Maximum age between 71 and 72

Age between 71 and 72

0.0070 (0.0015‐0.0161)
4. Religion value NONPatient does not report a religious association0.0047 (0.0015‐0.0128)
5. Laboratorytest

32623‐1 with value in (5.30, 7.4]

21000‐5 with value in (11.099, 12.9]

4544‐3 with value in (16.799, 36.8]

777‐3 with value in (25.999, 190.0]

785‐6 with value in (15.699, 29.5]

786‐4 with value in (29.799, 33.7]

787‐2 with value in (52.499, 86.3]

789‐8 with value in (2.149, 4.09]

between 750 and 1000 days of alignment

Blood measurements of

platelet mean volume

erythrocyte distribution

hematocrit

erythrocyte mean corpuscular hemoglobin

0.0041 (0.0026‐0.0074)
6. Laboratory test

736‐9 with value in (0.399, 16.6]

5905‐5 with value in (0.099, 6.1]

704‐7 with value in (0.000, 0.7]

731‐0 with value in (0.099, 1.1]

742‐7 with value in (0.000, 0.4]

751‐8 with value in (0.099, 3.0]

between 500 and 750 days of alignment

Blood measurements of

lymphocytes

monocytes

basophils

neutrophils

0.0037 (0.0005‐0.0093)
7. Diagnosis code V04.8 along with procedures 9065x and G000x between 250 and 500 days before alignment

Vaccines for influenza, pneumococcal disease

Revision mastoidectomy

Injection of samarium lexidrona

0.0028 (0.0006‐00073)
8. Non‐invasive systolic blood pressure in (127, 136] between 500 and 750 days before alignmentElevated blood pressure/hypertension0.0023 (0.0004‐0.0041)
9. Procedure 8260x and lab test 2132‐9 with value in (89.999, 382.8] between 0 and 250 days before alignment

Measurements of

blood cyanide

vitamin B12

transcobalamin

0.0021 (0.0012‐0.0031)
10. Laboratory test

50557‐8 with value negative

27297‐1 with value negative

50561‐0 with value negative

50563‐6 with value < 1 mg/dl

53327‐3 with value negative

53328‐1 with value negative

57747‐8 with value negative

between 250 and 500 days of alignment

Urine measurements of

ketones

leukocyte esterase

protein

urobilinogen

total bilirubin

glucose

erythrocytes

0.0021 (0.0009‐0.0044)

Summary of the top 10 most important feature groups, as determined by permutation importance. The letter “x” is used to denote any character. Laboratory tests, diagnoses, and procedures are represented as LOINC, ICD9, and CPT codes respectively.

Abbreviations: AUROC, area under the receiver operating characteristics curve; CI, confidence interval; CPT, current procedural terminology; ICD9, International Classification of Diseases, Ninth Revision; LOINC, Logical Observation Identifiers Names and Codes.

Comparison of electronic health record (EHR) data contributions. A, Analysis of individual EHR data fields. Comparison of model performance when trained with specific fields of EHR data. In this experiment, all data up to 1000 days prior to alignment were used. Error bars represent 95% confidence intervals. B, Analysis of longitudinal data. Comparison of model performance when trained on information from all encounters up to 1000 days prior to alignment versus training on information from up to 500 days before alignment and information from alignment only. In this experiment, data from all EHR components were used. Error bars represent 95% confidence intervals. The black dashed line represents the receiver operating characteristic curve for random predictions. C, Analysis of individual features. Broad categories in which the features from Table 2 can fall. Number correspond to those found in Table 2 Important features Maximum age between 59 and 68 Age between 59 and 68 Patient has an outpatient visit Time between visits is in (0, 2] days Maximum age between 71 and 72 Age between 71 and 72 32623‐1 with value in (5.30, 7.4] 21000‐5 with value in (11.099, 12.9] 4544‐3 with value in (16.799, 36.8] 777‐3 with value in (25.999, 190.0] 785‐6 with value in (15.699, 29.5] 786‐4 with value in (29.799, 33.7] 787‐2 with value in (52.499, 86.3] 789‐8 with value in (2.149, 4.09] between 750 and 1000 days of alignment Blood measurements of platelet mean volume erythrocyte distribution hematocrit erythrocyte mean corpuscular hemoglobin 736‐9 with value in (0.399, 16.6] 5905‐5 with value in (0.099, 6.1] 704‐7 with value in (0.000, 0.7] 731‐0 with value in (0.099, 1.1] 742‐7 with value in (0.000, 0.4] 751‐8 with value in (0.099, 3.0] between 500 and 750 days of alignment Blood measurements of lymphocytes monocytes basophils neutrophils Vaccines for influenza, pneumococcal disease Revision mastoidectomy Injection of samarium lexidrona Measurements of blood cyanide vitamin B12 transcobalamin 50557‐8 with value negative 27297‐1 with value negative 50561‐0 with value negative 50563‐6 with value < 1 mg/dl 53327‐3 with value negative 53328‐1 with value negative 57747‐8 with value negative between 250 and 500 days of alignment Urine measurements of ketones leukocyte esterase protein urobilinogen total bilirubin glucose erythrocytes Summary of the top 10 most important feature groups, as determined by permutation importance. The letter “x” is used to denote any character. Laboratory tests, diagnoses, and procedures are represented as LOINC, ICD9, and CPT codes respectively. Abbreviations: AUROC, area under the receiver operating characteristics curve; CI, confidence interval; CPT, current procedural terminology; ICD9, International Classification of Diseases, Ninth Revision; LOINC, Logical Observation Identifiers Names and Codes.

DISCUSSION

Research in predicting AD risk , , , , , , , , , , , , , , has focused on datasets specifically curated for the purpose of studying AD (eg, Alzheimer's Disease Neuroimaging Initiative [ADNI]). While such studies can be used to identify predictors of disease progression, many of the studied variables, for example, CSF composition, are not collected during routine clinical care, especially in the decades before symptom onset. Moreover, because of the costs associated with such data collection, study populations are relatively small (∼1700 patients) and prediction horizons relatively short (2–4 years). In contrast, EHR data consist of routinely collected data, have been collected for over a decade at some institutions, and are available for a large portion of the population, as highlighted by Stephan et al. Given this potential, we sought to explore the utility of EHRs in modeling the progression of AD 10 years before clinical diagnosis. We developed and validated an automated EHR‐based cohort discovery tool for identifying AD patients and then applied this tool to a large cohort of patients aligned between 68 and 72 years. Using these data and machine learning techniques, we developed a model for predicting AD conversion within 10 years. While EHR data have been leveraged to model other conditions, , , , they have been largely underused in modeling AD progression. Most related studies focus on cohort discovery, , , characterizing the incidence of AD, and modeling the risk of dementia more generally while controlling for age to a lesser extent. , We differ from previous work in that we focus on only AD, while prior work has focused on AD and related dementias. We chose to focus on AD alone, because it is the most common form of dementia. Previously proposed identification rules required at least five encounters with a dementia diagnosis code or AD associated medication. On RDW, this rule had a lower F1 score compared to our proposed rule. In addition, we differ from previous risk stratification models in that we consider AD specifically, , use a 10‐year horizon instead of 5 years or less, , and focus on a broader set of input covariates or potential risk factors. , , We also control for age to a larger extent, as it has been demonstrated that previous models performed similarly to predicting based on age alone. , Compared to curated datasets like ADNI, EHR data present additional challenges. In the context of AD, EHRs do not have a set of ground truth diagnoses. We relied on the fact that a subset of individuals in RDW were also volunteers in the Michigan‐ADRC for whom we had ground truth diagnoses. In addition, data from prospective studies such as ADNI are collected at fixed time intervals, while EHR data are irregularly sampled. Despite these challenges, there are many advantages in working with EHR data. First, EHR data may contain more longitudinal data per patient than ADNI. For example, 25% of ADNI participants had >10 encounters, compared to more than 50% in our study population. This allowed us to predict AD onset over longer horizons (10 years) with modest performance. Approximately half of the patients who converted between 8.4 and 10 years after alignment were correctly identified, demonstrating the possibility of early detection. The ability to predict over longer horizons could be crucial, as the physiological changes in the brain are suspected to take place at least 20 years before symptom onset. Over time, as more EHR data are collected, we may be able to improve model performance and investigate longer time horizons. Second, study populations from ADNI are highly enriched with AD individuals and AD‐specific data, while EHR‐derived study populations are more likely to represent the general population and the types of data routinely available. We identified laboratory tests and procedures associated with AD onset up to 10 years in advance. While identification of EHR variables known to be associated with AD for model building is useful, EHR variables with no known association to AD could lead to the discovery of unknown biological mechanisms, interactions, and novel biomarkers. Similarly, an EHR‐based predictive tool may be used in a cost‐effective strategy to screen which at‐risk patients should undergo earlier testing using more invasive (eg, CSF fluid) or imaging‐based established biomarkers. Many of the features identified as important matched previous findings in the literature. In particular, features related to health‐care use appeared to be strong predictors, in line with previous work that has reported an increase in health‐care use one year prior to AD diagnosis. , In addition, many of the important features related to laboratory blood tests have been previously associated with AD. Specifically, Chen et al. and Winchester et al. found that changes in blood cell composition may be associated with AD development. , Wang et al. found an association between vitamin B12 and AD development. In line with Cao et al. and Le Page et al., we identified immune system biomarkers as beneficial in early detection. , In terms of comorbidities we identified as associated with increased risk, hypertension has previously been identified. In addition, urine tests are associated with diabetes testing, another related risk factor. In terms of procedures, mastoid procedures could act as a possible surrogate for hearing loss, which has been suspected to be associated with AD. Finally, the receipt of vaccinations may be indiciative of an overall poorer state of health, increasing susceptibility to infection and disease. Importantly, all of the predictive factors identified in our retrospective analysis are merely associations and not necessarily indicative of a causal relationship. Our study is not without limitation. We relied on imperfect labels from our cohort discovery tool. As a result, the model may not generalize to predicting the full spectrum of patients that convert to AD. In addition, inaccuracies in labeling the date of AD onset may introduce additional noise. Another limitation stems from our decision to exclude censored patients. We excluded censored patients because they did not have sufficient follow‐up to assign a label. Going forward, approaches for incorporating censored patients could increase the size of the study population. Furthermore, although we aligned patients between 68 and 72 years to control for the effects of age on our prediction task, age appeared as an important predictor. Though aligning patients at a single age (eg, 68 years) could have mitigated this effect, this ultimately would have decreased the size of the study population. In summary, we demonstrated the potential for EHRs as a novel source of data for developing models that characterize AD progression. Going forward, such analyses could be applied to other EHRs to generate hypotheses regarding novel early predictors and mechanisms of AD. In addition, longitudinal clinical studies involving early interventions may selectively target recruitment efforts toward “at‐risk" patients well before symptom onset.

CONFLICTS OF INTEREST

The authors have no conflicts of interest to report. Supporting Information Click here for additional data file.
  38 in total

Review 1.  Cardiovascular risk factors and dementia.

Authors:  Howard Fillit; David T Nash; Tatjana Rundek; Andrea Zuckerman
Journal:  Am J Geriatr Pharmacother       Date:  2008-06

2.  Predicting diagnosis and cognition with 18F-AV-1451 tau PET and structural MRI in Alzheimer's disease.

Authors:  Niklas Mattsson; Philip S Insel; Michael Donohue; Jonas Jögi; Rik Ossenkoppele; Tomas Olsson; Michael Schöll; Ruben Smith; Oskar Hansson
Journal:  Alzheimers Dement       Date:  2019-01-25       Impact factor: 21.566

3.  Multi-Layer Multi-View Classification for Alzheimer's Disease Diagnosis.

Authors:  Changqing Zhang; Ehsan Adeli; Tao Zhou; Xiaobo Chen; Dinggang Shen
Journal:  Proc Conf AAAI Artif Intell       Date:  2018-02

4.  A Clinically-Translatable Machine Learning Algorithm for the Prediction of Alzheimer's Disease Conversion in Individuals with Mild and Premild Cognitive Impairment.

Authors:  Massimiliano Grassi; Giampaolo Perna; Daniela Caldirola; Koen Schruers; Ranjan Duara; David A Loewenstein
Journal:  J Alzheimers Dis       Date:  2018       Impact factor: 4.472

Review 5.  Dementia risk prediction in the population: are screening models accurate?

Authors:  Blossom C M Stephan; Tobias Kurth; Fiona E Matthews; Carol Brayne; Carole Dufouil
Journal:  Nat Rev Neurol       Date:  2010-05-25       Impact factor: 42.937

6.  Diabetes mellitus and risk of Alzheimer's disease and dementia with stroke in a multiethnic cohort.

Authors:  J A Luchsinger; M X Tang; Y Stern; S Shea; R Mayeux
Journal:  Am J Epidemiol       Date:  2001-10-01       Impact factor: 4.897

7.  Benefits and drawbacks of electronic health record systems.

Authors:  Nir Menachemi; Taleah H Collum
Journal:  Risk Manag Healthc Policy       Date:  2011-05-11

8.  Accurate multimodal probabilistic prediction of conversion to Alzheimer's disease in patients with mild cognitive impairment.

Authors:  Jonathan Young; Marc Modat; Manuel J Cardoso; Alex Mendelson; Dave Cash; Sebastien Ourselin
Journal:  Neuroimage Clin       Date:  2013-05-19       Impact factor: 4.881

9.  Altered peripheral profile of blood cells in Alzheimer disease: A hospital-based case-control study.

Authors:  Si-Han Chen; Xian-Le Bu; Wang-Sheng Jin; Lin-Lin Shen; Jun Wang; Zheng-Qian Zhuang; Tao Zhang; Fan Zeng; Xiu-Qing Yao; Hua-Dong Zhou; Yan-Jiang Wang
Journal:  Medicine (Baltimore)       Date:  2017-05       Impact factor: 1.889

10.  Multimodal and Multiscale Deep Neural Networks for the Early Diagnosis of Alzheimer's Disease using structural MR and FDG-PET images.

Authors:  Donghuan Lu; Karteek Popuri; Gavin Weiguang Ding; Rakesh Balachandar; Mirza Faisal Beg
Journal:  Sci Rep       Date:  2018-04-09       Impact factor: 4.379

View more
  2 in total

Review 1.  The medicinal chemistry of mitochondrial dysfunction: a critical overview of efforts to modulate mitochondrial health.

Authors:  Maximillian Taro William Lee; William Mahy; Mark David Rackham
Journal:  RSC Med Chem       Date:  2021-06-04

2.  Use of blood pressure measurements extracted from the electronic health record in predicting Alzheimer's disease: A retrospective cohort study at two medical centers.

Authors:  Donna Tjandra; Raymond Q Migrino; Bruno Giordani; Jenna Wiens
Journal:  Alzheimers Dement       Date:  2022-04-16       Impact factor: 16.655

  2 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.