Literature DB >> 35271664

Utility of machine learning in developing a predictive model for early-age-onset colorectal neoplasia using electronic health records.

Hisham Hussan1,2, Jing Zhao3, Abraham K Badu-Tawiah1,4,5, Peter Stanich1, Fred Tabung2,6, Darrell Gray1,2, Qin Ma3, Matthew Kalady2,7, Steven K Clinton2,6.   

Abstract

BACKGROUND AND AIMS: The incidence of colorectal cancer (CRC) is increasing in adults younger than 50, and early screening remains challenging due to cost and under-utilization. To identify individuals aged 35-50 years who may benefit from early screening, we developed a prediction model using machine learning and electronic health record (EHR)-derived factors.
METHODS: We enrolled 3,116 adults aged 35-50 at average-risk for CRC and underwent colonoscopy between 2017-2020 at a single center. Prediction outcomes were (1) CRC and (2) CRC or high-risk polyps. We derived our predictors from EHRs (e.g., demographics, obesity, laboratory values, medications, and zip code-derived factors). We constructed four machine learning-based models using a training set (random sample of 70% of participants): regularized discriminant analysis, random forest, neural network, and gradient boosting decision tree. In the testing set (remaining 30% of participants), we measured predictive performance by comparing C-statistics to a reference model (logistic regression).
RESULTS: The study sample was 55.1% female, 32.8% non-white, and included 16 (0.05%) CRC cases and 478 (15.3%) cases of CRC or high-risk polyps. All machine learning models predicted CRC with higher discriminative ability compared to the reference model [e.g., C-statistics (95%CI); neural network: 0.75 (0.48-1.00) vs. reference: 0.43 (0.18-0.67); P = 0.07] Furthermore, all machine learning approaches, except for gradient boosting, predicted CRC or high-risk polyps significantly better than the reference model [e.g., C-statistics (95%CI); regularized discriminant analysis: 0.64 (0.59-0.69) vs. reference: 0.55 (0.50-0.59); P<0.0015]. The most important predictive variables in the regularized discriminant analysis model for CRC or high-risk polyps were income per zip code, the colonoscopy indication, and body mass index quartiles. DISCUSSION: Machine learning can predict CRC risk in adults aged 35-50 using EHR with improved discrimination. Further development of our model is needed, followed by validation in a primary-care setting, before clinical application.

Entities:  

Mesh:

Year:  2022        PMID: 35271664      PMCID: PMC9064446          DOI: 10.1371/journal.pone.0265209

Source DB:  PubMed          Journal:  PLoS One        ISSN: 1932-6203            Impact factor:   3.752


Introduction

Colorectal cancer (CRC) is the most gastrointestinal cancer, affecting over 150,000 adults in the U.S. each year. Despite a declining CRC incidence and mortality in older adults due to effective screening, CRC incidence and mortality is rising in adults ≤50 years of age [1-3]. The duration of preclinical CRC is estimated to be between 4 and 6 years [4]. Thus, adults may harbor asymptomatic CRC for years before undergoing CRC screening [5]. Although multiple professional societies now recommend initiating CRC screening at age 45 as opposed to 50, simulations raise concerns about cost, risks, and efficacy even when fecal immunochemical testing (FIT) is used [6]. Thus, there is an urgent need to establish novel and targeted CRC screening strategies for young adults that are cost-effective and easy to implement. Such efforts are challenged by the perceived lower risk among young adults and medical providers, even when gastrointestinal symptoms are present [7-9]. One potential strategy for the early detection of CRC and premalignant polyps, is to apply evidence-based risk stratification tools to identify individuals at greater risk who can benefit from screening. Such efforts may reduce diagnostic delays for young adults, particularly when leveraging the power of electronic medical records to alert caregivers. Novel risk assessment tools are being developed and validated for other malignancies and applied to clinical practice so as to improve care with acceptable costs [10]. However, thus far, the available CRC risk assessment tools focus on asymptomatic adults over the age of 50 and do not capture adults aged 35–44 who account for 50% of early-onset CRC cases [11, 12]. Existing CRC prediction tools also lack discriminatory power or are cumbersome to use, which has reduced their utilization and dissemination [13]. Therefore, developing a sensitive and specific, and yet easy to implement, CRC risk assessment tool for adults aged 35–50 is necessary to classify young adults into meaningful risk groups so as to identify those at high risk, while reducing interventions such as colonoscopy, in those at low risk. In that regard, machine learning is an aspect of artificial intelligence that uses software algorithms to improve the analysis by learning and identifying patterns in large datasets [14]. Therefore, Incorporation of machine learning offers potential for the development of an effective CRC risk assessment tool for young adults. For instance, machine learning methods that integrate clinical risk factors have been applied to breast cancer risk prediction and improve predictive accuracy from 60% to 90% [15]. In addition, deep learning with an artificial neural network based on personal health data has been shown to robustly stratify CRC risk in the large national database [16]. Thus, we hypothesize that machine learning can integrate readily available and complex factors from electronic health records (EHRs) to create a prediction model for CRC that applies to adults aged 35–50. To test our hypothesis, we derived and internally validated a prediction model for CRC or high-risk polyps in adults aged 35–50 years who underwent colonoscopy due to symptoms or screening age indications.

Methods

Participants

We conducted a retrospective predictive study at the Ohio State University after receiving approval from the Ohio State University’s (OSU) Institutional Review Board (IRB protocol number 2020H0190). The Ohio State University IRB approved the waiver of informed consent since this is a retrospective chart review that involves no interaction with study participants; and the study accessed information which would normally be accessed during clinical care for these patients. To develop the model, we used data from average-risk adults aged 35–50 who underwent their first colonoscopy between November 2017 and February 2020. Our cohort and their colonoscopy data were obtained from the GI Quality Improvement Consortium (GIQuic) database at the Ohio State University. GIQuic is a collaborative, nonprofit, scientific organization between the American College of Gastroenterology and the American Society for Gastrointestinal Endoscopy [17]. The OSU GIQuIC database collects all the colonoscopies performed at OSU and includes colonoscopy quality measures such as adequacy of bowel preparation, indication, cecal intubation rate, and adenoma detection rate [18].

Inclusion and exclusion criteria

Our study plot is included in Fig 1. We included adults aged 35–50 years due to the increased incidence of early-onset CRC in adults aged 35–49 years compared to younger adults [1]. Moreover, a substantial portion of asymptomatic early-onset CRC patients are not diagnosed until the initiation of screening at age 50 [5]. Some health plans previously approved CRC screening colonoscopy in adults aged 45 and older according with the American Cancer Society (ACS) 2018 guidelines [19]. However all patients <45 are generally referred for a diagnostic colonoscopy (e.g., diarrhea, constipation, abdominal pain, irritable bowel syndrome, bleeding, etc). These gastrointestinal symptoms are found in a significant proportion of Americans, most of whom do not undergo diagnostic colonoscopy or have no organic causes on a colonoscopy [20-23]. Therefore, we included adults who underwent either a diagnostic or screening colonoscopy, which is the standard of care for the diagnosis of polyps or CRC. We further investigated if including diagnostic colonoscopy may lead to possible bias by comparing our predictors between adults aged 46–49 undergoing screening colonoscopy to vs. diagnostic colonoscopy (S1 Table). We selected 46–49 because the numbers of diagnostic and screening colonoscopies were similar in this age range (407 vs. 296, respectively). Only diagnostic symptoms, tobacco use, and triglyceride levels differed significantly, suggesting symptomatic and asymptomatic adults are similar for most of the predictors included in this study. All included adults had a complete colonoscopy and an adequate bowel prep for detection of polyps >5 mm (Boston Bowel Prep Scale ≥2 in every colon segment) [24]. As early-onset CRC primarily occurs in adults with no strong familial predisposition or pre-existing colitis [25], we included only average-risk adults in our model. Of the 5,588 participants considered for the study, we excluded patients with: (1) inflammatory bowel disease or colitis on subsequent biopsies; (2) personal history of polyps or CRC, elevated cancer makers (e.g., CEA or CA199) or metastatic cancer requiring colonoscopy; (3) family history of CRC in one first degree or two second degree relatives, or (4) hereditary CRC syndromes including polyposis syndromes. After these exclusions, we retained 3,116 participants. This study was conducted and reported in accordance with the guidelines for transparent reporting of a multivariable prediction model for individual prognosis or diagnosis (TRIPOD) [26].
Fig 1

Study plot detailing study flow as well as inclusion and exclusion criteria.

Outcomes

Our two outcomes were (1) CRC and (2) CRC or high-risk polyps confirmed by pathology. Our outcomes were recorded by a personnel who was blinded to the study patient clinical characteristics and laboratory measurements. High-risk polyps were defined as adenomatous or serrated polyps ≥10 mm in size with high grade dysplasia or villous component or ≥3 adenomas or serrated polyps of any size as done before [27, 28]. We did not include hyperplastic polyps in our outcomes due to benign nature. Because the rate of early-onset CRC is low (9.5–14 per 100,000) [29], we included high-risk polyps as an outcome to increase the pre-test probability of a positive test. We chose to include only high-risk polyps as 91% of proven high-risk polyps grow by 20% per year, whereas 63% of non-high-risk polyps remain stable or regress [30].

Predictors

The predictors for machine learning models were chosen from routinely available data in the EHRs using a priori knowledge. Our predictors were pulled from EHRs by a data analyst who was blinded to the study and our outcomes. The predictors included in this study are summarized in Table 1. Predictors included patient age at time of colonoscopy, reported sex and race; American Society of Anesthesiology (ASA) comorbidity category; the symptom indicating the colonoscopy; recorded height and weight; calculated body mass index (BMI) classified by mean, quartile, and overweight/obesity category per WHO criteria; social history (use of alcohol, tobacco, or recreational drugs); medication (aspirin and statins); and laboratory studies (hemoglobin and cholesterol panels). We used the ratio of triglyceride to high-density lipoprotein (HDL) as a surrogate of insulin resistance and stratified the ratio by mean, quartile, and <3 or ≥3 as previously reported [31, 32]. Zip code-derived social determinants of health were retrieved as well. Specifically, we used the 2018 IRS public data to link zip codes to mean adjusted gross income, adjusted gross income percentage within income brackets, and single return percentile [33] for inclusion in our model. Rural-urban commuting area (RUCA) codes, which are a detailed and flexible scheme for delineating sub-county components of rural and urban areas, were also linked to zip codes and included as predictors in our model [34].
Table 1

Included predictors and baseline demographics.

PredictorsPercentages and meansMissing data
Total number of included patients3116
Mean age (standard deviation or S.D.)46.5 (4.73)0.0%
Female Gender55.1%0.0%
Race1.8%
    Non-Hispanic White67.2%
    African American18.7%
    Hispanic2.6%
    Asian1.5%
    Other8.1%
Rural-urban commuting area code (RUCA2)0.0%
    Mean (S.D.)1.33 (1.17)
    Metropolitan [RUCA 1–3]93.3%
    Micropolitan [RUCA 4–6]5.0%
    Small town [RUCA 7–9]1.4%
    Rural [RUCA 10]0.3%
Percentage of returns within income brackets per zip code [mean (S.D.)]0.4%
    $1 to under $25,00030.60% (9.48)
    $25,000 to under $50,00024.59% (6.98)
    $50,000 to under $75,00014.64% (2.63)
    $75,000 to under $100,0009.50% (2.83)
    $100,000 to under $200,00014.70% (7.69)
    $200,000 or more5.97% (5.88)
Percentage of single tax returns per Zip code [mean (S.D.)]49.04% (8.67)0.4%
Adjusted gross income per zip code [mean (S.D.)]$1,399,101.15 (862,516.76)0.4%
American Society of Anesthesiology (ASA) Physical Status Classification System0.0%
    ASA I (healthy patient)25.6%
    ASA II (mild systemic disease)67.4%
    ASA III (severe systemic disease)6.9%
    ASA IV (life threatening systemic disease)0.1%
Colorectal cancer screening indication53.0%0.0%
All diagnostic colonoscopy indications47.0%0.0%
    Functional gastrointestinal symptoms:32.8%
        •Abdominal pain11.3%
        •Constipation5.9%
        •Diarrhea3.3%
        •Rectal pain0.6%
        •Pelvic pain0.3%
        •Obstipation0.1%
        •Irritable bowel syndrome0.3%
    Weight loss1.0%
    Gastrointestinal bleeding20.1%
    Anemia3.8%
    Change in bowel habits2.8%
    Change in stool caliber0.7%
    Personal history of cancer other than CRC0.4%
    Colorectal neoplasm in distant relative3.0%
    Family history of cancer other than CRC0.0%
    Prior diverticulitis prior diverticulitis2.1%
Height in feet [mean (S.D.)]5.59 (0.34)0.7%
Weight in pounds [mean (S.D.)]194.58 (53.11)6.0%
BMI (kg/m2)6.0%
    Mean (S.D.)30.24 (7.42)
    ≥ 25 Kg/m270.7%
    ≥ 30 Kg/m240.5%
    ≥ 35 Kg/m219.4%
    ≥ 40 Kg/m29.4%
    Median [Inter quartile Range (IQR)]28.8 (25–33.9)
Alcohol use0.9%
    Never1.1%
    No33.8%
    Not currently3.0%
    Yes61.2%
Tobacco use0.4%
    Never61.9%
    Passive0.2%
    Quit23.1%
    Yes14.3%
Intravenous drug user1.4%
    No98.4%
    Yes0.2%
Illicit drug user1.4%
    Never5.5%
    No84.3%
    Not currently2.0%
    Yes6.8%
Total cholesterol (mg/dL)29.2%
    Mean (S.D.)185.57 (41.11)
    ≥ 200 mg/dL23.9%
    < 200 mg/dL46.8%
    ≥ 170 mg/dL45.3%
    < 170 mg/dL25.5%
    Median (IQR)183 (159–210)
High Density Lipoprotein (HDL, mg/dL)29.9%
    Mean (S.D.)51.90 (15.82)
    ≥35 mg/dL63.5%
    <35 mg/dL6.6%
    ≥40 mg/dL55.4%
    <40 mg/dL14.7%
    Median (IQR)49 (41–60)
Low Density Lipoprotein (LDL, mg/dL)30.3%
    Mean (S.D.)107.01 (34.45)
    ≥100 mg/dL40.3%
    <100 mg/dL29.4%
    ≥150 mg/dL7.2%
    <150 mg/dL62.5%
    Median (IQR)106 (84–129)
Triglyceride (TG, mg/dL)29.4%
    Mean (S.D.)143.74 (176.88)
    ≥150 mg/dL21.9%
    <150 mg/dL48.7%
    Median (IQR)110 (76–167)
Triglyceride: High Density Lipoprotein (TG: HDL) ratio29.9%
    Mean (S.D.)3.31 (5.92)
    High (ratio ≥3)25.0%
    Low (ratio <3)45.1%
    Median (IQR)2.24 (1.35–3.77)
Hemoglobin (mg/dL)40.3%
    Mean (S.D.)13.75 (1.71)
    Females with anemia (<12 mg/dL)6.3%
    Males with anemia (<13.5 mg/dL)3.2%
    Median (IQR)13.9 (12.8–14.9)
Reported non-steroidal anti-inflammatory drugs use12.5%0.0%
Statin medications use14.3%0.0%

Data preparation

Dummy variables were created by converting categorical variables to corresponding as many numerical variables as there are categories. Then, training and testing datasets were generated by randomly splitting the data into 70% and 30%, respectively, using the createDataPartition function in the R package caret. Missing values were imputed by creating a bag imputation model and using the imputation model to predict the values of missing data points. Imputation via bagging fits a bagged tree model for each predictor as a function of all other predictors. Finally, we centered, scaled, and transformed the predictor values using the R function preProcess in the caret package to generate comparable continuous predictors with dynamic ranges.

Predictive modeling

To predict the probability of each outcome, we first fit a logistic regression model as the reference model including all of the aforementioned predictors. Then four machine learning models were constructed: (1) regularized discriminant analysis, (2) random forest, (3) neural network, and (4) gradient boosting decision tree. Regularized discriminant analysis is a generalization of linear discriminant analysis and quadratic discriminant analysis that increases the power of discriminant analysis to penalize large coefficients from small sample sizes. For our regularized discriminant analysis model, we performed a random search for two parameters (gamma and lambda) using the R package klaR [35]. Random forests are a bagging approach derived from many decision trees and are created with bootstrap samples of training data and random feature selection. For our random forest model, we used random search in the randomForest package to generate 20 random values of mtry and selected the value with the highest accuracy [36]. Neural networks are computational learning systems that use a network of functions to understand and translate a data input of one form into a desired output. For our neural network model, we performed a random search for two hyper-parameters (size and decay) using the R package nnet [37]. Gradient boosting decision trees are a boosting approach that builds an additive model of decision trees estimated by gradient descent. For our gradient boosting decision tree model, we applied a random search to tune parameters, number of iterations, and interaction depth while holding shrinkage constant in the R package gbm [38]. Finally, to account for potential overfitting of the machine learning models, we employed repeated five-fold cross-validation in the R package caret [34]. We assessed the predictive performance of each model by computing C-statistics [area under the receiver operating characteristic curve (AUC/AUROC)] and prediction metrics, including sensitivity, specificity, positive predictive value, and negative predictive value, in the testing dataset using functions provided in the R package pROC [39]. To account for the class imbalance caused by a low proportion of outcomes, we applied the SMOTE resampling method to generate artificial samples [40] and selected cutoffs based on the ROC curve. To evaluate the contribution of each predictor to the machine learning models, we calculated variable importance in the best performing models. Finally, we used the DeLong test to compare ROC curves to the reference model using the function roc.test in the R package pROC [39]. A p-value <0.05 was considered statistically significant. All analyses were performed with R version 4.0.2 (The R Foundation for Statistical Computing).

Results

Altogether, 3,116 adults aged 35–50 were included in our study. The characteristics of the participants are described in Table 1. The cohort was 55.1% female, 32.8% non-white, and 93.3% belonged to a metropolitan area per RUCA classifications. Approximately 54% of the cohort belonged to zipcode that earned less than $50,000 a year, and more than two-thirds (72%) of the cohort were overweight or obese. A screening colonoscopy was performed in 53% of participants, and functional gastrointestinal symptoms were the main indication for a diagnostic colonoscopy (32.8%).

Prediction of colorectal cancer

Overall, 16 (0.05%) patients had CRC on colonoscopy. The C-statistics for CRC are presented as ROC curves in Fig 2 and comparisons of AUC characteristics in Fig 3. The reference model had the lowest discriminative ability when compared to the machine learning models. For example, the neural network and gradient boosting decision tree models had higher AUC values compared to the reference model but did not reach significance (neural network: 0.75; 95%CI, 0.48–1.00; P = 0.07; stochastic gradient boosting: 0.76; 95%CI, 0.46–1.00; P = 0.14). The performance metrics for all models are detailed in Table 2. All machine learning models had similar sensitivity to the reference model except regularized discriminant analysis, which had lower sensitivity. The specificity and accuracy were much higher for machine learning models compared to the reference model (e.g., specificity of 96% and accuracy of 96% for regularized discriminant analysis vs. 10% and 11%, respectively, for the reference model). The balanced accuracy, a better metric for a rare event like CRC, was higher in the machine learning models compared to the reference model (e.g., 73% for the neural network vs. 42% for the reference). Due to the low incidence of CRC, the positive predictive value was low in all models (maximum of 5% for regularized discriminant analysis), and the negative predictive value was 99% in all models.
Fig 2

Receiver Operator Curves (ROC) of the reference and machine learning models in the test set for colorectal cancer (CRC) and CRC or high-risk polyps (bottom).

Fig 3

Area Under the Curve (AUC) of our reference and machine learning models in the test set for colorectal cancer (CRC) and CRC or high-risk polyps.

The p value compares the machine learning models to the reference model using the DeLong test.

Table 2

Performance metrics of different prediction models.

Colorectal cancer
Metric/ModelLogistic regressionRegularized discriminantRandom forestNeural networkStochastic gradient boosting
    Accuracy0.110.960.660.710.86
    Accuracy (lower)0.090.950.620.680.84
    Accuracy (upper)0.130.970.690.740.88
    Balanced accuracy0.420.730.700.730.80
    Sensitivity0.750.500.750.750.75
    Specificity0.100.960.650.710.86
    Positive predictive value0.000.0500.010.02
    Negative predictive value0.990.990.990.990.99
Colorectal cancer or high-risk polyps
    Accuracy0.520.610.610.620.54
    Accuracy (lower)0.490.580.580.590.51
    Accuracy (upper)0.550.640.650.650.58
    Balanced accuracy0.540.60.590.600.54
    Sensitivity0.580.570.550.560.54
    Specificity0.510.620.630.630.55
    Positive predictive value0.170.210.210.210.18
    Negative predictive value0.870.890.880.890.87
Receiver Operator Curves (ROC) of the reference and machine learning models in the test set for colorectal cancer (CRC) and CRC or high-risk polyps (bottom).

Area Under the Curve (AUC) of our reference and machine learning models in the test set for colorectal cancer (CRC) and CRC or high-risk polyps.

The p value compares the machine learning models to the reference model using the DeLong test.

Prediction of colorectal cancer or high-risk polyps

There were 478 (15.34%) participants with CRC or high-risk polyps in our cohort. The C-statistics are described in Figs 2 and 3. All machine learning approaches had a significantly higher ability to predict CRC or high-risk polyps, except for gradient boosting decision tree. For example, the regularized discriminant analysis model had an AUC of 0.64 (95%CI, 0.59–0.69), whereas the reference model had an AUC of 0.55 (95%CI, 0.5–0.6; P<0.0015). Compared with the reference model, all machine learning models had comparable sensitivity and slightly higher specificity, accuracy, and balanced accuracy (Table 2). The positive predictive value was higher in the machine learning models compared to the reference model (e.g., 0.27 for the regularized discriminant analysis model vs. 0.17 for the reference model). The negative predictive value was comparable in all models with a maximum of 0.89 in the regularized discriminant analysis and neural network models.

Variable importance

The importance of variables in predicting the risk of CRC and CRC or high-risk polyps is demonstrated in Fig 4. These variables were calculated in the best performing models based on our AUC comparisons to reference illustrated in Fig 3: The neural network for CRC and regularized discriminant analysis for CRC or high-risk polyps. The leading predictors of CRC in the neural network model were ASA comorbidity category, HDL quartile, gastrointestinal bleeding as an indication for diagnostic colonoscopy, mean percent of single returns per zip code, BMI quartiles, and mean triglyceride:HDL ratio. The most important predictive variables for CRC or high-risk polyps in the regularized discriminant analysis model were income returns within income brackets per zip code, the indication for colonoscopy (screening vs. diagnostic), BMI quartiles, triglyceride:HDL ratio (<3 vs ≥3), alcohol use, ASA comorbidity category, statin use, HDL category, and gastrointestinal bleeding as an indication for diagnostic colonoscopy.
Fig 4

Comparison of the reference Area Under the Curve (AUC) to machine learning models in the test set for colorectal cancer (CRC) and CRC or high-risk polyps.

Discussion

Cost-effective tools are needed to improve CRC screening in young adults and to reduce the morbidity and mortality associated with delayed diagnosis. We assessed the utility of machine learning for creating a predictive model for colorectal neoplasia using single-center data from 3,116 colonoscopy patients aged 35–50. To develop this model, we carefully selected four machine-learning approaches (regularized discriminant analysis, random forest, neural network, and gradient boosting decision tree) and compared them to a logistic regression model. Regularized discriminant analysis minimizes the misclassification probability compared to logistic regression [41]. Random forest and gradient boosting decision trees are more powerful than logistic regression when there are higher-order interactions between predictors [42, 43]. Neural networks have a high tolerance for noise and are able to diagnose networks on their own [44]. In our analyses, the machine learning models achieved better predictive performance for CRC or high-risk polyps using data routinely available in EHRs (e.g., indication, zip code, BMI, and laboratory studies). The machine learning models also achieved higher specificity, positive predictive values, and accuracy for predicting our outcomes (i.e., leading to less over-utilization of testing). To our knowledge, this is the first study that has applied modern machine learning approaches to predict colorectal neoplasia in adults aged 35–50 with or without symptoms. Multiple professional societies now recommend initiating CRC screening at age 45 as opposed to 50. However, timely adoption of CRC screening in adults younger than 50 remains challenging, due to the overall low population risk of CRC or pre-malignant polyps, costs of screening, and risks of colonoscopy. An alternative screening strategy is to perform FIT annually starting at 45 years of age followed by colonoscopies starting at 50 years of age. In this scenario, the number of colonoscopies would be reduced, but stool samples would have to be mailed yearly. Ultimately, this cumbersome strategy may not be as cost-effective as expanding screening colonoscopies in older more at-risk adults [6]. Furthermore, despite the high specificity of the FIT for CRC (94.9%), its sensitivity is low for CRC (73.8%), advanced adenomas (23.8%), and advanced serrated lesions (5%) [45]. One example is the Colorectal Risk Assessment Tool (CCRAT), which is endorsed by the National Cancer Institute [46]. CCRAT relies on patients to report risk factors, was only validated in adults over the age of 50, and modestly discriminates CRC (AUC of 0.61) [47]. Although newer tools can stratify the risk of high-risk polyps or CRC, they too have suboptimal discriminatory performance [27]. For instance, a recent head-to-head comparison of 17 risk models yielded an AUC of 0.58 to 0.65 for advanced adenomas or CRC [13]. More recent models had modest prediction ability even when using an extensive list of predictors, which can be cumbersome for patients and providers [27, 48]. Furthermore, most risk stratification tools were created for asymptomatic adults ≥50 years of age and are not specifically tailored for adults <50 years of age or those with symptoms [13]. Indeed, machine learning approaches are reported to be powerful tools for predictive analytics in healthcare [49, 50] and have demonstrated substantial success in many applications such as biomarker identification [51] and outcome prediction [52, 53]. Therefore, the present study builds on and extends these reports by demonstrating the superior ability of modern machine learning approaches to predict CRC (AUC of 0.75 for the neural network model) and CRC or high-risk polyps (AUC of 0.64 for the regularized discriminant analysis model) compared to conventional logistic regression using variables routinely available in EHRs. When it comes to practical applications, and based on the improved positive prediction of CRC or high-risk serrated or adenomatous polyps (21.9% in the neural network model compared to conventional regression) observed in our study, we suggest that combination of machine learning-based risk assessment and FIT could offer a cost-effective early screening strategy for adults under the age of 50 and would help to reduce the burden of colonoscopy referrals on the healthcare system. Therefore, future studies should combine machine learning models with non-invasive methods (e.g., FIT and screening for symptoms) to improve the effectiveness of CRC detection in adults under the age of 50. Due to the recent recommendation to lower the age of CRC screening to 45, this approach would be critical to conserve colonoscopy resources by stratifying adults into risk categories.

Strengths and limitations

Our proof-of-concept study has several strengths and limitations. The strengths include the significance and innovation of the model, rigorous methods and reporting, inclusion of average-risk patients with or without symptoms, use of predictors collected during routine clinical care, and internal validation using a split sample. For instance, symptoms (e.g., gastrointestinal bleeding) were important predictors for CRC and can risk-stratify adults in the primary care setting based on symptoms and other CRC predictors. There are several possible explanations for the incremental gains in predictive ability achieved by the machine learning models. For example, although we integrated several well known risk factors for CRC, the categorical formats of continuous variables with clinically meaningful cutoffs may contribute to risk prediction. Machine learning accounts for linear and non-linear relationships between variables, which enhances predictive performance without assuming additivity compared to conventional statistical models [54]. Assembling methods that combine several basic models to produce one optimal model, such as the random forest and gradient boosting decision tree models, results in a well-generalized model and reduces the risk of overfitting [55, 56]. Although machine learning improves predictive ability, the predictions remain imperfect. This is likely due to the subjectivity of symptoms, timing of BMI measurement relative to CRC (early-life vs. pre-diagnosis), and lack of broader predictors, such as diet and physical activity. Although these variables are risk factors for CRC, our objective was to harness the limited set of clinical data that are available in EHRs to develop machine learning models. Machine learning approaches are also data driven and therefore depend on accurate data. This becomes problematic when data is missing. For example, patients referred for screening colonoscopy may have symptoms that were not reported at time of referral [57]. Patients may also under-report recreational drugs’ use or the use of over-the-counter medications such as aspirin. Second, the imputation of missing data is a potential source of bias. Nevertheless, imputation by machine learning is a rigorous technique, especially when compared to regression [58]. Third, due to the rarity of CRC, our dataset was imbalanced, which may bias predictions towards the dominant class. We applied the “SMOTE” oversampling method to adjust for this bias when developing our models [40]; however, we anticipate the collection of additional patients with CRC in future studies to better address this issue. Forth, the inclusion of non-high-risk polyps may undermine the discrimination of the models. To evaluate the effect of this inclusion, we compared the sensitivity of the machine learning and regression models after exclusion of non-high-risk polyps (data not shown). We did not observe an improvement in the predictive power of any of the models.

Conclusions

In this analysis of data routinely collected in EHRs for clinical purposes, we demonstrated that machine learning has a superior ability to predict the risk of colorectal neoplasia in adults aged 35–50 compared to conventional logistic regression. Our machine learning models improved specificity, positive predictive values, and accuracy compared to logistic regression and therefore have the potential to reduce invasive testing. Future research should aim to validate our model in large primary care and referral settings and to expand machine learning models by using a broader set of predictors. Upon successful completion of this work, machine learning models have the potential to stratify adults aged 35–50 years with or without symptoms into CRC risk categories, which will lead to precise and cost-effective prevention and early detection of CRC. Our ultimate vision is for machine learning risk assessment tools to be seamlessly integrated into the health care electronic medical system for real-time monitoring of patient risk. We expect that using EHR-based risk assessment tool will also reduce barriers to adoption of our model and improve uptake and value of screening.

Comparison of predictors in a sub-sample of patients aged 46–49 undergoing diagnostic and screening colonoscopy.

(DOCX) Click here for additional data file. 28 Jan 2022
PONE-D-21-34521
Utility of machine learning in developing a predictive model for early-age-onset colorectal neoplasia using electronic health records
PLOS ONE Dear Dr. Hussan, Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process. Please submit your revised manuscript by Mar 14 2022 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file. Please include the following items when submitting your revised manuscript:
A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'. A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'. An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'. If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter. If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols. We look forward to receiving your revised manuscript. Kind regards, Hsu-Heng Yen Academic Editor PLOS ONE Journal Requirements: When submitting your revision, we need you to address these additional requirements. 1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf 2. Please include the ethics statement in the Methods section, including details of the name of the ethics committee, the approval number, and the fact that consent was waived [Note: HTML markup is below. Please do not edit.] Reviewers' comments: Reviewer's Responses to Questions Comments to the Author 1. Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented. Reviewer #1: Yes Reviewer #2: Yes Reviewer #3: Yes ********** 2. Has the statistical analysis been performed appropriately and rigorously? Reviewer #1: Yes Reviewer #2: Yes Reviewer #3: Yes ********** 3. Have the authors made all data underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified. Reviewer #1: Yes Reviewer #2: Yes Reviewer #3: Yes ********** 4. Is the manuscript presented in an intelligible fashion and written in standard English? PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here. Reviewer #1: Yes Reviewer #2: Yes Reviewer #3: Yes ********** 5. Review Comments to the Author Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters) Reviewer #1: 1. The authors discussing machine learning in this paper but there isn't any clear definition about it. If the authors can provide a clear definition of machine learning and how these machine learning have been use at the introduction section that would helpful 2. As the study involved 3,116 adults aged 35-60, I would be interested to know how these sample were recruited for the study 3. Overall, the paper has demonstrated clear methodology of research Reviewer #2: Dear Authors, The manuscripts is well-organized and well written. The topic is interesting and significant in the current situation. I would suggest few minor revisions to enhance the quality of your manuscript. i) You have added some supporting information. However, the graphs and charts are not clearly visible. Is it possible to upload better quality image? ii) The references you used are good enough in quantity, however, it would be beneficial to add few more references which are up to date (preferably from 2020-2022). It will ensure the research gap. iii) Please add the managerial and practical significance of the research in a separate paragraph. Thank you and good luck! Reviewer #3: I am pleased to have an opportunity to review this scientific paper about the predictive models for early-age-onset colorectal neoplasia. I think this paper is scientifically informative and interesting. I have some comments and suggestions as below. 1.Major The author should add the work plot in this manuscript and it's necessary to add the inclusion and exclusion criteria in this plot. The figure legend and table should put in the end of manuscript. You should put the P value in your figure. The history of colorectal neoplasia in family member may be added in this manuscript, which was an important factor associated with colorectal cancer. Patients with abnormal laboratory studies (CEA and CA199) should be excluded in this manuscript which may result a bias of the cohort. ********** 6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy. Reviewer #1: No Reviewer #2: Yes: Syed Far Abid Hossain Reviewer #3: No [NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.] While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step. 4 Feb 2022 REVIEWERS’ COMMENTS: Comments to the Author Reviewer #1: 1. The authors discussing machine learning in this paper but there isn't any clear definition about it. If the authors can provide a clear definition of machine learning and how these machine learning have been use at the introduction section that would helpful Answer: Thank you for this important point. We included a definition of machine learning with a reference in introduction. A description of the utilized machine learning algorithms is also detailed in the methods and in our discussion. 2. As the study involved 3,116 adults aged 35-60, I would be interested to know how these sample were recruited for the study Answer: This is a very good comment. As described, this is a retrospective cohort available at the Ohio State University. We clarified further that our cohort and their colonoscopy data were obtained from the Ohio State University colonoscopy quality reporting GI Quality Improvement Consortium, Ltd (GIQuic) database. GIQuic is a collaborative, nonprofit, scientific organization between the American College of Gastroenterology and the American Society for Gastrointestinal Endoscopy1. We linked to electronic health records data to our database to form a retrospective sample of 3,116 adults aged 35-50 with EHR data and colonoscopy-based colorectal neoplasia outcomes. We clarified this point in our methods and Fig 1. 3. Overall, the paper has demonstrated clear methodology of research Answer: Thank you for the kind comment. Reviewer #2: Dear Authors, The manuscripts is well-organized and well written. The topic is interesting and significant in the current situation. I would suggest few minor revisions to enhance the quality of your manuscript. 1. You have added some supporting information. However, the graphs and charts are not clearly visible. Is it possible to upload better quality image? Answer: We apologize the quality. It appears the images lost quality in the submission process and we have since rectified this mistake. 2. The references you used are good enough in quantity, however, it would be beneficial to add few more references which are up to date (preferably from 2020-2022). It will ensure the research gap. Answer: Thank you for pointing that out. We added more up-to-date references. We also revised and enhanced the quality of the paper. 3. Please add the managerial and practical significance of the research in a separate paragraph. Answer: This is a very reasonable comment. We included practical significance of this work in one paragraph. Thank you and good luck! Reviewer #3: I am pleased to have an opportunity to review this scientific paper about the predictive models for early-age-onset colorectal neoplasia. I think this paper is scientifically informative and interesting. I have some comments and suggestions as below. 1. Major, the author should add the work plot in this manuscript and it's necessary to add the inclusion and exclusion criteria in this plot. Answer: Thank you for helping improve the presentation of our paper. We added a plot describing our inclusion and exclusions (Fig 1). 2. The figure legend and table should put in the end of manuscript. Answer: Thank you. We are following PLOS ONE's style requirements when it comes to figure legends and tables. Specifically, per the instructions: “Each figure caption should appear directly after the paragraph in which they are first cited.” “Tables should be included directly after the paragraph in which they are first cited.” Link to instructions: https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf 3. You should put the P value in your figure. Answer: Many thanks. To clarify, the p values compare the machine learning methods to conventional regressions and are already included in Fig 3. We clarified the meaning of our p values in Fig 3 in the manuscript. Fig 2 and 4 do not have p values to insert. 4. The history of colorectal neoplasia in family member may be added in this manuscript, which was an important factor associated with colorectal cancer. Answer: As early-onset CRC primarily occurs in adults with no strong familial predisposition or pre-existing colitis2, we included only average-risk adults without family history of CRC. History of colorectal neoplasia was indeed an included factor in our prediction module, but however it was not high enough compared to environmental risk factors such as social factors, obesity and insulin resistance. This is likely since environmental risk factors are likely more contributing than genetic factors for the increased risk in average-risk adults with early-onset neoplasia. 5. Patients with abnormal laboratory studies (CEA and CA199) should be excluded in this manuscript which may result a bias of the cohort. Answer: We completely agree. None of the patients in the cohort were referred due to an abnormal CEA or CA199. We clarified that in our exclusions. REFERENCES: 1. GIQuIC. www.giquic.org. Accessed January 27, 2022. 2. Dozois EJ, Boardman LA, Suwanthanma W, et al. Young-onset colorectal cancer in patients with no known genetic predisposition: can we increase early recognition and improve outcome? Medicine (Baltimore). 2008;87(5):259-263. Submitted filename: Response to reviewers.docx Click here for additional data file. 28 Feb 2022 Utility of machine learning in developing a predictive model for early-age-onset colorectal neoplasia using electronic health records PONE-D-21-34521R1 Dear Dr. Hussan, We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements. Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication. An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org. If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org. Kind regards, Hsu-Heng Yen Academic Editor PLOS ONE Reviewers' comments: Reviewer's Responses to Questions Comments to the Author 1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation. Reviewer #2: All comments have been addressed Reviewer #3: All comments have been addressed ********** 2. Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented. Reviewer #2: Yes Reviewer #3: Yes ********** 3. Has the statistical analysis been performed appropriately and rigorously? Reviewer #2: Yes Reviewer #3: Yes ********** 4. Have the authors made all data underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified. Reviewer #2: Yes Reviewer #3: Yes ********** 5. Is the manuscript presented in an intelligible fashion and written in standard English? PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here. Reviewer #2: Yes Reviewer #3: Yes ********** 6. Review Comments to the Author Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters) Reviewer #2: The revised version is good. Thank you for addressing the comments properly. The manuscript is well organized now. Reviewer #3: (No Response) ********** 7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy. Reviewer #2: Yes: Syed Far Abid Hossain, Coordinator, MBA Program, IUBAT Reviewer #3: No
  38 in total

1.  Cost-Effectiveness and National Effects of Initiating Colorectal Cancer Screening for Average-Risk Persons at Age 45 Years Instead of 50 Years.

Authors:  Uri Ladabaum; Ajitha Mannalithara; Reinier G S Meester; Samir Gupta; Robert E Schoen
Journal:  Gastroenterology       Date:  2019-03-28       Impact factor: 22.682

Review 2.  Machine Learning and Artificial Intelligence: Definitions, Applications, and Future Directions.

Authors:  J Matthew Helm; Andrew M Swiergosz; Heather S Haeberle; Jaret M Karnuta; Jonathan L Schaffer; Viktor E Krebs; Andrew I Spitzer; Prem N Ramkumar
Journal:  Curr Rev Musculoskelet Med       Date:  2020-02

3.  Global patterns and trends in colorectal cancer incidence in young adults.

Authors:  Rebecca L Siegel; Lindsey A Torre; Isabelle Soerjomataram; Richard B Hayes; Freddie Bray; Thomas K Weber; Ahmedin Jemal
Journal:  Gut       Date:  2019-09-05       Impact factor: 23.059

4.  Missing data imputation using statistical and machine learning methods in a real breast cancer problem.

Authors:  José M Jerez; Ignacio Molina; Pedro J García-Laencina; Emilio Alba; Nuria Ribelles; Miguel Martín; Leonardo Franco
Journal:  Artif Intell Med       Date:  2010-07-16       Impact factor: 5.326

5.  Colonoscopic evaluation of minimal rectal bleeding in average-risk patients for colorectal cancer.

Authors:  Shahriar Nikpour; Ali Ali Asgari
Journal:  World J Gastroenterol       Date:  2008-11-14       Impact factor: 5.742

6.  Validation of a colorectal cancer risk prediction model among white patients age 50 years and older.

Authors:  Yikyung Park; Andrew Nathan Freedman; Mitchell H Gail; David Pee; Albert Hollenbeck; Arthur Schatzkin; Ruth M Pfeiffer
Journal:  J Clin Oncol       Date:  2008-12-29       Impact factor: 44.544

7.  Transparent Reporting of a multivariable prediction model for Individual Prognosis or Diagnosis (TRIPOD): the TRIPOD statement.

Authors:  Gary S Collins; Johannes B Reitsma; Douglas G Altman; Karel G M Moons
Journal:  Ann Intern Med       Date:  2015-01-06       Impact factor: 25.391

8.  Machine learning techniques for personalized breast cancer risk prediction: comparison with the BCRAT and BOADICEA models.

Authors:  Chang Ming; Valeria Viassolo; Nicole Probst-Hensch; Pierre O Chappuis; Ivo D Dinov; Maria C Katapodi
Journal:  Breast Cancer Res       Date:  2019-06-20       Impact factor: 6.466

9.  Scoring colorectal cancer risk with an artificial neural network based on self-reportable personal health data.

Authors:  Bradley J Nartowt; Gregory R Hart; David A Roffman; Xavier Llor; Issa Ali; Wazir Muhammad; Ying Liang; Jun Deng
Journal:  PLoS One       Date:  2019-08-22       Impact factor: 3.240

10.  Derivation and validation of a predictive model for advanced colorectal neoplasia in asymptomatic adults.

Authors:  Thomas F Imperiale; Patrick O Monahan; Timothy E Stump; David F Ransohoff
Journal:  Gut       Date:  2020-09-29       Impact factor: 23.059

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.