Literature DB >> 36123745

An ensemble learning with active sampling to predict the prognosis of postoperative non-small cell lung cancer patients.

Danqing Hu1,2, Huanyao Zhang1,2, Shaolei Li3, Huilong Duan1,2, Nan Wu4, Xudong Lu5,6.   

Abstract

BACKGROUND: Lung cancer is the leading cause of cancer death worldwide. Prognostic prediction plays a vital role in the decision-making process for postoperative non-small cell lung cancer (NSCLC) patients. However, the high imbalance ratio of prognostic data limits the development of effective prognostic prediction models.
METHODS: In this study, we present a novel approach, namely ensemble learning with active sampling (ELAS), to tackle the imbalanced data problem in NSCLC prognostic prediction. ELAS first applies an active sampling mechanism to query the most informative samples to update the base classifier to give it a new perspective. This training process is repeated until no enough samples are queried. Next, an internal validation set is employed to evaluate the base classifiers, and the ones with the best performances are integrated as the ensemble model. Besides, we set up multiple initial training data seeds and internal validation sets to ensure the stability and generalization of the model.
RESULTS: We verified the effectiveness of the ELAS on a real clinical dataset containing 1848 postoperative NSCLC patients. Experimental results showed that the ELAS achieved the best averaged 0.736 AUROC value and 0.453 AUPRC value for 6 prognostic tasks and obtained significant improvements in comparison with the SVM, AdaBoost, Bagging, SMOTE and TomekLinks.
CONCLUSIONS: We conclude that the ELAS can effectively alleviate the imbalanced data problem in NSCLC prognostic prediction and demonstrates good potential for future postoperative NSCLC prognostic prediction.
© 2022. The Author(s).

Entities:  

Keywords:  Active sampling; Ensemble learning; Non-small cell lung cancer; Prognostic prediction

Mesh:

Year:  2022        PMID: 36123745      PMCID: PMC9487160          DOI: 10.1186/s12911-022-01960-0

Source DB:  PubMed          Journal:  BMC Med Inform Decis Mak        ISSN: 1472-6947            Impact factor:   3.298


Background

Lung cancer is a type of cancer that begins in the lungs and may spread to lymph nodes or other organs in the body. It is the most diagnosed cancer and the leading cause of cancer death globally [1]. The two main types of lung cancer are small-cell lung cancer (SCLC) and non-small cell lung cancer (NSCLC). NSCLC is the most common type and accounts for about 85% of all lung cancer cases. The prognosis of NSCLC patients is poor and only 23.3% of cases can survive for more than 5 years [2]. In the era of precision medicine, more and more treatment options have become available. Besides the characteristics of cancer, cancer stage, treatment history, etc., prognosis is also of importance on the choice of complicated multidisciplinary treatment [3]. At present, surgery remains the only potentially curative modality for resectable NSCLC patients. However, cancer may recur at any time after surgery and seriously threaten the survival of postoperative patients [4]. Thus, it is critical to predict the prognosis of postoperative patients accurately to optimize the clinical decisions, such as adjuvant treatment selection and personalized follow-up plan, so that patients can receive proper management to improve the quality of life and even prolong the survival time [5, 6]. To accurately assess the prognosis of patients, researchers have adopted multiple machine learning algorithms to develop prognostic models by exploiting various data like clinical, imaging, and genomic data [7]. Although these models are capable of mining nontrivial knowledge from historical data [8-12], the imbalanced data problem is still a bottleneck of building a robust prognostic prediction model, especially for patients who relapsed or died shortly after surgeries, which causes the algorithms to bias the majority-class cases and affects the predictive performance [13]. Therefore, we need an effective strategy to counteract this problem. In this paper, we propose a novel approach, i.e., ensemble learning with active sampling (ELAS), to alleviate the problem caused by imbalanced data. Active sampling has shown great potential to deal with the imbalanced data problem [14-17]. ELAS develops the first base classifier using a balanced initial training data seed, and then applies the active sampling mechanism to query samples to update the base classifier. Next, the base classifiers that achieve good performances on an internal validation set are integrated as the ensemble model. To evaluate the proposed method, extensive experiments were conducted on a clinical dataset consisting of 1848 postoperative NSCLC patients collected from a Chinese Cancer Hospital. Experimental results indicate that the ELAS outperforms several benchmark models, which shows the ability to alleviate the imbalanced data problem in postoperative NSCLC prognostic prediction.

Methods

Prognostic prediction for postoperative NSCLC patients is a typical imbalanced learning problem, especially for short-term prognosis prediction. Therefore, directly applying the traditional machine learning algorithms may lead to poor performance [13]. In this study, we propose the ELAS to alleviate the problem. Figure 1 illustrates the process of ELAS. The ELAS mainly consists of three parts, i.e., data initialization, active sampling, and model ensemble. We will elaborate on the details of the ELAS as follows.
Fig. 1

The process of the ELAS

The process of the ELAS

Data initialization

For training set where x is the patient sample and is the sample size of the training set. Before active sampling, we first randomly select 20% of the samples from the as the internal validation set . Note that the is designed for the selection of the base classifiers in the ELAS model, which is different from the traditional validation set for hyperparameter selection or early stopping. And the remaining 80% of samples in are regarded as the training data pool with sample size . When obtaining the , we randomly select samples with no replacement from the majority class and minority class of respectively as a balanced initial training data seed to train the first base classifier, where is the sample size of the . And the is updated by removing the samples in the .

Active sampling

Using the balanced , we train the first base classifier with any reasonable supervised machine learning algorithms. When the first base classifier is trained, we employ it to predict the risks of samples in the and select the most informative samples from using any reasonable query strategies. In this study, we employ the ranked batch-mode sampling (RBMS) described in the literature [18] as the query strategy. In comparison with the traditional active learning query strategies like uncertainty sampling, RBMS uses Eq. (1) to assign the final scores for a batch of samples not only considering the informativeness of each sample but also the similarity between the samples and the already selected ones. Note that the α parameter is responsible for weighting the impact of similarity score and uncertainty score in the sample’s final score . Using Eq. (2), α leads the query strategy to prioritize diversity on the initial iterations where the is much smaller than the while, with the increase of the queried samples, shift the priority to samples in which the classifier is uncertain about. is equal to at the first active sampling iteration. To determine the uncertainty of the sample, the RBMS uses the least confident uncertainty score. Let be the probability of a sample belonging to class j predicted by the classifier, then the uncertainty score can be calculated by Eq. (3). Moreover, RBMS employs Eq. (4) to find the similarity score, where is the current sample, is the dataset including samples in and the selected samples in this query round. is the similarity function to measure the distance between the and the sample in . We used the Euclidean distance as the similarity function in this study. Based on the RBMS, we can avoid the sub-optimal sample selection caused by traditional active learning query strategies when selecting informative samples. The queried patient samples are added into as the new training data and removed from . So far, the first active sampling process is done, and we obtain the first classifier , new training data , and training data pool . Based on the new and , we can start the next round of active sampling process until not enough samples in can be sampled into for base classifier development. During each active sampling iteration, one base classifier is trained and used to query new samples for the next base classifier. All the trained base classifiers during this process are stored in the base classifier list L waiting for the final base classifier selection. In this study, we do not use the stop criteria to early terminate the training process [19-21], because the discrimination ability of the base classifier does not always improve with the addition of queried samples when using the real clinical data.

Model ensemble

After the active sampling, we can obtain a base classifier list L with base classifiers, where is the sample size of the before training data seed sampling. Among these base classifiers, we select top K base classifiers with the best prediction performances on the internal validation set for the ensemble model. However, the only accounts for 20% of the , which may lead the selected base classifiers to overfit this and deteriorate the generalization ability of the ensemble model. Thus, we apply a stratified fivefold cross-validation mechanism to generate the . Each fold is regarded as one for base classifier evaluation, and the remaining 4 folds are combined as the for base classifier training. Using this strategy, each sample in the will be used to evaluate and select base classifiers, and we can obtain 5 base classifier lists where each list corresponds to a to avoid overfitting to one specific . Moreover, we also notice that the different initial training data seed will lead to the different first base classifier and the following active sampling results and then the different subsequent base classifiers. To obtain more stable and robust prognostic prediction performance, we initialize times with different random seeds and repeat the whole active sampling process separately to obtain base classifier lists during each fold. Thus, when using fivefold cross-validation for multiple generations and times initializations, we can obtain a total of base classifier lists. We select the top K base classifiers from each L based on their performances on corresponding internal validation sets. The ELAS will average the base classifiers’ outputs as the final ensemble result. The details of the whole training process of the ELAS are given in Algorithm I.

Experimental setup

To develop the ELAS model, we selected support vector machine (SVM) [22], logistic regression with L2 regularization (L2-LR) [23], and classification and regression trees (CART) [24], to train the base classifiers. We randomly divided 80% of samples as the training set and the remaining 20% as the test set. To tune the hyper-parameters, fivefold cross-validation was employed on the training set, and a grid search strategy was applied for the base classifiers on the hyper-parameter spaces: for SVM, for L2-LR, and for CART. To release the problem of massive possible value sets of the hyper-parameters, we selected radial basis function kernel for SVM, Gini impurity for CART, and , 10 for , 3 for , 20 for . Note that we should keep the no more than the sample size of minority class because we want the to be a balanced dataset. Besides, we should also keep the K no more than to ensure that the top K base classifiers can be selected from. In this study, we conducted extensive experiments to explore the effectiveness of the proposed ELAS approach. First, we compared the ELAS with the base classifier algorithms, i.e., SVM, L2-LR, and CART, to explore whether the ELAS can improve the performance of prognostic prediction. And then, as the ELAS is an ensemble method, we also selected two famous ensemble methods, i.e., AdaBoost [25] and Bagging [26, 27], as the benchmarks. Moreover, we also applied two resampling methods to deal with imbalanced data, namely SMOTE [28] and TomekLinks [29], to explore which strategy is better. To evaluate the ELAS and benchmarks’ performances, we employed the area under the receiver operating characteristic curve (AUROC) and the area under the precision–recall curve (AUPRC) as the metrics. To eliminate the bias caused by the test set partition, the whole data set segmentation, model development, and evaluation process was repeated 10 times with different random seeds so that we can obtain the averaged AUROC value and AUPRC value with their standard deviations (SD) for each prognostic task. The paired student t-test was performed to determine whether the AUROC and AUPRC values of ELAS are statistically significantly different from the values of the benchmark algorithms and a p value less than 0.05 was considered significant.

Results

Data

We reviewed 1848 NSCLC patients who had undergone curative surgery from 2006 to 2015 in the Department of Thoracic Surgery II of Peking University Cancer Hospital. The collected data covered patient demographic information, preoperative exams and treatments, pathological information of the primary tumor and lymph nodes, and the pathological TNM stage. Clinicians manually recorded all the clinical data to ensure its reliability and correctness. The details of the clinical data are listed in the Additional file 1. Before model development, we preprocessed the collected clinical data. Specifically, patient samples with missing feature values were excluded from the dataset. We removed the features with variance lower than 1% to ensure that no features have almost the same value for all samples. We converted all categorical features into a one-hot encoding form and binned the continuous features into intervals. To label the patient’s recurrence and death statuses, we used the Disease-Free-Survival (DFS) and Overall-Survival (OS) in the follow-up data. In this study, we selected three time periods, i.e., 1-year, 3-year, and 5-year, to explore the effectiveness of the proposed method to handle the different degrees of imbalance. Within each period, we labeled the patients who experienced the events (recurrence or death) as positive samples and those who did not experience any events as negative samples. Patients who lost follow-up within the period and had not experienced any events were excluded from this prognostic task. Table 1 lists the statistics of the 1-year, 3-year, and 5-year prognoses.
Table 1

The statistics of the 1-year, 3-year, and 5-year patient prognoses

OutcomesNumber of patients
1-year3-year5-year
Recurrence, n (%)102 (7.6%)296 (29.1%)377 (51.9%)
No recurrence, n (%)1,246 (92.4%)720 (70.9%)350 (48.1%)
Death, n (%)62 (4.6%)220 (21.8%)307 (43.7%)
No death, n (%)1,288 (95.4%)787 (78.2%)395 (56.3%)
The statistics of the 1-year, 3-year, and 5-year patient prognoses

In comparison with the base classifier algorithms

As an ensemble learning method, we first compared the ELAS with the base classifier algorithms to explore whether the base classifier algorithms can benefit from the ELAS. The AUROC and AUPRC values of the base classifier algorithms and the ELAS are illustrated in Tables 2 and 3. Figures 2 and 3 present the results in the bar graph manner. The sensitivity and specificity values are listed in the Additional file 2. We also calculated the paired student t-test to explore whether there are significant differences between the base classifier algorithms and the ELAS, and the results are listed in Table 4.
Table 2

The AUROC values of the base classifier algorithms and the ELAS

TaskBase classifier algorithmsELAS
SVML2-LRCARTSVM-ELASL2-LR-ELASCART-ELAS
MeanSDMeanSDMeanSDMeanSDMeanSDMeanSD
1-year recurrence0.6490.0630.6600.0720.6030.0720.7020.0790.6740.0710.6680.056
1-year death0.6530.0570.7540.0430.650.0720.7600.0420.7400.0570.7400.059
3-year recurrence0.7130.0410.6970.0270.6370.0310.7280.0330.7090.0290.7060.036
3-year death0.7020.0440.7110.0400.6630.0410.7330.0350.7200.0370.7370.040
5-year recurrence0.7510.0530.7300.0610.6680.0450.7480.0550.7350.0630.7240.051
5-year death0.7390.0330.7180.0280.6310.0440.7420.0290.7290.0260.6940.040
All tasks0.7010.0630.7110.0560.6420.0570.7360.0520.7180.0550.7110.054

The bold means the best results for corresponding tasks

Table 3

The AUPRC values of the base classifier algorithms and the ELAS

TaskBase classifier algorithmsELAS
SVML2-LRCARTSVM-ELASL2-LR-ELASCART-ELAS
MeanSDMeanSDMeanSDMeanSDMeanSDMeanSD
1-year recurrence0.1450.0280.1780.0700.1180.0460.1530.0520.1730.0740.1560.055
1-year death0.1230.0420.1370.0420.1090.0290.1290.0390.1330.0400.1360.041
3-year recurrence0.5180.0540.4970.0410.4060.0330.5270.0500.5090.0470.4860.044
3-year death0.4370.0540.4130.0480.3520.0610.4590.0470.4210.0470.4480.068
5-year recurrence0.7600.0570.7420.0650.6480.0540.7580.0550.7450.0640.7240.046
5-year death0.6940.0450.6800.0360.5320.0350.6950.0400.6900.0340.6340.052
All tasks0.4460.2500.4410.2340.3610.2030.4530.2470.4450.2390.4310.227

The bold means the best results for corresponding tasks

Fig. 2

The AUROC values of the base classifier algorithms and the ELAS

Fig. 3

The AUPRC values of the base classifier algorithms and the ELAS

Table 4

The paired student t-test results between the base classifier algorithms and the ELAS

MetricComparison1-year tasks3-year tasks5-year tasksAll tasks
AUROCSVM versus SVM-ELAS< 0.01< 0.010.411< 0.01
L2-LR versus L2-LR-ELAS0.487< 0.01< 0.01< 0.01
CART versus CART-ELAS< 0.01< 0.01< 0.01< 0.01
AUPRCSVM versus SVM-ELAS0.165< 0.010.378< 0.01
L2-LR versus L2-LR-ELAS0.093< 0.010.0110.015
CART versus CART-ELAS< 0.01< 0.01< 0.01< 0.01

The bold means the p-value is less than 0.05, which means the results between different models have statistically significant differences

The AUROC values of the base classifier algorithms and the ELAS The bold means the best results for corresponding tasks The AUPRC values of the base classifier algorithms and the ELAS The bold means the best results for corresponding tasks The AUROC values of the base classifier algorithms and the ELAS The AUPRC values of the base classifier algorithms and the ELAS The paired student t-test results between the base classifier algorithms and the ELAS The bold means the p-value is less than 0.05, which means the results between different models have statistically significant differences Based on the experimental results above, we find that the ELAS achieves significant improvements compared with all base classifier algorithms under both AUROC and AUPRC metrics when combining all prognostic tasks together. Moreover, the more serious the data imbalance, the more improvements on AUROC values obtained by the ELAS, which indicates the ELAS strategy can better tackle the imbalanced problem than the base classifiers. From Fig. 3 we can notice that the AUPRC values increase in a step-like manner with the extension of the time of the prognostic prediction task, which is because the AUPRC is mainly affected by the degree of data imbalance. For 1-year prognostic prediction tasks, only CART significantly benefited from the ELAS method under AUPRC metric, but SVM and L2-LR did not gain significant improvements on AUPRC values via using ELAS. For 3-year prognostic prediction tasks, all base classifier algorithms achieve better performances when using ELAS. For 5-year prognostic prediction tasks, both L2-LR and CART have significant improvements, but SVM does not. By comparing the three ELAS models, we notice that the SVM-ELAS achieved the best overall performances with 0.736 AUROC value and 0.453 AUPRC value. So, we select the SVM-ELAS as the representative model to compare with other benchmarks in the following experiments.

In comparison with the benchmark algorithms

Ensemble learning is one of the effective approaches to handling the imbalanced data problem [16, 17, 30]. In this study, the proposed ELAS also averages the outputs of the selected base classifiers as the final predictive result. So here, we apply two state-of-the-art ensemble learning algorithms, i.e., AdaBoost [25] and Bagging [26, 27], as the benchmarks to compare with the ELAS. AdaBoost is one popular boosting algorithm that fits a sequence of weak classifiers on repeatedly reweighted samples and follows to weighted sum the outputs of weak classifiers as the predictive results. Bagging is another ensemble strategy that randomly samples subsets of the training set without concern for performance to build base classifiers and then averages their outputs as the predictions. Unlike them, the ELAS bias towards selecting samples that are hard to distinguish to train the base classifier step by step and averages the outputs of the base classifiers with the best performances as the ensemble predictions. Besides ensemble learning, resampling techniques are also widely used to alleviate the effect of the skewed class distribution by rebalancing the sample space for an imbalanced dataset [16, 17]. In this subsection, we also select two resampling techniques, i.e., SMOTE [28] and TomekLinks [29], as the benchmarks to compare with the proposed method. SMOTE is an over-sampling method that generates new samples from the vector between the current sample and one of its k nearest neighbors to enrich the minority class. TomekLinks is an under-sampling method that first detects if the two samples of different classes are the nearest neighbors of each other and then deletes the one in the majority class to reduce the majority class. Tables 5 and 6 show the AUROC values and AUPRC values of the benchmarks and the ELAS, respectively. Figures 4 and 5 present the experimental results more intuitively. The sensitivity and specificity values of the benchmarks and the ELAS are listed in the Additional file 3. To further prove the performance improvements of the ELAS, the paired student t-test is also conducted to compare the performances of the benchmarks and the ELAS and listed in Table 7.
Table 5

The AUROC values of the ensemble algorithms, resampling algorithms, and the ELAS

TaskEnsemble algorithmsResampling algorithmsProposed
SVM-AdaBoostSVM-BaggingSVM-SMOTESVM-TomekLinksSVM-ELAS
MeanSDMeanSDMeanSDMeanSDMeanSD
1-year recurrence0.6820.0820.6730.0720.6200.0730.6500.0650.7020.079
1-year death0.7680.0550.7260.0470.6700.0580.6680.0580.7600.042
3-year recurrence0.6920.0380.7230.0370.7060.0310.7230.0380.7280.033
3-year death0.7070.0430.7210.0390.7100.0300.7110.0430.7330.035
5-year recurrence0.7520.0550.7520.0530.7510.0530.7520.0530.7480.055
5-year death0.7240.0310.7390.0320.7320.0310.7380.0360.7420.029
All tasks0.7210.0620.7220.0540.6980.0650.7070.0620.7360.052

The bold means the best results for corresponding tasks

Table 6

The AUPRC values of the ensemble algorithms, resampling algorithms, and the ELAS

TaskEnsemble algorithmsResampling algorithmsProposed
SVM-AdaBoostSVM-BaggingSVM-SMOTESVM-TomekLinksSVM-ELAS
MeanSDMeanSDMeanSDMeanSDMeanSD
1-year recurrence0.1500.0470.1530.0490.1140.0330.1510.0430.1530.052
1-year death0.1340.0330.1250.0430.1010.0340.1240.0500.1290.039
3-year recurrence0.4930.0420.5240.0510.4900.0460.5240.0510.5270.050
3-year death0.4200.0450.4540.0540.4040.0480.4490.0440.4590.047
5-year recurrence0.7620.0520.7650.0540.7620.0570.7630.0570.7580.055
5-year death0.6810.0400.6930.0450.6780.0480.6850.0500.6950.040
All tasks0.4400.2430.4520.2490.4250.2570.4490.2480.4530.247

The bold means the best results for corresponding tasks

Fig. 4

The AUROC values of the ensemble algorithms, resampling algorithms, and the ELAS

Fig. 5

The AUPRC values of the ensemble algorithms, resampling algorithms, and the ELAS

Table 7

The paired student t-test results between the benchmark algorithms and the ELAS

MetricComparison1-year tasks3-year tasks5-year tasksAll tasks
AUROCSVM-AdaBoost versus SVM-ELAS0.231< 0.010.041< 0.01
SVM-Bagging versus SVM-ELAS< 0.01< 0.010.490< 0.01
SVM-SMOTE versus SVM-ELAS< 0.01< 0.010.104< 0.01
SVM-TomekLinks versus SVM-ELAS< 0.01< 0.010.454< 0.01
AUPRCSVM-AdaBoost versus SVM-ELAS0.428< 0.010.096< 0.01
SVM-Bagging versus SVM-ELAS0.3960.0460.1460.337
SVM-SMOTE versus SVM-ELAS< 0.01< 0.010.084< 0.01
SVM-TomekLinks versus SVM-ELAS0.3340.0410.2870.088

The bold means the p-value is less than 0.05, which means the results between different models have statistically significant differences

The AUROC values of the ensemble algorithms, resampling algorithms, and the ELAS The bold means the best results for corresponding tasks The AUPRC values of the ensemble algorithms, resampling algorithms, and the ELAS The bold means the best results for corresponding tasks The AUROC values of the ensemble algorithms, resampling algorithms, and the ELAS The AUPRC values of the ensemble algorithms, resampling algorithms, and the ELAS The paired student t-test results between the benchmark algorithms and the ELAS The bold means the p-value is less than 0.05, which means the results between different models have statistically significant differences Note that the SVM-ELAS achieved the best overall performance with 0.736 AUROC value and 0.453 AUPRC value for all tasks together compared with the benchmarks. For 1-year prognostic prediction tasks, the SVM-ELAS outperformed the benchmark algorithms on AUROC values significantly except for the SVM-AdaBoost. The possible reason is that there is a good similarity between the reweighting in AdaBoost and active sampling in the ELAS. Specifically, AdaBoost gives higher weights to misclassified samples so that subsequent base classifiers can tend to classify them correctly, while ELAS actively selects the indistinguishable samples into the training data and uses these samples for all subsequent base classifier developments. So, the samples hard to classify are paid extra attention in both AdaBoost and ELAS, which may lead the similar prediction performances for the 1-year prognostic prediction tasks. Although SVM-ELAS did not outperform the SVM-AdaBoost on 1-year prognostic prediction tasks but obtained significant improvements on both metrics when combining all tasks. Compared with the resampling methods, the SVM-ELAS outperforms the benchmark models for 1-year prediction tasks except for AUPRC of SVM-TomekLinks on 1-year prediction tasks, which indicates the ELAS is a competitive strategy to handle the imbalanced data problem compared with SMOTE and TomekLinks. For 3-year prognostic tasks, the SVM-ELAS achieves significant improvements on both AUROC and AUPRC values compared with all benchmarks. But for 5-year prognostic prediction tasks, the SVM-ELAS did not show significant improvements, probably due to the data imbalance problem becoming relatively weak.

Discussion

In this study, we proposed the ELAS to tackle the imbalanced data problem in NSCLC prognostic prediction. Our approach is generalizable for other biomedical data analyses with imbalanced prediction targets. The experimental results have demonstrated that the ELAS has robust predictive performance, especially for short-term prognostic prediction, when compared with the state-of-the-art techniques. Although the ELAS achieves comparative performances for NSCLC prognostic prediction, there are multiple directions we would like to further work into for more meaningful discoveries. In the current study, we just employed one kind of query strategy described in the literature [18] to select the most informative patient samples. However, exploiting multiple query criteria together shows great potential to improve the performance for classification problems [21, 31]. In the future, we can attempt to use multiple query criteria to select the representative samples from different perspectives to facilitate the development of the base classifier, e.g., using information density to take the structure of the data into account [32], combining base classifiers from different initial training data set as a committee to select the samples with the most disagreements [32]. Moreover, although we applied multiple internal validation sets for the base classifier selection to alleviate the overfitting problem, this selection strategy makes the selected base classifiers easily overfit to the corresponding internal validation set. In the future, we can attempt to sample the base classifiers using the distribution generated from the performances of base classifiers or randomly select a subset of top N classifiers to further reduce the overfitting problem.

Conclusions

In this study, we proposed the ELAS approach to predict the prognosis for postoperative NSCLC patients. Experimental results indicate that the ELAS achieves the best overall performance with an averaged 0.736 AUROC value and 0.453 AUPRC value in comparison with the benchmark models, which indicates it can effectively alleviate the imbalanced data problem in NSCLC prognostic prediction. Additional file 1. The details of the experimental datasets. Additional file 2. The sensitivity and specificity values of the base classifier algorithms and the ELAS. Additional file 3. The sensitivity and specificity values of the ensemble algorithms, resampling algorithms, and the ELAS.
  13 in total

1.  Ensemble learning with active example selection for imbalanced biomedical data classification.

Authors:  Sangyoon Oh; Min Su Lee; Byoung-Tak Zhang
Journal:  IEEE/ACM Trans Comput Biol Bioinform       Date:  2011 Mar-Apr       Impact factor: 3.710

2.  MACE prediction of acute coronary syndrome via boosted resampling classification using electronic medical records.

Authors:  Zhengxing Huang; Tak-Ming Chan; Wei Dong
Journal:  J Biomed Inform       Date:  2017-01-05       Impact factor: 6.317

3.  Prediction of mortality after radical cystectomy for bladder cancer by machine learning techniques.

Authors:  Guanjin Wang; Kin-Man Lam; Zhaohong Deng; Kup-Sze Choi
Journal:  Comput Biol Med       Date:  2015-05-29       Impact factor: 4.589

4.  Global Cancer Statistics 2020: GLOBOCAN Estimates of Incidence and Mortality Worldwide for 36 Cancers in 185 Countries.

Authors:  Hyuna Sung; Jacques Ferlay; Rebecca L Siegel; Mathieu Laversanne; Isabelle Soerjomataram; Ahmedin Jemal; Freddie Bray
Journal:  CA Cancer J Clin       Date:  2021-02-04       Impact factor: 508.702

5.  Relationship between cancer patients' predictions of prognosis and their treatment preferences.

Authors:  J C Weeks; E F Cook; S J O'Day; L M Peterson; N Wenger; D Reding; F E Harrell; P Kussin; N V Dawson; A F Connors; J Lynn; R S Phillips
Journal:  JAMA       Date:  1998-06-03       Impact factor: 56.272

6.  Factors that can make an impact on decision-making and decision implementation in cancer multidisciplinary teams: an interview study of the provider perspective.

Authors:  Rozh Jalil; Maria Ahmed; James S A Green; Nick Sevdalis
Journal:  Int J Surg       Date:  2013-03-14       Impact factor: 6.071

7.  Identification and Construction of Combinatory Cancer Hallmark-Based Gene Signature Sets to Predict Recurrence and Chemotherapy Benefit in Stage II Colorectal Cancer.

Authors:  Shanwu Gao; Chabane Tibiche; Jinfeng Zou; Naif Zaman; Mark Trifiro; Maureen O'Connor-McCourt; Edwin Wang
Journal:  JAMA Oncol       Date:  2016-01       Impact factor: 31.777

8.  Machine-learning prediction of cancer survival: a retrospective study using electronic administrative records and a cancer registry.

Authors:  Sunil Gupta; Truyen Tran; Wei Luo; Dinh Phung; Richard Lee Kennedy; Adam Broad; David Campbell; David Kipp; Madhu Singh; Mustafa Khasraw; Leigh Matheson; David M Ashley; Svetha Venkatesh
Journal:  BMJ Open       Date:  2014-03-17       Impact factor: 2.692

Review 9.  Machine learning applications in cancer prognosis and prediction.

Authors:  Konstantina Kourou; Themis P Exarchos; Konstantinos P Exarchos; Michalis V Karamouzis; Dimitrios I Fotiadis
Journal:  Comput Struct Biotechnol J       Date:  2014-11-15       Impact factor: 7.271

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.