Literature DB >> 31979006

Flexible Data Trimming Improves Performance of Global Machine Learning Methods in Omics-Based Personalized Oncology.

Victor Tkachev1, Maxim Sorokin1,2, Constantin Borisov3, Andrew Garazha1, Anton Buzdin1,2,4,5, Nicolas Borisov1,2,4.   

Abstract

(1) Background: Machine learning (ML) methods are rarely used for an omics-based prescription of cancer drugs, due to shortage of case histories with clinical outcome supplemented by high-throughput molecular data. This causes overtraining and high vulnerability of most ML methods. Recently, we proposed a hybrid global-local approach to ML termed floating window projective separator (FloWPS) that avoids extrapolation in the feature space. Its core property is data trimming, i.e., sample-specific removal of irrelevant features. (2)
Methods: Here, we applied FloWPS to seven popular ML methods, including linear SVM, k nearest neighbors (kNN), random forest (RF), Tikhonov (ridge) regression (RR), binomial naïve Bayes (BNB), adaptive boosting (ADA) and multi-layer perceptron (MLP). (3)
Results: We performed computational experiments for 21 high throughput gene expression datasets (41-235 samples per dataset) totally representing 1778 cancer patients with known responses on chemotherapy treatments. FloWPS essentially improved the classifier quality for all global ML methods (SVM, RF, BNB, ADA, MLP), where the area under the receiver-operator curve (ROC AUC) for the treatment response classifiers increased from 0.61-0.88 range to 0.70-0.94. We tested FloWPS-empowered methods for overtraining by interrogating the importance of different features for different ML methods in the same model datasets. (4) Conclusions: We showed that FloWPS increases the correlation of feature importance between the different ML methods, which indicates its robustness to overtraining. For all the datasets tested, the best performance of FloWPS data trimming was observed for the BNB method, which can be valuable for further building of ML classifiers in personalized oncology.

Entities:  

Keywords:  bioinformatics; chemotherapy; machine learning; omics profiling; oncology; personalized medicine

Year:  2020        PMID: 31979006      PMCID: PMC7037338          DOI: 10.3390/ijms21030713

Source DB:  PubMed          Journal:  Int J Mol Sci        ISSN: 1422-0067            Impact factor:   5.923


1. Introduction

A personalized approach in oncology was proven helpful for increasing efficacy of drugs prescription in many cancers [1,2]. Generally, it is based on finding specific biomarkers which can be mutations, protein levels or patterns of gene expression [3]. High throughput gene expression data can be connected with responsiveness on treatment using two major approaches. First, drug efficacy can be simulated using hypothesis-driven drug scoring algorithms which utilize knowledge of drugs molecular specificities and up/downregulated statuses of target genes and molecular pathways in a tumor [1,3,4,5,6]. In turn, agnostic drug scoring approach, including machine learning (ML) methods can offer even a wider spectrum of opportunities by non-hypothesis-driven direct linkage of specific molecular features with clinical outcomes, such as responsiveness on certain types of treatment [7,8]. ML has a variety of methods that could be used for such agnostic approach, e.g., decision trees, DT [9,10], random forests, RF [11], linear [12], logistic [13], lasso [14,15], and ridge [16] regressions, multi-layer perceptron, MLP [10,17,18], support vectors machines, SVM [9,10,19], adaptive boosting [20,21,22]. The high throughput transcriptomic data, including microarray- and next-generation sequencing gene expression profiles can be utilized for building such classifiers/predictors of clinical response to a certain type of treatment. However, the direct use of ML to personalize prediction of clinical outcomes is problematic, due to the lack of sufficient amounts of preceding clinically annotated cases supplemented with the high-throughput molecular data (~thousands or tens thousands of cases per treatment scheme) [23]. Several ML methods have been recently successfully applied for distinguishing between cancer patients with positive and negative responses on various treatments [20,24,25,26]. However, they were not successful (area under curve (AUC) < 0.66) in predicting clinical outcomes for several model datasets, including multiple myeloma expression dataset associated with known clinical responses on cancer drug bortezomib [20,24,25,26,27]. For the classical ML approaches, most of the clinical genetic datasets are insufficient for effectively solving the task of differentiating treatment responders from non-responders [9,28]. Features measured by sequencing (e.g., polymorphisms, mutations or gene expression values) are far more numerous than the cohorts of individual patients with traced clinical outcomes. For generating statistically significant predictions, extensive reduction of a pool of features under consideration is needed to make their number comparable with the number of individual samples available [10,29,30,31]. To leverage the performance of ML in biomedicine, we recently developed an approach called flexible data trimming (Data trimming (DT) is the process of removing or excluding extreme values, or outliers, from a dataset [32]) [8,29,33,34,35]. This approach is heuristic and based on a common geometrical sense (Figure 1). It utilizes the following basic principles: (i) When a new sample is analyzed to make a prediction, the predictor has to be adapted to a new observation, or re-learned; (ii) the re-learned predictor must be built within a new specific subspace, while using reduced (trimmed) training data.
Figure 1

Area under curve (AUC) (a–d), sensitivity (SN) (e–h) and specificity (SP) (i–l) calculated for treatment response classifiers for eleven non-equalized datasets. The classifiers were based on SVM (a,e,i), RF (b,f,j), binomial naïve Bayes (BNB) (c,g,k) and multi-layer perceptron (MLP) (d,h,l) machine learning (ML) methods. The color legend shows the absence or presence of FloWPS in the classifier analytic pipeline and the value of relative balance factor B. On each panel, each violin plot shows the distribution of values for eleven cancer datasets.

Excluding non-informative features helps ML classifiers to avoid extrapolation, which is a well-known Achilles heel of ML [36,37,38,39]. Thus, for every point of a validation dataset, the training dataset is adjusted to form a floating window. We, therefore, called the respective ML approach, floating window projective separator (FloWPS) [8]. In a pilot trial of this approach, it significantly enhanced robustness of the SVM classifier in all ten clinical gene expression datasets totally representing 992 cancer patients either responding or not on the different types of chemotherapy [8]. FloWPS demonstrated surprisingly high performance (the ROC (receiver-operator curve) is a widely used graphical plot that illustrates the diagnostic ability of a binary classifier system as its discrimination threshold is varied. The ROC is created by plotting the true positive rate against the false positive rate at various threshold settings. The area under the ROC curve, called ROC AUC, or simply AUC, is routinely used for assessment of the quality of the classifier. AUC can vary from 0.5 till 1 and the standard threshold discriminating good vs. poor classifiers is AUC > 0.7 or more) of AUC > 0.7 for the leave-one-out scheme in all datasets, including those where responders and non-responders were poorly distinguishable algorithmically in the previous works [20,24,25,26,27]. However, the applicability and usefulness of FloWPS for a wide variety of ML methods remained unstudied. Here, we investigated FloWPS performance for seven popular ML methods, including linear SVM, k nearest neighbors (kNN), random forest (RF), Tikhonov (ridge) regression (RR), binomial naïve Bayes (BNB), adaptive boosting (ADA) and multi-layer perceptron (MLP). We performed computational experiments for 21 high throughput gene expression datasets (41–235 samples per dataset) corresponding to 1778 cancer patients with known responses on chemotherapy treatments. We showed that FloWPS essentially improved the classifier quality for all global ML methods (SVM, RF, BNB, ADA, MLP), where the AUC for the treatment response classifiers increased from 0.65–0.85 range to 0.80–0.95. For all the datasets tested, the best performance of FloWPS data trimming was observed for the BNB method, which can be valuable for further building of ML classifiers in personalized oncology. Additionally, to test the robustness of FloWPS-empowered ML methods against overtraining, we interrogated agreement/consensus features between the different ML methods tested, which were used for building mathematical models for the classifiers. The lack of such agreement/consensus could indicate overtraining of the ML classifiers built, suggesting random noise instead of extracting significant features distinguishing between the treatment responders and non-responders. If ML methods indeed tend to amplify random noise during overtraining, then one could expect a lack of correlation between the features for geometrically different ML models. However, we found here that (i) there were statistically significant positive correlations between different ML methods in terms of relative feature importance, and (ii) that this correlation was enhanced for the ML methods with FloWPS. We, therefore, conclude that the beneficial role of FloWPS is not due to overtraining.

2. Results

2.1. Performance of FloWPS for Equalized Datasets Using All ML Methods with Default Settings

In this study, we used FloWPS in combination with seven ML methods, namely, linear support vector machines (SVM), k nearest neighbors (kNN), random forest (RF), ridge regression (RR), binomial naïve Bayes (BNB), adaptive boosting (ADA) and multi-layer perceptron (MLP). First ten over twenty-one gene expression datasets investigated here had equal numbers of known responders and non-responders and were investigated first. The basic quality characteristics of using seven above ML methods for discriminating between responders and non-responders in these datasets are shown in Supplementary Figures S1_1, S1_2, S1_3, including AUC, sensitivity (SN) and specificity (SP). Each ML method was applied with its default settings using Python package sklearn [40], both with and without data trimming, separately for each dataset. Although different values of relative balance factor B and discrimination threshold τ (see Materials and Methods, Section 4.3) did not affect the ROC AUC characteristics, they were crucial for sensitivity and specificity (Supplementary Figures S1_1, S1_2, S1_3). We found that the use of FloWPS has considerably improved the AUC metric for all global ML methods investigated (SVM, RF, BNB, ADA and MLP), but had no effect on the performance of local methods kNN and RR (Supplementary Figures S1_1, S1_2, S1_3). For the global ML methods, FloWPS improved the classifier quality and increased AUC from 0.61–0.88 range to 0.70–0.94 (Supplementary Figures S1_1, S1_2, S1_3), and AUC median values—from 0.70–0.77 range to 0.76–0.82 (Table 1). In addition, kNN and RR also showed poor SN and SP for B > 1 and B < 1, respectively (Supplementary Figures S1_1, S1_2, S1_3).
Table 1

Performance metrics for seven ML methods with default settings for datasets with equal numbers of responders and non-responders.

ML MethodMethod TypeMedian AUC without FloWPSMedian AUC with FloWPSPaired t-Test p-Value for AUC with-vs.-w/o FloWPS Advantage of FloWPSMedian SN at B = 4Median SP at B = 0.25
SVMGlobal0.740.801.3 × 10−5Yes0.450.42
kNNLocal0.760.750.53No0.250.34
RFGlobal0.740.821.3 × 10-5Yes0.450.42
RRLocal0.800.790.16No0.360.41
BNBGlobal0.770.822.7 × 10−4Yes0.510.58
ADAGlobal0.700.762.4 × 10−4Yes0.320.41
MLPGlobal0.730.826.4 × 10−5Yes0.530.53

Yes–FloWPS is beneficial for ML quality, No–FloWPS is not beneficial for ML quality.

These findings are summarized in Table 1. Considering quality criterion of combining the highest AUC, the highest SN at B = 4 and the highest SP at B = 0.25, the top three methods identified for the default settings were BNB, MLP and RF (Supplementary Figures S1_1, S1_2, S1_3; Table 1).

2.2. Performance of FloWPS for Equalized Datasets Using BNB, MLP and RF Methods with the Advanced Settings

We then checked the performance of three best ML methods (BNB, MLP and RF) for the same ten datasets with equal numbers of responders and non-responders using advanced settings, see Materials and Methods (Supplementary Figures S2_1, S2_2, S2_3; Table 2). FloWPS improved the classifier quality for these three ML methods and increased AUC from 0.75–0.78 range to 0.83-0.84 (Table 2).
Table 2

Performance metrics for BNB, MLP and RF methods with the advanced settings for datasets with equal numbers of responders and non-responder samples.

ML MethodMedian AUC without FloWPSMedian AUC with FloWPSPaired t-Test p-Value for AUC with-vs.-w/o FloWPSMedian SN at B = 4Median SP at B = 0.25
RF0.750.833.5 × 10−60.500.56
BNB0.780.836.7 × 10−40.500.60
MLP0.770.842.4 × 10−40.500.51
For RF, the best results were obtained with the following parameter settings: n_estimators = 30, criterion = “entropy” (Supplementary Figures S2_1, S2_2, S2_3). For BNB, the best settings were alpha = 1.0, binarize = 0.0, and fit_prior = False (Supplementary Figures S2_1, S2_2, S2_3). For MLP, the best settings were hidden_layer_sizes = 30, alpha = 0.001 (Supplementary Figures S2_1, S2_2, S2_3). Among these three ML methods, the best results were obtained for BNB with alpha = 1.0, binarize = 0.0, and fit_prior = False (Supplementary Figures S2_1, S2_2, S2_3). BNB with these parameter settings can be, therefore, recommended for further development and implementation of the expression-based classifiers of individual treatment response, because it showed simultaneously acceptable AUC, SN and SP for the maximum spectrum of datasets tested (Supplementary Figures S2_1, S2_2, S2_3; Table 2).

2.3. Performance of FloWPS for Non-Equalized Datasets Using BNB, MLP, RF and SVM Methods with the Advanced Settings

We then applied the best settings previously found for BNB, MLP and RF methods using responder-equalized data for the new eleven datasets containing different proportions of treatment responders’ and non-responders’ samples. In addition, we also used linear SVM method (Figure 1, Table 3) with penalty parameter C = 1 because our previous results [8] showed that C ≤ 1 minimizes the risk of overtraining for SVM. The output ML classifier quality metrics were obtained for these four methods, including AUC (Figure 1a–d), SN (Figure 1e–h) and SP (Figure 1i–l). In this trial, the number of responders and non-responder samples were not equal. To compensate for the possible influence of the variable proportion of samples in the two classes, SVM and RF calculations were performed using the balanced-class option.
Table 3

Performance metrics for BNB, MLP, RF and SVM methods with the advanced settings for eleven datasets with variable numbers of responders and non-responder samples.

MethodMedian AUC without FloWPSMedian AUC with FloWPSPaired t-Test p-Value for AUC with-vs.-w/o FloWPSMedian SN at B = 4Median SP at B = 0.25
SVM0.810.830.0130.650.70
RF0.760.864.9 × 10−60.560.71
BNB0.840.897.5 × 10−40.780.75
MLP0.830.881.0 × 10−40.630.71
The application of FloWPS improved the classifier quality for these four ML methods, as the median AUC for the treatment response classifiers increased from 0.76–0.84 range to 0.83–0.89 (Figure 1a–d, Table 3). In this experiment, we confirmed the advantage of using FloWPS for all four ML methods tested and the best performance of BNB also for eleven datasets with non-equal numbers of responders and non-responder samples.

2.4. Correlation Study Between Different ML Methods at the Level of Feature Importance

We showed positive pairwise correlations between the different ML methods at the level of relative importance (I, see Materials and Methods) of different features tested (Table 4, Supplementary Figures S3_1, S3_2, Supplementary Table S4_1). Greater similarities between I marks in the different ML methods reflect more robust applications of the ML. Importantly, the correlations for the ML methods with FloWPS were always higher than for the methods without FloWPS (Table 4, Supplementary Figures S3_1, S3_2). This clearly suggests the beneficial role of FloWPS for extracting informative features from the noisy data. In this model, the biggest similarity was observed for the pair of RR and BNB methods.
Table 4

Median pairwise Pearson/Spearman correlation at feature (gene expression) importance (I) level. Figures above main diagonal: With FloWPS; figures below: Without FloWPS.

SVMRFRRBNBMLP
SVM10.53/0.550.40/0.390.37/0.340.46/0.46
RF0.34/0.4010.51/0.320.48/0.310.59/0.38
RR0.19/0.140.35/0.0410.93/0.790.89/0.52
BNB0.24/0.140.33/0.090.88/0.6410.81/0.46
MLP0.33/0.300.40/0.170.76/0.060.61/0.121

3. Discussion

Many ML methods which were designed for global separation of different classes of points in the feature space are prone to overtraining when the number of preceding cases is low. Global ML methods may also fail if there is only local rather than global order in the placement of different classes in the feature space (Figure 2a).
Figure 2

Schematic view of global-local order hybrid ML analytic pipeline (adopted after [8]; copyright belongs to the authors of [8], who wrote also the current paper). (a) Global machine learning methods may fail to separate classes for datasets without global order. (b) ML, coupled with FloWPS, works locally and handles that cases more accurately.

To improve performance of ML, FloWPS approach includes some elements of the local methods, e.g., using the flexible data trimming that avoids extrapolation in the feature space for each validation point and by selecting only several nearest neighbors from the training dataset. In such a way, the whole ML classifier becomes hybrid, both global and local (Figure 2b). In this hybrid approach, for each validation point training of ML models is performed in the individually tailored feature space. Every validation point is surrounded by a floating window from the points of the training dataset, and the irrelevant features are avoided using the rectangular projections in the feature space. This approach was initially tested for the SVM method [8,33,34,35], and in this study, we for the first time applied it to supplement other six popular ML techniques. We used twenty-one clinically annotated gene expression datasets totally, including 1778 patient samples with known clinical treatment responses. These datasets contained 41–235 samples and represented breast cancer (10) multiple myeloma (4), acute myeloid leukemia (3), pediatric acute lymphoblast leukemia (1), pediatric Wilms kidney tumor (1), low grade gliomas (1) and lung cancer (1). The chemotherapeutic treatment schemes included taxanes, bortezomib, vincristine, trastuzumab, letrozole, tipifarnib, temozolomide, busulfan and cyclophosphamide. We confirmed the efficiency of FloWPS for all tested global ML methods: Linear support vector machines (SVM), random forest (RF), binomial naïve Bayes (BNB), adaptive boosting (ADA) and multi-layer perceptron (MLP). The paired t-test for FloWPS-vs.-no-FloWPS comparison assures that the AUC values for FloWPS-empowered ML methods are significantly higher. For all the datasets tested, the use of FloWPS could increase the quality of binary classifiers for clinical response on chemotherapy. The regression-like methods, including FloWPS-assisted ML techniques, produce as their outputs the continuous values for likelihood of a sample belonging to a specific class. A discrimination threshold (τ) applied to these output values makes it possible to classify the samples as either responders or non-responders. To set up this threshold, it is important to evaluate the relative penalties of false positive and false negative errors. In most clinically relevant applications, this relative balance factor (B) varies between 0.25 and 4 [41,42,43,44,45]. For higher B values, the test sensitivity (SN) is low, and lower B means lower specificity (SP). We found that FloWPS-assisted global ML methods RF, BNB and MLP, exhibited the highest SN and SP in the range 0.25 ≤ B ≤ 4 (Supplementary Figures S1_1, S1_2, S1_3; Table 1). Our further and more detailed trial with advanced ML settings confirmed this finding, with the best results shown by the binomial naïve Bayesian (BNB) method with the settings alpha = 1.0, binarize = 0.0, fit_prior = False (Supplementary Figures S2_1, S2_2, S2_3; Table 2). When the best settings identified were applied to eleven cancer datasets with different proportions of the responders and non-responders, FloWPS again was found beneficial for all local ML techniques, and the BNB method showed the best performance (Figure 1c,g,k; Table 3). Overtraining, together with extrapolation, is very frequently considered also an Achilles heel of ML. We, therefore, tested if FloWPS helps to extract truly significant features or if it simply adapts to random noise, thus, causing overfitting. We compared four global ML methods (SVM, RF, BNB and MLP) and one local ML method (RR) to check similarities between them in terms of relative importance of distinct individual features. We confirmed that all these five ML methods were positively correlated at the level of feature importance (Table 4, Supplementary Figures S3_1, S3_2). Moreover, using FloWPS significantly enhanced such correlations in all the cases examined (Table 4 Supplementary Figures S3_1, S3_2, Supplementary Table S4_1). These results clearly suggest that FloWPS is helpful for extracting relevant information rather than merely follows the random noise and overfits the ML model. Overall, we propose that using correlations between different ML methods at the level of relative importance of distinct features may be used as an evaluation metric of ML suitability for building classifiers utilizing omics data (Table 5, Supplementary Figure S5_1). In this case, the higher is the correlation, the bigger should be the probability that the separation of responders from non-responders is robust and non-overtrained.
Table 5

Minimal, median, mean and maximal Pearson/Spearman correlation values for pairwise comparison of different ML methods with FloWPS at the level of feature importance (I).

Dataset #Dataset IDMinMedianMeanMax
1GSE250660.41/0.280.72/0.440.67/0.460.93/0.81
2GSE41998−0.02/−0.100.55/0.390.49/0.350.87/0.83
3GSE97820.37/0.190.58/0.410.62/0.410.97/0.88
4GSE397540.34/0.280.50/0.370.54/0.410.84/0.72
5GSE688710.50/0.430.62/0.600.68/0.640.95/0.93
6GSE551450.32/0.290.57/0.420.60/0.450.85/0.70
7TARGET500.34/0.570.69/0.740.66/0.720.95/0.82
8TARGET100.32/0.300.50/0.450.58/0.480.90/0.77
9TARGET20 busulfan0.63/0.550.70/0.660.76/0.700.97/0.89
10TARGET20 no busulfan0.16/0.350.63/0.530.60/0.550.92/0.79
11GSE187280.38/0.210.54/0.460.62/0.450.95/0.79
12GSE201810.33/0.170.43/0.430.56/0.430.96/0.79
13GSE201940.06/0.040.50/0.300.49/0.340.93/0.80
14GSE239880.28/0.180.46/0.350.55/0.390.96/0.82
15GSE326460.23/0.110.37/0.280.49/0.320.95/0.74
16GSE379460.40/0.260.62/0.450.62/0.440.92/0.69
17GSE428220.34/0.030.52/0.400.58/0.380.89/0.82
18GSE51220.12/−0.060.40/0.200.46/0.250.93/0.79
19GSE595150.37/0.260.47/0.470.59/0.490.96/0.74
20TCGA-LGG0.27/0.130.64/0.470.63/0.420.94/0.76
21TCGA-LC0.44/0.230.62/0.550.66/0.530.95/0.90
Surely, very few gene expression/mutation datasets have enough number of clinically annotated preceding cases that are sufficient for building any ML model. For the datasets, which does not have enough cases, the transfer learning approach may be applied. This approach implies that the ML model is trained on a bigger, similar, but quite different, dataset, and then applied to a smaller (validation) dataset. The FloWPS technique has been already tested for transfer learning, and gene expression profiles of cell cultures treated with chemotherapeutic drugs served as training datasets [33,34,35]. Another possibility is to aggregate different smaller datasets into bigger ones. For such aggregation, a new harmonizing technique, which is capable to merge arbitrary number of datasets obtained using arbitrary experimental platforms [46], may be applied. Of course, transformations in the feature space aimed to adapt it to individual preceding cases is not a new idea in ML [47,48,49]. However, our flexible data trimming approach FloWPS is different because it does not use any pre-selected analytical form of transformation kernels, but instead adapts the feature space aoristically for every particular validation case. The success of using FloWPS for the real-world gene expression datasets, including tens to hundreds of samples prompts further trials of its applicability in biomedicine and in the other fields where increased accuracy of ML classifiers is needed.

4. Materials and Methods

4.1. Clinically Annotated Molecular Datasets

We used 21 publicly available datasets, including high throughput gene expression profiles associated with clinical outcomes of the respective patients (Table 6). The biosamples were obtained from tumor biopsies before chemotherapy treatments. The outcomes were response or lack of response on the therapy used, as defined in the original reports (Table 6).
Table 6

Clinically annotated gene expression datasets used in this study.

ReferenceDataset IDDisease TypeTreatmentExperimental PlatformNumber NC of Cases (R vs. NR)Number S of Core Marker Genes
[50,51]GSE25066Breast cancer with different hormonal and HER2 statusNeoadjuvant taxane + anthracyclineAffymetrix Human Genome U133 Array235 (118 R: Complete response + partial response; 117 NR: Residual disease + progressive disease) 20
[52]GSE41998 Breast cancer with different hormonal and HER2 statusNeoadjuvant doxorubicin + cyclophosphamide, followed by paclitaxelAffymetrix Human Genome U133 Array68 (34 R: Complete response + partial response; 34 NR: Residual disease + progressive disease)11
[27]GSE9782Multiple myelomaBortezomib monotherapyAffymetrix Human Genome U133 Array169 (85 R: Complete response + partial response; 84 NR: No change + progressive disease)18
[53]GSE39754Multiple myelomaVincristine + adriamycin + dexamethasone followed by autologous stem cell transplantation (ASCT)Affymetrix Human Exon 1.0 ST Array124 (62 R: Complete, near-complete and very good partial responders, 62 NR: Partial, minor and worse)16
[54]GSE68871Multiple myelomaBortezomib-thalido-mide-dexamethasoneAffymetrix Human Genome U133 Plus98 (49 R: Complete, near-complete and very good partial responders, 49 NR: Partial, minor and worse)12
[55]GSE55145 Multiple myelomaBortezomib followed by ASCTAffymetrix Human Exon 1.0 ST Array56 (28 R: Complete, near-complete and very good partial responders, 28 NR: Partial, minor and worse) 14
[56,57]TARGET-50Pediatric kidney Wilms tumorVincristine sulfate + cyclosporine, cytarabine, daunorubicin + conventional surgery + radiation therapyIllumina HiSeq 200072 (36 R, 36 NR: See Reference [8])14
[56,58]TARGET-10 Pediatric acute lymphoblastic leukemiaVincristine sulfate + carboplatin, cyclophosphamide, doxorubicinIllumina HiSeq 200060 (30 R, 30 NR: See Reference [8])14
[56]TARGET-20 Pediatric acute myeloid leukemiaNon-target drugs (asparaginase, cyclosporine, cytarabine, daunorubicin, etoposide; methotrexate, mitoxantrone), including busulfan and cyclophosphamideIllumina HiSeq 200046 (23 R, 23 NR: See Reference [8])10
[56]TARGET-20 Pediatric acute myeloid leukemiaSame non-target drugs, but excluding busulfan and cyclophosphamideIllumina HiSeq 2000124 (62 R, 62 NR: See Reference [8])16
[59]GSE18728Breast cancerDocetaxel, capecitabineAffymetrix Human Genome U133 Plus 2.0 Array61 (23R: Complete response + partial response; 38 NR: Residual disease + progressive disease)16
[60,61]GSE20181Breast cancerLetrozoleAffymetrix Human Genome U133A Array52 (37 R: Complete response + partial response; 15 NR: Residual disease + progressive disease)11
[62]GSE20194Breast cancerPaclitaxel; (tri)fluoroacetyl chloride; 5-fluorouracil, epirubicin, cyclophosphamideAffymetrix Human Genome U133A Array52 (11 R: Complete response + partial response; 41 NR: Residual disease + progressive disease)10
[63]GSE23988Breast cancerDocetaxel, capecitabineAffymetrix Human Genome U133A Array61 (20 R: Complete response + partial response; 41 NR: Residual disease + progressive disease)18
[64]GSE32646Breast cancerPaclitaxel, 5-fluorouracil, epirubicin, cyclophosphamideAffymetrix Human Genome U133 Plus 2.0 Array115 (27 R: Complete response + partial response; 88 NR: Residual disease + progressive disease)17
[65]GSE37946Breast cancerTrastuzumabAffymetrix Human Genome U133A Array50 (27 R: Complete response + partial response; 23 NR: Residual disease + progressive disease)14
[66]GSE42822Breast cancerDocetaxel, 5-fluorouracil, epirubicin, cyclophosphamide, capecitabineAffymetrix Human Genome U133A Array91 (38 R: Complete response + partial response; 53 NR: Residual disease + progressive disease)13
[67]GSE5122Acute myeloid leukemiaTipifarnibAffymetrix Human Genome U133A Array57 (13 R: Complete response + partial response + stable disease; 44 R: Progressive disease)10
[68]GSE59515Breast cancerLetrozoleIllumina HumanHT-12 V4.0 expression beadchip75 (51 R: Complete response + partial response; 24 NR: Residual disease + progressive disease)15
[69]TCGA-LGGLow-grade gliomaTemozolomide + (optionally) mibefradilIllumina HiSeq 2000131 (100 R: Complete response + partial response + stable disease; 31 NR: Progressive disease)9
[69]TCGA-LCLung cancerPaclitaxel + (optionally),cisplatin/carboplatin, reolysinIllumina HiSeq 200041 (24 R: Complete response + partial response + stable disease; 17 NR: Progressive disease)7
The datasets preparation for the analysis included the following steps [8]: Labelling each patient as either responder or non-responder on the therapy used; For each dataset, finding top marker genes having the highest AUC values for distinguishing responder and non-responder classes; Performing the leave-one-out (LOO) cross-validation procedure to complete the robust core marker gene set used for building the ML model.

4.2. Principles of Flexible Data Trimming

We first introduced [33,34,35] flexible data trimming as a preprocessing tool for transferring to real patients the gene expression data obtained for cell cultures treated with anti-cancer drugs. Then this method was overhauled and used to increase the SVM-based classifier’s performance for the datasets that contained only gene expression data for cancer patients [8,29]. Since the number of patients with annotated case histories (when treatment method and its clinical success is known, together with the high-throughput gene expression/mutation profile) is limited, we have tailored the whole data trimming scheme to match the leave-one-out (LOO) methodology. This LOO approach in our method is employed three times [8,29]: First, it helped us to specify the core marker gene sets (see Materials and Methods), which form the feature space F = (f1,…,f) for subsequent application of data trimming; Second, it was applied for every ML prediction act for the wide range of data trimming parameters, m and k; Third, it was used for the final prediction of the treatment response for every patient and optimized (for all remaining patients) values of parameters m and k. Now let us describe flexible data trimming in more detail. Imagine that we have to classify the clinical response for a certain patient I (called patient of interest) from a given dataset. Let the whole dataset contain N patients, so that the remaining N − 1 patients form the preceding dataset D, for the patient of interest. For ML without data trimming, in the feature space F = (f1,…,f) all N – 1 remaining patients are used to build the classifier. However, in the case of FloWPS, LOO procedure will be applied to classify every sample j ≠ i from the preceding dataset D without sample i, and N − 2 remaining samples may be used for such a classification of sample j. To avoid extrapolation in the feature space, we consider the subset F of relevant features [8]. A feature f is considered relevant for the sample j if on its axis there are at least m projections from N − 2 training samples, which are larger than f(i,j), and, at the same time, at least m, which are smaller than f(i,j), when m is a non-negative integer parameter (Figure 3a). The maximum possible m value is (N – 2)/2, since if m is less than (N – 2)/2, then no relevant features may be chosen. Similarly, the minimal case of m = 0 also corresponds to no feature selection. Note that the resulting subset of relevant features F(m) will be individual for every pair of samples i and j [8].
Figure 3

Outline of floating window projective separator (FloWPS) approach. Selection of relevant features (a) and nearest neighbors (b) are schematized.

Moreover, in the space F (m) only k closest samples to sample j will be allowed for training among the remaining (N – 2) cases. As a measure for proximity, the Euclidean distance is used [8]. Here k is another integer parameter that specifies the number of nearest neighbors in the subspace of selected features (Figure 3b). The maximal possible k is N – 2, which corresponds to no training sample selection. In contrast, when k is too low, there is an increased risk of ML error, due to the presence of a too-small number of training points among the k nearest neighbors (Figure 3b). After selection of relevant features and nearest neighbors for the sample j, the ML model is trained using nearest neighbors only, and used for prediction of a clinical response, P(m,k), for the patient j. After repeating this procedure for all other j ≠ i, we obtain the area-under the ROC curve, AUC (m,k), for all, but i-th samples for fixed values of data trimming parameters m and k. The AUCi (m,k) can be then analyzed as a function of m and k [8]. Over the range of possible m and k values, we compare the AUCi function [8]. All pairs of (m,k) values that provide AUCi (m,k) > p·max (AUCi (m,k)) form the prediction-accountable set Si for the patient of interest i [8], where p is the confidence threshold, which could vary from 0.90 till 0.95 in our previous computational experiments [8]. Finally, the FloWPS prediction P for the sample of interest i, is calculated by averaging the ML predictions over the prediction-accountable set S: . By repeating this procedure for all other samples, a set of FloWPS predictions will be obtained for the whole dataset [8]. The overview of LOO cross-validation algorithm for FloWPS-empowered ML-based predictor is shown in Figure 4.
Figure 4

The algorithm of data trimming used for binomial naïve Bayes (LOO) cross-validation of the clinically annotated gene expression datasets. Indexes i and j denote samples (patients), index s denotes pairs of (m0,k0)-values in the prediction-accountable set, and indexes m and k denote the data trimming parameters.

The application of ML methods without FloWPS means that prediction is made for each sample i using the parameter values m = 0, k = N − 1, and a training dataset D (without sample i).

4.3. Application of ML Methods

All the ML calculations were performed using our R package flowpspkg.tar.gz, ffsdf available at Gitlab through the link: https://gitlab.com/borisov_oncobox/flowpspkg. This package, which was prepared for convenience of R users, is a wrapper over a Python code, which is also runnable. The Python code is based on library sklearn [40]. For the default settings trial, linear support vector machines (SVM), k nearest neighbors (kNN), random forest (RF), ridge regression (RR), binomial naïve Bayes (BNB), adaptive boosting (ADA) and multi-layer perceptron (MLP) were used with the default parameter settings for the sklearn package. For the advanced settings trial, three ML methods, which showed the best sensitivity and specificity for default settings within the range of relative balance factor 0.25 ≤ B ≤ 4, were run under the following conditions. For RF, the parameter n_estimators = 10, 30 or 100, and criterion = “gini” or “entropy” were used (totally 3 × 2 = 6 setting cases). For BNB, the parameters alpha = 0.0 or 1.0, binarize = 0.0 or 1.0, and fit_prior = True or False, were tried (totally 2 × 2 × 2 = 8 setting cases). For MLP, the parameters hidden_layer_sizes = 30 or 100, and alpha = 0.01, 0.001 or 0.0001 were checked (totally 2 × 3 = 6 setting cases). For the datasets with an unequal number of responders and non-responder samples (Table 6), linear SVM and RF calculations were done with setting class_weight = “balanced” and class_weight = “balanced_subsample”, respectively. All other parameters were used with the default settings.

4.4. False Positive Vs. False Negative Error Balance

For all ML methods, the FloWPS predictions (P) were made which were likelihoods for attribution of samples to one of the two classes (clinical responders or non-responders). The discrimination threshold (τ), which may be applied to distinguish between the two classes, should be determined according to the cost balance between false positive (FP) and false negative (FN) errors. In our previous study [8], for determination of the τ value, we considered the costs for FP and FN errors to be equal, and then maximized the overall accuracy rate, ACC = (TP + TN)/(TP + TN + FP + FN), since the class sizes were equal. In a more general case, the penalty value p = B·FP + FN is minimized; here, B is called relative balance factor. B is less than 1 for the situations when the FN error (e.g., refusal of prescription of a drug which might help the patient) is more dangerous than the FP error (e.g., prescription of a useless treatment). Contrary, B is greater than 1, when it is safer not to prescribe treatment for a patient than to prescribe it. Several practitioners of clinical diagnostic tests have different opinions on how high/low should be this balance factor. In different applications, the preferred values can be B = 4 [41,42,45], B < 0.16 [70], 4.5 < B < 5 [44], B < 5 [43], B > 10 for emergency medicine only [71], B > 5 for toxicology [72]. In case of oncological disease, B should be low when only one or few treatment options is/are available for a certain patient, because the refusal to give a treatment may cause serious harm to the patient. Contrarily, in the situation when the best treatment plan must be selected among multiple options available, the risk of wrong drug prescription will be higher, and B should be high as well. For our analyses, we used five model settings of B equal to 0.1, 0.25, 1, 4 or 10.

4.5. Feature Importance Analysis

For linear SVM, RF, RR, BNB and MLP methods and for all transcriptomic datasets tested, we calculated relative importance, I, of each gene expression feature f in the dataset, using the following attributes of ML classes in Python library sklearn [40]: For linear SVM: I = |coef_[0]|, where coef_[0] is the normal vector to the separation hyperplane between responders and non-responders in the feature space in the training model. For RF, I = |feature_importances| from the training model. For RR, , where the summation runs through every sample t in the training model. For BNB,, where the values named feature_count denote the number of samples encountered for each class c and feature f during fitting of the training model. For MLP, , where is the coefficient matrix in the first layer of the neural network for feature f of sample t in the training model. For each validation point I, the I was averaged over all predication-accountable set S.

5. Conclusions

We applied a flexible data trimming technique FloWPS to enhance performance of seven popular ML methods, including linear SVM, k nearest neighbors (kNN), random forest (RF), Tikhonov (ridge) regression (RR), binomial naïve Bayes (BNB), adaptive boosting (ADA) and multi-layer perceptron (MLP). We performed computational experiments for 21 high throughput gene expression datasets (41–235 samples per dataset) totally, including 1778 cancer patients with known responses on chemotherapy treatments. FloWPS essentially improved the classifier quality for all global ML methods (SVM, RF, BNB, ADA, MLP), where the area under the receiver-operator curve (ROC AUC) for the treatment response classifiers increased from 0.61–0.88 range to 0.70–0.94. The comparison of five best ML methods (SVM, RF, RR, BNB and MLP) at the level of relative importance for different features confirmed that ML models used here were not overtrained and that the usage of FloWPS increased the correlations between the different ML methods at the level of feature importance. For all the datasets tested, the best performance of FloWPS data trimming was observed for the BNB method, which can be valuable for further building of ML classifiers in personalized oncology.
  55 in total

1.  Identification of molecular predictors of response in a study of tipifarnib treatment in relapsed and refractory acute myelogenous leukemia.

Authors:  Mitch Raponi; Jean-Luc Harousseau; Jeffrey E Lancet; Bob Löwenberg; Richard Stone; Yi Zhang; Wayne Rackoff; Yixin Wang; David Atkins
Journal:  Clin Cancer Res       Date:  2007-04-01       Impact factor: 12.531

2.  Sequential changes in gene expression profiles in breast cancers during treatment with the aromatase inhibitor, letrozole.

Authors:  W R Miller; A Larionov; T J Anderson; D B Evans; J M Dixon
Journal:  Pharmacogenomics J       Date:  2010-08-10       Impact factor: 3.550

3.  Gene expression pathway analysis to predict response to neoadjuvant docetaxel and capecitabine for breast cancer.

Authors:  Larissa A Korde; Lara Lusa; Lisa McShane; Peter F Lebowitz; LuAnne Lukes; Kevin Camphausen; Joel S Parker; Sandra M Swain; Kent Hunter; Jo Anne Zujewski
Journal:  Breast Cancer Res Treat       Date:  2010-02       Impact factor: 4.872

4.  Effect of training-sample size and classification difficulty on the accuracy of genomic predictors.

Authors:  Vlad Popovici; Weijie Chen; Brandon G Gallas; Christos Hatzis; Weiwei Shi; Frank W Samuelson; Yuri Nikolsky; Marina Tsyganova; Alex Ishkin; Tatiana Nikolskaya; Kenneth R Hess; Vicente Valero; Daniel Booser; Mauro Delorenzi; Gabriel N Hortobagyi; Leming Shi; W Fraser Symmans; Lajos Pusztai
Journal:  Breast Cancer Res       Date:  2010-01-11       Impact factor: 6.466

5.  Changes in expression of oestrogen regulated and proliferation genes with neoadjuvant treatment highlight heterogeneity of clinical resistance to the aromatase inhibitor, letrozole.

Authors:  William R Miller; Alexey Larionov
Journal:  Breast Cancer Res       Date:  2010-07-20       Impact factor: 6.466

6.  Optimal type I and type II error pairs when the available sample size is fixed.

Authors:  John P A Ioannidis; Iztok Hozo; Benjamin Djulbegovic
Journal:  J Clin Epidemiol       Date:  2013-05-09       Impact factor: 6.437

7.  Shambhala: a platform-agnostic data harmonizer for gene expression data.

Authors:  Nicolas Borisov; Irina Shabalina; Victor Tkachev; Maxim Sorokin; Andrew Garazha; Andrey Pulin; Ilya I Eremin; Anton Buzdin
Journal:  BMC Bioinformatics       Date:  2019-02-06       Impact factor: 3.169

8.  FLOating-Window Projective Separator (FloWPS): A Data Trimming Tool for Support Vector Machines (SVM) to Improve Robustness of the Classifier.

Authors:  Victor Tkachev; Maxim Sorokin; Artem Mescheryakov; Alexander Simonov; Andrew Garazha; Anton Buzdin; Ilya Muchnik; Nicolas Borisov
Journal:  Front Genet       Date:  2019-01-15       Impact factor: 4.599

9.  A Machine Learning Approach for Identifying Gene Biomarkers Guiding the Treatment of Breast Cancer.

Authors:  Ashraf Abou Tabl; Abedalrhman Alkhateeb; Waguih ElMaraghy; Luis Rueda; Alioune Ngom
Journal:  Front Genet       Date:  2019-03-27       Impact factor: 4.599

10.  Data Mining and Machine Learning Models for Predicting Drug Likeness and Their Disease or Organ Category.

Authors:  Abraham Yosipof; Rita C Guedes; Alfonso T García-Sosa
Journal:  Front Chem       Date:  2018-05-09       Impact factor: 5.221

View more
  7 in total

1.  Bioinformatics Methods in Medical Genetics and Genomics.

Authors:  Yuriy L Orlov; Ancha V Baranova; Tatiana V Tatarinova
Journal:  Int J Mol Sci       Date:  2020-08-28       Impact factor: 5.923

2.  Cancer gene expression profiles associated with clinical outcomes to chemotherapy treatments.

Authors:  Nicolas Borisov; Maxim Sorokin; Victor Tkachev; Andrew Garazha; Anton Buzdin
Journal:  BMC Med Genomics       Date:  2020-09-18       Impact factor: 3.063

3.  Editorial: Next Generation Sequencing Based Diagnostic Approaches in Clinical Oncology.

Authors:  Anton Buzdin; Ira Ida Skvortsova; Xinmin Li; Ye Wang
Journal:  Front Oncol       Date:  2021-01-28       Impact factor: 6.244

4.  System, Method and Software for Calculation of a Cannabis Drug Efficiency Index for the Reduction of Inflammation.

Authors:  Nicolas Borisov; Yaroslav Ilnytskyy; Boseon Byeon; Olga Kovalchuk; Igor Kovalchuk
Journal:  Int J Mol Sci       Date:  2020-12-31       Impact factor: 5.923

5.  Machine Learning Applicability for Classification of PAD/VCD Chemotherapy Response Using 53 Multiple Myeloma RNA Sequencing Profiles.

Authors:  Nicolas Borisov; Anna Sergeeva; Maria Suntsova; Mikhail Raevskiy; Nurshat Gaifullin; Larisa Mendeleeva; Alexander Gudkov; Maria Nareiko; Andrew Garazha; Victor Tkachev; Xinmin Li; Maxim Sorokin; Vadim Surin; Anton Buzdin
Journal:  Front Oncol       Date:  2021-04-15       Impact factor: 6.244

6.  Recent Trends in Cancer Genomics and Bioinformatics Tools Development.

Authors:  Anastasia A Anashkina; Elena Y Leberfarb; Yuriy L Orlov
Journal:  Int J Mol Sci       Date:  2021-11-10       Impact factor: 5.923

Review 7.  Machine Learning: A New Prospect in Multi-Omics Data Analysis of Cancer.

Authors:  Babak Arjmand; Shayesteh Kokabi Hamidpour; Akram Tayanloo-Beik; Parisa Goodarzi; Hamid Reza Aghayan; Hossein Adibi; Bagher Larijani
Journal:  Front Genet       Date:  2022-01-27       Impact factor: 4.599

  7 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.