Literature DB >> 22927888

Recursive feature selection with significant variables of support vectors.

Chen-An Tsai1, Chien-Hsun Huang, Ching-Wei Chang, Chun-Houh Chen.   

Abstract

The development of DNA microarray makes researchers screen thousands of genes simultaneously and it also helps determine high- and low-expression level genes in normal and disease tissues. Selecting relevant genes for cancer classification is an important issue. Most of the gene selection methods use univariate ranking criteria and arbitrarily choose a threshold to choose genes. However, the parameter setting may not be compatible to the selected classification algorithms. In this paper, we propose a new gene selection method (SVM-t) based on the use of t-statistics embedded in support vector machine. We compared the performance to two similar SVM-based methods: SVM recursive feature elimination (SVMRFE) and recursive support vector machine (RSVM). The three methods were compared based on extensive simulation experiments and analyses of two published microarray datasets. In the simulation experiments, we found that the proposed method is more robust in selecting informative genes than SVMRFE and RSVM and capable to attain good classification performance when the variations of informative and noninformative genes are different. In the analysis of two microarray datasets, the proposed method yields better performance in identifying fewer genes with good prediction accuracy, compared to SVMRFE and RSVM.

Entities:  

Mesh:

Year:  2012        PMID: 22927888      PMCID: PMC3426197          DOI: 10.1155/2012/712542

Source DB:  PubMed          Journal:  Comput Math Methods Med        ISSN: 1748-670X            Impact factor:   2.238


1. Introduction

Feature selection is a critical issue for disease subtyping and cancer classification. There are two goals when optimizing classification procedures, attaining highest accuracy and selecting smallest set of features. With the development of microarray technology, experimenters can detect expression profiles for more than ten thousands gene at a time. Classification problem with such huge amount of genes lead to inefficiency, inconsistency, and bias. There are many discussions on reducing the features by univariate rankings of gene and selecting the genes with highest rankings to build a classifier, for example, t-statistics (Golub et al. [1], Furey et al. [2], and Li et al. [3]) and F-score (Chen and Lin [4]) related ranking methods. Feature selection is the most challenging task in the pattern classification research especially for high-dimensional data. Depending on the classification algorithms, feature selection techniques can be classified into three main groups: embedded, filter, and wrapper approaches. The filter methods rank features according to some criteria or indices of relevance which are completely independent of the classification algorithm such as P value. The filter approach is a stand-alone prior step, regardless of which classification algorithm will be used. Afterwards, the selected feature subset will be applied to the classification algorithm. The effects of the selected features on the performance of the algorithm are not taken into account. Classical classification algorithms, such as Fisher's linear discriminant analysis and k-nearest neighbour, often use the filter approach to select relevant predictors prior to classification (Chen and Lin [4], Roepman et al. [5], Mark and Kung [6], Pavlidis et al. [7], and Yu and Liu [8]). In the wrapper approach, the model selection algorithms are wrapped in the search process of feature subsets. This has the advantages that the feature selection process can take into account feature dependencies while building the classification model. However, these methods have high computational complexity to repeatedly training and testing predictors each time a new subset is defined. West et al. [9] used binary regression coefficients as importance in scoring genes to the contribution to the classification. Díaz-Uriarte and Alvarez de Andrés [10] proposed the use of Gini index as variable importance to perform gene selection in the classification algorithm of random forest. Sharma et al. [11] proposed a wrapper-like feature selection method based on null linear discriminant analysis (LDA) method. Embedded methods are an integral part of feature selection techniques and specific classification algorithms, for example, decision trees or neural networks with regularization. In addition, Sharma et al. [12] proposed a successive feature selection algorithm to iteratively eliminate redundant features with minimal information in terms of classification accuracy. Their method combines filters and wrappers together to search for the best top-r feature subset. A recent review (Saeys et al. [13]) summarized many more feature selection techniques and discussed their use for bioinformatics applications. Over the recent years, support vector machine (SVM; Cristianini and Shawe-Taylor [14]), a supervised machine learning, is widely used in classification problem especially with high-dimensional data such as microarray gene expression profile. SVM maps input data points to construct maximal-marginal hyperplane in higher dimension space to classify data with different class labels. The hyperplane is constructed using only the support vectors (i.e., data that lie on the margin). The general form of the hyperplane is represented as where the training set is of size n, x is the input data of sample i  (i = 1,2,…, n), α is the Lagrange multiplier solved from the training set. y ∈ (−1, 1) is the class label for sample i, K(x, x ) is the kernel function, x is the variable vector of a sample, and b can be viewed as intercept. The optimized weight vector equals to ∑ α y ϕ(x), where α is nonzero if i belongs to support vector, otherwise α equals to 0, and ϕ(·) is a nonlinear mapping function from input space to feature space. Support vector machine has many applications and performs very well on microarray related classification problems (Furey et al. [2] and Brown et al. [15]). In addition, many gene selection methods based on SVM have been proposed previously. For example, Guyon et al. [16] proposed a support vector machine recursive feature elimination algorithm (SVMRFE), which uses the coefficients of weight vector to compute the feature ranking score. However, like the concept of slope, the selected genes with higher coefficients of weight means that they will have higher expression values compared to nonselected genes. Hence, some noisy but high-expression value genes have high possibility to be selected. On the other hand, Zhang et al. [17] proposed the recursive SVM feature selection (RSVM), which combines the weight and data information (i.e., class mean) to formulate selection criterion. Such method takes all data information into consideration and the outlier data is also included. However, such feature selection result is greatly affected by class label assignment. There is no general rule for assigning the positive and negative signs to the two classes. Hence, such unstable criterion is difficult to use. Hence, we propose a new feature selection criterion SVM-t based on the use of t-statistics embedded in support vector machine. We use the univariate ranking method on support vectors for identifying significant genes and the backward elimination procedure follows the workflow of RSVM with nested subsets of feature. The aim of combining procedures is to identify more significant genes among the nearest support vectors. In this paper, we compare the performance of the three SVM-based gene selection methods via extensive simulations and real microarray data analyses.

2. Materials and Methods

2.1. SVMRFE

In general, all classification problems can be generalized to two-class classification problem. If there are more than two-classes, the simplest and widely used approach is the multiple one-against-all scheme. Hence, multiple class problems can be reduced to multiple simple two-class problem. Consider a binary decision function with linear kernel, the function in (1) can be represented as: where w is weight vector of the classifier. The optimized weight vector equals to ∑ α y x, where α is nonzero if i belongs to support vector, otherwise α equals to 0. Hence, for support vector machine, the binary decision function is determined by the linear combination of support vectors. The class of new sample x can be easily determined by the sign of f(x). For achieving the objective of better feature selection, Guyon et al. [16] proposed SVMRFE method to find important gene subset for cancer classification. To evaluate the importance of features, SVMRFE uses w 2 as ranking criterion. Features with the smallest ranking scores are eliminated. This criterion can also be interpreted as weighted sums of support vectors [17]. In addition, based on SVMRFE, Duan et al. [18] used cross-validation method to estimate multiple w 2 and take signal-to-noise ratio of the multiple weight value as the ranking criterion. Hence, the multiple weights SVMRFE is robust to small perturbations for single experiment.

2.2. RSVM

To enhance the robustness to noise and outliers, Zhang et al. [17] proposed recursive support vector machine (i.e., RSVM). They developed a feature selection method to construct the stepwise classification rules and reduce the gene numbers at the same time. From (2), one can obtain the weight of specific feature. Unlike SVMRFE, RSVM selects important features by the value of the product of weight and corresponding feature: where w denotes weight of the jth feature, m + and m − stand for the means of feature j in the respective classes. This method takes both the classifier information (i.e., weight) and the data (i.e., class mean of two classes) into consideration. With the ranking score calculated from (3), we can set a threshold to filter out any features with score below it in the next steps. Furthermore, the classification can be performed with the selected features step by step. After a series of iterations with smaller and smaller feature sets, the best rule is constructed by a prespecified number of features selected from the highest selected-frequency list. Finally, the best rule and number of selected features can be decided simultaneously. However, such feature selection result is greatly affected by class label assignment. We found that different assignments of class labelling (+ or −) will result in different selected gene sets. Hence, such unstable criterion is ambiguous to use.

2.3. Proposed SVM-t Method

Support vector machine uses only the information of support vectors to construct the maximal separation hyperplane and determine the classes for new samples. The support vectors, the set of closest points between two classes, play an important role in SVMRFE and RSVM for feature selection. These two methods use the weights of corresponding features to build the selection criteria; instead, we combined the univariate ranking methods (i.e., absolute t-statistics) with support vector machine. The proposed method uses the most important subset (i.e., SVs) of the data points to construct the selection criteria (4). In other words, we use the algorithm of support vector machine as a sampling technique for data points of two classes simultaneously. The standard two-sample t-statistic is used as a surrogate statistic to evaluate the significant differences between two classes. Our proposed method is easy to implement and keeping the computational complexity comparable to that of SVMRFE and RSVM. Therefore, with the variation of samples, we can identify the most significant differences for specific genes among the closest points: where n + (resp., n −) indicates the number of support vectors for class +1 (resp., −1). We can calculate mean u + (resp., u −) and standard deviation s + (resp., s −) by using only the support vectors of feature j-labelled class +1 (resp., −1) to obtain the score of each feature. The features with highest scores are the features with most significant difference between the two classes. It is intuitive to select features with the highest score as our feature set. However, there are some exceptions according to such a criterion. We further break down the selection criterion into four situations as follows: the data set can be well separated by two SVs. Both classes have only one SV. Equation (4) can thus be reduced as the difference of particular feature of SVs; one class has only one SV, and the other class has more than one SV, and (4) becomes a single mean t-test statistics. Considering class −1 with only one SV, (4) reduces to similar to (b), class +1 has only one SV. The selection criterion is finally, in the most general condition, we usually encounter classification problems with more than one support vectors for each class. For practical use, this condition is more complicated than previous ones, and selection criterion in (4) shall be used. We follow the workflow of feature selection suggested in [17] as in Figure 1. First, a monotone decreasing sequence {d ,…, d 0} for the number of selected features is specified where k is the total number of runs. After executing the selection criteria for each run, the number of input features d will be determined. Secondly, the cross-validation method is adopted to carry out gene selection procedure and execute the SVM procedures with d features. For the comparison among the difference of selection methods, we apply these three different criteria in this step. For each run, the features with top d ranking scores are selected until the sequence ends. Finally, the gene set with minimal CV error is selected. The set with fewest genes will be chosen when ties occur for CV errors.
Figure 1

Workflow of feature selection. Workflow of the SVM-t algorithm.

3. Results

3.1. Simulation Experiments

First, we evaluate performance of the three methods using simulated data sets. Our simulation data contains different means and standard deviations for informative and noninformative genes. We first generate a training set with 100 samples (50 samples for each class), each containing expression values of 1000 genes where 300 of them are informative genes and the rest 700 noninformative genes. To validate the selected gene sets, we also generate another 1000 independent samples (500 samples for each class). We performed 100 simulations for each data set and used “Leave-one-out” version CV method. The following two simulation cases use the same aforementioned scheme with varying parameter setting.

Simulation Case I

In this simulation, we separate the informative gene set into two parts. The first 150 genes independently follow the Gaussian distribution N(0.15, 0.5) for class 1 and N(−0.15, 0.5) for class 2. On the contrary, the last 150 genes are independently distributed from the Gaussian distribution N(−0.15, 0.5) for class 1 and N(0.15, 0.5) for class 2. The rest 700 “noninformative” genes independently follow N(0,1) distribution. Table 1 summarizes the simulation result with 100 runs using average and standard deviation of the 100 runs with the percentage of informative genes coverage and the number of support vectors (SVs). With this parameter setting, the informative genes cannot be easily differentiated due to the small mean difference. However, taking the standard deviation into consideration, we found that the proposed method shows better results than RSVM and SVMRFE.
Table 1

Comparison of selection methods on simulation data 1.

Num. of genesRSVMSVMRFESVM-t
Test (%)Rec (%)nSVTest (%)Rec (%)nSVTest (%)Rec (%)nSV
8001.06 ± 0.3798.97 ± 0.5196.42 ± 0.081.07 ± 0.3798.37 ± 0.6398.25 ± 0.040.97 ± 0.3799.59 ± 0.3795.36 ± 0.07
600 0.94 ± 0.3693.76 ± 1.2692.90 ± 0.120.99 ± 0.3992.49 ± 1.4497.26 ± 0.090.67 ± 0.2998.52 ± 0.6187.45 ± 0.45
500 0.92 ± 0.3887.38 ± 1.7090.62 ± 0.160.95 ± 0.3885.95 ± 1.9195.96 ± 0.130.44 ± 0.2496.96 ± 0.8781.14 ± 1.09
400 1.02 ± 0.4277.00 ± 2.0584.91 ± 0.211.01 ± 0.3876.34 ± 2.2090.52 ± 0.200.19 ± 0.1593.48 ± 1.3471.48 ± 1.06
300 1.26 ± 0.5463.12 ± 2.0676.51 ± 0.291.31 ± 0.6062.99 ± 2.0980.02 ± 0.300.04 ± 0.0784.28 ± 1.4760.63 ± 0.91
200 2.14 ± 0.9267.68 ± 2.8961.99 ± 0.372.07 ± 0.9368.56 ± 2.8363.93 ± 0.400.04 ± 0.694.89 ± 1.7849.37 ± 0.99
150 2.88 ± 1.1371.35 ± 3.8356.73 ± 0.402.61 ± 1.2572.76 ± 3.7059.05 ± 0.370.10 ± 0.1297.79 ± 1.4943.67 ± 1.21
100 4.36 ± 1.9374.62 ± 4.7543.35 ± 0.323.75 ± 1.7677.62 ± 4.6144.48 ± 0.280.53 ± 0.2899.14 ± 1.0536.07 ± 0.75
90 4.58 ± 2.0276.32 ± 4.7345.30 ± 0.304.01 ± 1.9179.34 ± 4.7746.57 ± 0.310.79 ± 0.3499.14 ± 1.0034.64 ± 0.82
80 5.20 ± 2.2177.29 ± 4.8542.66 ± 0.314.21 ± 1.8681.33 ± 4.8043.66 ± 0.321.14 ± 0.4499.41 ± 1.0132.81 ± 0.69
70 5.64 ± 2.6178.84 ± 5.2839.27 ± 0.334.47 ± 1.9083.10 ± 5.1240.04 ± 0.271.66 ± 0.5899.60 ± 0.8630.81 ± 0.54
60 6.41 ± 2.5680.37 ± 5.3935.21 ± 0.265.02 ± 1.9785.62 ± 5.1435.82 ± 0.322.57 ± 0.7299.65 ± 1.0428.62 ± 0.36
50 7.43 ± 2.5782.44 ± 5.8530.89 ± 0.275.77 ± 2.0988.06 ± 5.6731.21 ± 0.253.91 ± 0.8999.62 ± 0.9326.32 ± 0.41

Num. of genes: number of selected features in each recursive step. Results of 1000 features are the same and we do not list them here.

Test: test error rate.

Rec: the percentage of recovery informative genes (among 300 informative genes).

nSV: the average number of support vectors selected in each recursive step.

In addition, we increase the standard deviation for informative genes to 1 so that the difference of variation effect between informative and noninformative genes is reduced. Table 2 shows that RSVM yields better average test error rate and selects more informative genes than SVMRFE and the proposed method. However, the proposed method selected fewer support vectors in each recursive step.
Table 2

Comparison of selection methods on simulation data 2.

Num. of genesRSVMSVMRFESVM-t
Test (%)Rec (%)nSVTest (%)Rec (%)nSVTest (%)Rec (%)nSV
8006.15 ± 0.8491.24 ± 1.5596.60 ± 0.096.06 ± 0.8590.48 ± 1.4098.01 ± 0.066.24 ± 0.8291.59 ± 1.3296.02 ± 0.14
600 6.44 ± 0.8880.45 ± 2.2694.58 ± 0.146.28 ± 0.8678.86 ± 2.2797.60 ± 0.096.72 ± 0.8881.19 ± 1.9791.29 ± 0.41
500 6.64 ± 0.9373.63 ± 2.3992.98 ± 0.156.51 ± 0.9171.95 ± 2.2697.15 ± 0.107.15 ± 0.9774.24 ± 2.1487.72 ± 0.99
400 7.23 ± 1.0465.58 ± 2.3989.59 ± 0.257.11 ± 1.0363.74 ± 2.4595.19 ± 0.217.91 ± 1.2365.67 ± 2.4881.56 ± 1.58
300 8.07 ± 1.0155.53 ± 2.1883.60 ± 0.368.05 ± 1.1953.62 ± 2.2388.72 ± 0.309.13 ± 1.5055.35 ± 2.4073.08 ± 1.79
200 10.05 ± 1.4863.29 ± 3.1771.21 ± 0.5110.32 ± 1.5161.08 ± 3.1174.08 ± 0.5111.90 ± 1.5962.46 ± 2.8561.41 ± 1.52
150 11.80 ± 1.8267.97 ± 3.5265.14 ± 0.4512.16 ± 1.8365.53 ± 3.8769.57 ± 0.4514.16 ± 1.6867.13 ± 3.2954.77 ± 1.21
100 14.81 ± 1.9274.24 ± 4.1550.51 ± 0.4015.37 ± 1.9471.13 ± 4.1252.82 ± 0.4918.09 ± 2.0472.74 ± 3.5645.75 ± 0.50
90 15.70 ± 1.7675.37 ± 4.1752.23 ± 0.4016.24 ± 1.9372.36 ± 4.5854.89 ± 0.4319.28 ± 2.3774.02 ± 3.9843.81 ± 0.40
80 16.70 ± 1.8776.80 ± 4.1449.25 ± 0.4617.34 ± 1.8874.05 ± 4.7251.68 ± 0.4020.51 ± 2.2675.19 ± 4.4741.50 ± 0.39
70 17.94 ± 1.8678.27 ± 4.4045.24 ± 0.4318.61 ± 1.8575.73 ± 4.3447.22 ± 0.3822.13 ± 2.3576.69 ± 4.5039.12 ± 0.47
60 19.28 ± 1.8780.15 ± 4.8540.49 ± 0.4119.91 ± 2.1277.18 ± 4.8342.17 ± 0.4123.91 ± 2.3677.63 ± 5.0336.59 ± 0.66
50 21.21 ± 1.8782.22 ± 4.7935.40 ± 0.3721.55 ± 1.9479.06 ± 4.9436.58 ± 0.3625.83 ± 2.6679.90 ± 5.5933.86 ± 0.64

Num. of genes, Test, Rec, and nSV: the same as Table 1.

Based on the previous setting, we further increase the variance of noninformative genes. We set the standard deviation of noninformative gene twice as large as that of informative genes and follows N(0,2) distribution. Table 3 shows that all three methods give worse test error rates than before because of the more complex nature of the simulation scheme. The coverage percentages of informative genes by RSVM and SVMRFE decrease with the decreasing gene levels, but the proposed method maintains high coverage percentages of informative genes. Besides, the test error rate of proposed method stays at around 25% which is lower than the error rates of RSVM and SVMRFE.
Table 3

Comparison of selection methods on simulation data 3.

Num. of genesRSVMSVMRFESVM-t
Test (%)Rec (%)nSVTest (%)Rec (%)nSVTest (%)Rec (%)nSV
80026.39 ± 1.6586.52 ± 1.6097.68 ± 0.0426.66 ± 1.7684.41 ± 1.7598.63 ± 0.0326.13 ± 1.6591.56 ± 1.3897.07 ± 0.05
600 27.36 ± 1.6467.53 ± 2.0696.16 ± 0.127.99 ± 1.7065.72 ± 2.2698.41 ± 0.0425.46 ± 1.7481.20 ± 2.0991.60 ± 0.29
500 28.63 ± 1.7256.83 ± 2.1895.08 ± 0.1129.12 ± 1.6755.57 ± 2.2798.17 ± 0.0525.05 ± 1.7074.70 ± 2.0887.77 ± 1.05
400 30.42 ± 1.9144.77 ± 2.4092.24 ± 0.1630.76 ± 1.9544.57 ± 2.3496.53 ± 0.1024.31 ± 2.1766.70 ± 2.1580.84 ± 1.18
300 32.98 ± 1.9832.58 ± 2.3087.17 ± 0.2033.05 ± 2.2132.90 ± 2.4190.65 ± 0.2523.30 ± 2.2956.71 ± 2.1971.85 ± 1.17
200 36.59 ± 2.2130.20 ± 3.0074.61 ± 0.3036.49 ± 2.0831.66 ± 2.8476.09 ± 0.3122.26 ± 2.5365.50 ± 2.9560.63 ± 0.66
150 38.31 ± 2.0429.55 ± 3.6568.60 ± 0.3338.09 ± 2.2231.37 ± 3.6171.66 ± 0.3521.60 ± 2.9371.07 ± 3.7954.36 ± 0.48
100 40.85 ± 2.2027.15 ± 4.5053.31 ± 0.3340.37 ± 2.4530.40 ± 4.4054.78 ± 0.3321.53 ± 3.1977.63 ± 4.9146.03 ± 0.49
90 41.27 ± 2.1927.43 ± 4.7654.61 ± 0.3040.63 ± 2.5230.49 ± 5.0956.22 ± 0.3021.90 ± 3.3778.89 ± 5.0644.07 ± 0.54
80 41.62 ± 1.9627.61 ± 4.8851.27 ± 0.3040.61 ± 2.3931.05 ± 5.4452.55 ± 0.3022.44 ± 3.5180.35 ± 5.3241.95 ± 0.56
70 41.99 ± 2.4027.56 ± 5.2947.08 ± 0.2841.17 ± 2.6131.47 ± 5.8448.03 ± 0.3222.90 ± 3.4181.69 ± 5.4039.66 ± 0.64
60 42.51 ± 2.1626.98 ± 5.3442.10 ± 0.2741.56 ± 2.5631.90 ± 6.1742.73 ± 0.2724.18 ± 3.1283.02 ± 5.4136.98 ± 0.63
50 42.99 ± 2.5626.84 ± 6.2736.77 ± 0.2741.53 ± 2.7132.52 ± 6.5937.05 ± 0.2725.56 ± 3.3084.58 ± 6.1034.18 ± 0.66

Num. of genes, Test, Rec, and nSV: the same as Table 1.

Simulation Case II

Here, we increase the mean values of the informative genes from 0.15 to 0.25 and keep the distribution of the 700 “noninformative” at N(0,1). In Table 4, we find that this simulated data is well separated with the higher difference between the two classes. The test error rate and percentage of selected informative genes are comparable among the three methods. The prediction results are almost perfect (i.e., test errors are nearly zero) for all conditions in this simulation. The proposed method utilizes fewer support vectors than the other two methods though.
Table 4

Comparison of selection methods on simulation data 4.

Num. of genesRSVMSVMRFESVM-t
Test (%)Rec (%)nSVTest (%)Rec (%)nSVTest (%)Rec (%)nSV
8000.00 ± 0.00 100.00 93.82 ± 0.090.00 ± 0.00 100.00 96.88 ± 0.060.00 ± 0.00 100.00 92.54 ± 0.10
600 0.00 ± 0.0099.95 ± 0.1388.22 ± 0.150.00 ± 0.0099.88 ± 0.1994.37 ± 0.130.00 ± 0.00 100.00 82.03 ± 0.77
500 0.00 ± 0.0099.68 ± 0.3282.72 ± 0.200.00 ± 0.0099.39 ± 0.4590.14 ± 0.170.00 ± 0.0099.99 ± 0.0673.90 ± 1.23
400 0.00 ± 0.0098.02 ± 0.7770.36 ± 0.240.00 ± 0.0096.80 ± 1.0177.60 ± 0.280.00 ± 0.0099.94 ± 0.1461.64 ± 0.83
300 0.00 ± 0.0087.36 ± 1.3757.45 ± 0.320.00 ± 0.0085.18 ± 1.4363.84 ± 0.390.00 ± 0.0097.26 ± 0.7646.16 ± 0.33
200 0.00 ± 0.0094.49 ± 1.6346.07 ± 0.310.00 ± 0.0093.15 ± 1.5848.56 ± 0.320.00 ± 0.0099.94 ± 0.1837.89 ± 1.60
150 0.00 ± 0.0096.91 ± 1.6142.99 ± 0.330.00 ± 0.0096.60 ± 1.8045.50 ± 0.350.00 ± 0.0099.96 ± 0.1633.85 ± 2.04
100 0.00 ± 0.0098.66 ± 1.2833.32 ± 0.290.00 ± 0.0098.72 ± 1.3634.71 ± 0.320.00 ± 0.0099.89 ± 0.3527.87 ± 1.13
90 0.00 ± 0.0099.03 ± 1.0936.60 ± 0.310.00 ± 0.0098.92 ± 1.3138.08 ± 0.350.00 ± 0.0099.88 ± 0.4426.93 ± 1.30
80 0.00 ± 0.0099.33 ± 1.0734.94 ± 0.280.00 ± 0.0099.35 ± 0.9836.35 ± 0.290.00 ± 0.0099.90 ± 0.3425.55 ± 1.13
70 0.01 ± 0.0299.43 ± 0.9732.49 ± 0.290.01 ± 0.0299.60 ± 0.8633.62 ± 0.330.01 ± 0.0399.90 ± 0.4223.97 ± 0.95
60 0.01 ± 0.0499.55 ± 0.8829.40 ± 0.270.02 ± 0.0499.53 ± 0.9230.31 ± 0.320.03 ± 0.0699.88 ± 0.4322.30 ± 0.67
50 0.03 ± 0.0799.76 ± 0.7125.91 ± 0.260.04 ± 0.0799.78 ± 0.6926.65 ± 0.290.09 ± 0.199.84 ± 0.5520.47 ± 0.48

Num. of genes, Test, Rec, and nSV: the same as Table 1.

We further set standard deviation of noninformative to 1 to eliminate the variation effect between informative and noninformative genes. The result in Table 5 is very similar to that of Table 2 with lower test error rates for the three methods. With decreasing gene number, RSVM gives better average test error rate than SVMRFE and proposed method. The percent of selected informative genes are comparable among the three methods which may due to the higher difference of informative genes between two classes. In addition, our proposed method also selected fewer support vectors.
Table 5

Comparison of selection methods on simulation data 5.

Num. of genesRSVMSVMRFESVM-t
Test (%)Rec (%)nSVTest (%)Rec (%)nSVTest (%)Rec (%)nSV
8000.05 ± 0.0798.09 ± 0.7892.07 ± 0.080.05 ± 0.0797.26 ± 0.9095.12 ± 0.070.05 ± 0.0798.44 ± 0.7091.05 ± 0.11
600 0.06 ± 0.0894.89 ± 1.1488.21 ± 0.130.04 ± 0.0693.00 ± 1.3793.74 ± 0.100.05 ± 0.0795.35 ± 1.0983.85 ± 0.72
500 0.06 ± 0.0891.47 ± 1.3985.42 ± 0.150.05 ± 0.0788.91 ± 1.7692.29 ± 0.120.06 ± 0.0892.08 ± 1.5879.30 ± 1.23
400 0.06 ± 0.0886.07 ± 1.8180.50 ± 0.170.06 ± 0.0882.49 ± 2.0088.29 ± 0.170.08 ± 0.0985.86 ± 1.9072.49 ± 1.66
300 0.07 ± 0.0875.82 ± 2.0373.51 ± 0.240.08 ± 0.1071.78 ± 2.1580.72 ± 0.280.11 ± 0.1175.15 ± 2.0063.95 ± 1.81
200 0.17 ± 0.1387.61 ± 2.0061.22 ± 0.290.21 ± 0.1682.71 ± 2.3065.46 ± 0.330.30 ± 0.2485.78 ± 2.3553.12 ± 1.72
150 0.34 ± 0.2192.33 ± 2.2356.24 ± 0.300.40 ± 0.2088.59 ± 2.5160.89 ± 0.310.69 ± 0.3790.68 ± 2.6747.09 ± 1.60
100 1.17 ± 0.4496.48 ± 1.9943.75 ± 0.321.28 ± 0.3893.68 ± 2.5946.26 ± 0.352.15 ± 0.6594.29 ± 2.6338.94 ± 0.68
90 1.45 ± 0.4797.01 ± 2.1146.10 ± 0.261.63 ± 0.4994.88 ± 2.4948.90 ± 0.342.74 ± 0.7495.09 ± 2.7137.32 ± 0.63
80 1.90 ± 0.5997.83 ± 1.9743.71 ± 0.322.08 ± 0.5695.70 ± 2.4946.21 ± 0.313.52 ± 0.8895.88 ± 2.9235.30 ± 0.45
70 2.51 ± 0.6598.26 ± 1.8840.37 ± 0.302.79 ± 0.6796.49 ± 2.6642.40 ± 0.324.60 ± 1.0696.74 ± 2.5033.16 ± 0.32
60 3.52 ± 0.7198.82 ± 1.5436.36 ± 0.313.72 ± 0.8297.33 ± 2.4837.93 ± 0.305.92 ± 1.2597.52 ± 2.5930.80 ± 0.39
50 4.94 ± 0.8799.10 ± 1.4631.99 ± 0.265.00 ± 0.8998.32 ± 2.0033.11 ± 0.297.93 ± 1.4397.88 ± 2.4628.32 ± 0.45

Num. of genes, Test, Rec, and nSV: the same as Table 1.

Finally, we set the standard deviation of noninformative gene twice as large as that of informative genes with distribution N(0,2). The three methods all yield higher test error rate than previous settings but the proposed method performs better than RSVM and SVMRFE with significant margins for such complex simulation data.

Stability of Feature Selection

To investigate the stability of feature selection, we perform 200 times of simulation case I with switched class labels. The frequency plot of selected features in Figure 2 shows that different class labels in RSVM will result in selecting the different sets of features, while our proposed method does not alter the frequencies of selected features by switching the class labels. In addition, the RSVM tends to select noninformative genes much more frequently than our approach. Figure 3 plots the Jaccard's coefficient of RSVM to comparing the similarity of feature selections between two switched class labels over 200 simulations. We observe that the class labelling in RSVM is crucial for achieving better agreement of feature selection. The low Jaccard's coefficient reveals that there is heterogeneity between two different class labels for some simulations. In contrast to the RSVM, our proposed method is independent of the class labels and appears to be better than RSVM for selecting a small number of discriminative genes.
Figure 2

Stability of feature selection for switching class labels. Grey vertical and solid black lines represent frequencies of selected features for different class labels. The first 150 genes and the last 150 genes are discriminative genes.

Figure 3

Jaccard's coefficients over 200 simulations for RSVM.

3.2. Application on the Human Breast Cancer Data

We next evaluated performance of the three methods by using two microarray datasets. The first dataset is from Affymetrix DNA microarray analysis of a series of primary breast cancer samples [9]. This data contains gene expression profiles of 7129 detected probe sets for 49 samples (24 ER positive and 25 negative patients). Preprocessings including background correction, normalization, PM correction and expression summary were conducted, and the gene expression levels were log2-transformed. All 7129 genes were included in the gene selection process. The minimal gene number is set at 5 with decreasing ration 0.85 for generating the decreasing gene selection sequence (Table 6).
Table 6

Comparison of selection methods on simulation data 6.

Num. of genesRSVMSVMRFESVM-t
Test (%)Rec (%)nSVTest (%)Rec (%)nSVTest (%)Rec (%)nSV
8004.96 ± 0.9096.92 ± 0.9096.88 ± 0.075.02 ± 0.9195.78 ± 1.0798.43 ± 0.044.72 ± 0.8698.63 ± 0.6495.99 ± 0.07
600 4.91 ± 0.9687.79 ± 1.6294.16 ± 0.115.09 ± 1.0386.10 ± 1.6697.77 ± 0.073.86 ± 0.8395.76 ± 1.1188.66 ± 0.44
500 5.30 ± 1.0579.05 ± 1.9792.60 ± 0.135.46 ± 1.1577.77 ± 2.1197.07 ± 0.113.06 ± 0.7592.73 ± 1.2983.23 ± 1.19
400 5.88 ± 1.2867.74 ± 2.3388.34 ± 0.216.10 ± 1.3167.05 ± 2.5693.20 ± 0.202.09 ± 0.7287.26 ± 1.6674.71 ± 1.06
300 7.34 ± 1.6464.62 ± 2.6381.00 ± 0.297.44 ± 1.6264.78 ± 2.9783.88 ± 0.271.22 ± 0.5392.60 ± 2.1064.77 ± 1.00
200 9.99 ± 1.9856.40 ± 2.9766.42 ± 0.369.85 ± 2.0557.43 ± 2.9868.16 ± 0.350.85 ± 0.4688.83 ± 1.9453.17 ± 0.80
150 11.91 ± 2.4158.77 ± 3.5860.70 ± 0.3711.40 ± 2.2460.39 ± 3.6263.23 ± 0.381.00 ± 0.5193.74 ± 2.2546.96 ± 0.83
100 14.84 ± 2.9160.30 ± 4.2246.77 ± 0.3013.49 ± 2.8264.10 ± 4.2847.76 ± 0.352.19 ± 0.7896.98 ± 2.3639.08 ± 0.48
90 15.62 ± 3.1761.60 ± 4.6848.39 ± 0.2913.62 ± 2.9065.53 ± 4.8649.56 ± 0.302.67 ± 0.9597.36 ± 2.1137.39 ± 0.58
80 15.85 ± 3.3763.10 ± 5.3045.47 ± 0.3314.23 ± 3.1567.33 ± 5.3746.45 ± 0.333.43 ± 0.9097.93 ± 2.0235.39 ± 0.41
70 16.70 ± 3.6264.47 ± 5.7841.75 ± 0.3114.67 ± 3.4669.53 ± 5.7042.48 ± 0.314.30 ± 1.0398.19 ± 1.9033.25 ± 0.36
60 17.56 ± 3.8865.48 ± 6.3837.47 ± 0.2915.25 ± 3.5571.52 ± 5.9937.87 ± 0.285.69 ± 1.2998.65 ± 1.7530.95 ± 0.38
50 18.80 ± 4.0166.70 ± 6.9132.81 ± 0.2815.78 ± 3.5474.04 ± 6.6333.03 ± 0.297.42 ± 1.3299.10 ± 1.6728.44 ± 0.41

Num. of genes, Test, Rec, and nSV: the same as Table 1.

Table 7 shows that RSVM selects fewer discriminant genes than SVMRFE and the proposed method. However, in earlier recursive steps, the proposed method attains the minimal CV error rate (2.04%). Proposed method may select more significant genes than SVMRFE and RSVM in the first few iterations. Taking also gene number into consideration we may choose 10 genes with CV error rate (8.16%) as our solution. SVMRFE can also attain CV error rate (8.16%) with 20 selected genes and RSVM can attain minimal error rate (4.08%) with 5 genes. For gene number fewer than 20, proposed method has higher number of support vectors than SVMRFE and RSVM, and these two methods have almost identical number of support vectors. Table 8 shows that proposed method selected 10 genes with P value < 0.0004.
Table 7

The CV results on the breast cancer dataset.

Num. of genesRSVM SVM-RFESVM-t
CV2Mean SVCV2Mean SVCV2Mean SV
71298.16% 47.02 8.16% 47.02 8.16% 47.02
6060 8.16% 45.94 8.16% 47.02 6.12% 46.88
5151 8.16% 45.31 8.16% 47.02 6.12% 44.88
4378 8.16% 44.55 8.16% 47.02 6.12% 42.84
3721 8.16% 43.20 8.16% 47.00 6.12% 39.55
3163 8.16% 41.98 8.16% 47.00 4.08% 37.41
2689 8.16% 40.92 8.16% 46.84 2.04% 35.67
2286 10.20% 39.63 10.20% 46.78 2.04% 35.18
1943 10.20% 38.41 12.24% 46.61 4.08% 34.31
1652 10.20% 37.47 12.24% 46.06 4.08% 32.43
1404 10.20% 36.41 12.24% 45.39 4.08% 32.31
1193 12.24% 35.31 12.24% 44.80 6.12% 30.90
1014 12.24% 34.12 12.24% 44.39 6.12% 30.24
862 12.24% 33.04 12.24% 43.14 6.12% 29.76
733 12.24% 32.10 12.24% 42.27 4.08% 29.02
623 12.24% 30.84 10.20% 41.31 6.12% 27.10
530 12.24% 30.10 12.24% 40.96 6.12% 26.16
450 12.24% 29.29 12.24% 39.49 6.12% 24.35
382 12.24% 29.00 10.20% 38.04 8.16% 23.37
325 10.20% 28.78 12.24% 37.49 8.16% 22.33
276 10.20% 28.14 10.20% 36.59 6.12% 20.78
235 12.24% 27.55 12.24% 35.27 6.12% 19.94
200 10.20% 26.67 12.24% 34.14 8.16% 19.24
170 10.20% 25.98 10.20% 33.53 8.16% 18.14
144 14.29% 25.12 12.24% 31.67 10.20% 17.94
122 14.29% 23.84 12.24% 30.67 12.24% 17.63
104 12.24% 22.12 12.24% 29.65 12.24% 17.53
88 10.20% 21.02 14.29% 27.67 14.29% 16.76
75 10.20% 20.39 14.29% 25.92 12.24% 16.78
64 12.24% 19.39 8.16% 24.61 14.29% 16.43
54 12.24% 18.37 8.16% 22.14 18.37% 16.33
46 10.20% 18.10 8.16% 21.57 20.41% 15.57
39 14.29% 16.61 10.20% 19.24 16.33% 15.57
33 14.29% 15.55 10.20% 16.86 16.33% 15.04
28 16.33% 13.98 10.20% 15.96 16.33% 14.31
24 16.33% 13.04 10.20% 15.02 18.37% 13.69
20 12.24% 12.02 8.16% 12.78 22.45% 13.39
17 10.20% 11.20 14.29% 11.65 22.45% 12.76
14 8.16% 10.29 12.24% 10.59 12.24% 12.43
12 6.12% 9.16 14.29% 9.73 14.29% 12.04
10 4.08% 8.35 14.29% 8.69 8.16% 12.16
8 6.12% 7.29 16.33% 7.51 10.20% 12.22
7 6.12% 7.22 16.33% 6.86 16.33% 12.86
6 8.16% 6.29 16.33% 6.20 12.24% 13.43
5 4.08% 5.98 18.37% 6.20 12.24% 13.63

Num. of genes: the number of selected genes in the recursive steps. CV2: total cross-validation error rate (CV2 error rate). Mean SV: mean number of support vectors used in the cross-validation at each level.

Table 8

Selected biomarkers on the breast cancer dataset.

Probe IdRankGene titleGene symbol t-statistics P value
X03635_at1Estrogen receptor 1ESR1 9.509 4.08E−09
U79293_at2Thrombospondin, type I, domain containing 4THSD4 8.198 2.21E−07
U39840_at3Forkhead box A1FOXA1 7.149 4.78E−06
X17059_s_at5N-acetyltransferase 1 (arylamine N-acetyltransferase)NAT1 7.175 4.78E−06
L08044_s_at10Trefoil factor 3 (intestinal)TFF3 6.328 4.14E−05
X83425_at12Basal cell adhesion molecule (Lutheran blood group)BCAM 6.028 1.08E−04
M59815_at14Complement component 4A (Rodgers blood group) /// complement Component 4B (Childo blood group)C4A /// C4B 5.927 1.27E−04
X76180_at17Sodium channel, nonvoltage-gated 1 alphaSCNN1A 5.796 1.69E−04
X53002_s_at63Integrin, beta 5ITGB5 4.639 2.71E−03
U52522_at70ADP-ribosylation factor interacting protein 2 (arfaptin 2)ARFIP2 4.533 3.51E−03

Rank: gene-ranking is computed using P values.

3.3. Application on the Lung Cancer Data

The second dataset is a lung cancer dataset used in CAMDA 2003 [19]. Gene expression profiles of 7129 genes for 86 lung cancer patients (67 stage I and 19 stage III tumors) are available for analysis. Similar preprocessing and parameter setting to the breast cancer analysis are adopted. Error rates shown in Table 9 tell us that this lung cancer data is more difficult than the breast cancer data for predicting the disease subtypes. SVM-RFE and RSVM can only reach error rate at above 20%, while the proposed method attains CV error rate below 10% at several feature levels. Table 10 shows the selected genes by the proposed method, and most of them have P values less than 0.05.
Table 9

The CV results on the lung cancer dataset.

Num. of genesRSVMSVM-RFESVM-t
CV2Mean SVCV2Mean SVCV2Mean SV
712922.09% 51.34 22.09% 51.34 22.09% 51.34
6060 20.93% 46.78 22.09% 51.35 20.93% 50.85
5151 20.93% 45.07 22.09% 51.40 19.77% 50.20
4378 20.93% 44.24 22.09% 51.42 18.60% 49.24
3721 20.93% 43.99 22.09% 51.35 15.12% 48.22
3163 20.93% 43.53 22.09% 51.34 15.12% 47.60
2689 20.93% 43.38 22.09% 51.36 15.12% 47.38
2286 20.93% 43.12 20.93% 51.86 15.12% 46.71
1943 22.09% 43.17 20.93% 52.79 12.79% 46.05
1652 22.09% 43.26 20.93% 52.51 12.79% 45.10
1404 22.09% 43.12 20.93% 51.84 13.95% 43.78
1193 22.09% 42.87 20.93% 52.33 10.47% 42.60
1014 22.09% 42.86 20.93% 51.15 8.14% 41.84
862 22.09% 42.24 22.09% 51.64 8.14% 41.50
733 22.09% 41.47 22.09% 52.21 9.30% 41.03
623 22.09% 41.55 23.26% 51.72 12.79% 40.07
530 22.09% 40.85 23.26% 51.28 11.63% 39.71
450 22.09% 40.40 23.26% 50.59 10.47% 39.20
382 22.09% 39.08 23.26% 50.70 9.30% 38.34
325 22.09% 38.62 23.26% 50.77 11.63% 37.43
276 22.09% 38.52 24.42% 49.21 9.30% 36.62
235 22.09% 38.71 24.42% 48.27 11.63% 36.06
200 22.09% 37.28 24.42% 47.67 12.79% 35.59
170 22.09% 36.33 24.42% 45.84 12.79% 35.13
144 22.09% 36.19 23.26% 44.64 15.12% 33.88
122 24.42% 35.43 24.42% 42.81 15.12% 33.48
104 24.42% 34.30 24.42% 40.87 15.12% 32.59
88 24.42% 32.87 29.07% 39.02 16.28% 31.99
75 24.42% 31.28 26.74% 36.56 17.44% 31.92
64 26.74% 29.84 25.58% 34.08 16.28% 31.80
54 26.74% 27.41 25.58% 30.07 17.44% 31.97
46 26.74% 25.94 26.74% 27.50 15.12% 31.90
39 25.58% 24.42 25.58% 25.10 12.79% 32.30
33 23.26% 22.53 29.07% 23.09 11.63% 32.66
28 20.93% 20.42 26.74% 19.87 15.12% 32.97
24 20.93% 18.86 24.42% 19.31 13.95% 33.16
20 23.26% 16.94 25.58% 16.59 16.28% 33.51
17 23.26% 15.83 24.42% 15.71 16.28% 33.58
14 22.09% 16.42 23.26% 16.15 17.44% 34.08
12 24.42% 17.72 26.74% 17.03 16.28% 34.55
10 25.58% 19.76 26.74% 19.95 13.95% 35.15
8 24.42% 22.45 29.07% 22.60 15.12% 36.01
7 24.42% 23.64 27.91% 24.16 16.28% 36.44
6 29.07% 25.35 27.91% 25.97 17.44% 36.87
5 27.91% 27.86 30.23% 28.73 18.60% 37.50

Num. of genes: the number of selected genes in the recursive steps. CV2: total cross-validation error rate (CV2 error rate). Mean SV: mean number of support vectors used in the cross-validation at each level.

Table 10

Selected biomarkers on the lung cancer dataset.

MarkerRankGene titleGene symbol t-statistics P value
D13705_s_at 1 Cytochrome P450, family 4, subfamily A, polypeptide 11CYP4A11 5.743 9.52E−04
HG1103-HT1103_at 2 −5.434 1.77E−03
X07618_s_at 5 Cytochrome P450, family 2, subfamily D, polypeptide 6CYP2D6 4.785 9.89E−03
Z71460_at 9 ATPase, H+ transporting, lysosomal V0 subunit a1ATP6V0A1 4.644 0.0124
AFFX-HUMGAPDH/M33197_M_at 10 Glyceraldehyde-3-phosphate dehydrogenaseGAPDH −4.496 0.0131
M37583_at 13 H2A histone family, member ZH2AFZ −4.445 0.0132
AFFX-HUMGAPDH/M33197_5_at 19 Glyceraldehyde-3-phosphate dehydrogenaseGAPDH −4.161 0.0153
AFFX-HUMRGE/M10098_3_at 20 4.098 0.0153
D13748_at 22 Eukaryotic translation initiation factor 4A, isoform 1EIF4A1 −4.245 0.0153
HG2279-HT2375_at 23 −4.099 0.0153
L14922_at 26 Replication factor C (activator 1) 1, 145 kDaRFC1 −4.138 0.0153
U37143_at 35 Cytochrome P450, family 2, subfamily J, polypeptide 2CYP2J2 4.245 0.0153
HG1153-HT1153_at 48 −4.031 0.0157
HG651-HT5209_s_at 49 3.995 0.0157
X01677_f_at 60 Glyceraldehyde-3-phosphate dehydrogenaseGAPDH −4.008 0.0157
U25433_at 65 CATR tumorigenicity conversion 1CATR1 3.971 0.0161
U38864_at 70 Zinc finger protein 212ZNF212 3.941 0.0166
X77584_at 100 ThioredoxinTXN −3.807 0.0186
AFFX-HUMGAPDH/M33197_3_at 101 Glyceraldehyde-3-phosphate dehydrogenaseGAPDH −3.774 0.0191
L34060_at 103 Cadherin 8, type 2CDH8 3.789 0.0191
U96769_rna1_at 108 ChondroadherinCHAD 3.782 0.0191
HG2855-HT2995_at 122 −3.714 0.0205
U66617_at 128 SWI/SNF related, matrix associated, actin dependent regulator of chromatin, subfamily d, member 1SMARCD1 3.706 0.0205
U77970_at 191 Neuronal PAS domain protein 2NPAS2 3.544 0.0237
L25085_at 195 Sec61 beta subunitSEC61B −3.535 0.024
D45248_at 243 Proteasome (prosome, macropain) activator subunit 2 (PA28 beta)PSME2 −3.407 0.0285
U55764_at 273 Sulfotransferase family 1E, estrogen-preferring, member 1SULT1E1 3.359 0.0303
J03258_at 295 Vitamin D (1,25-dihydroxyvitamin D3) receptorVDR −3.314 0.0316
L19067_at 430 v-rel reticuloendotheliosis viral oncogene homolog A, nuclear factor of kappa light polypeptide gene enhancer in B-cells 3, p65 (avian)RELA −3.106 0.0424
D38583_at 433 S100 calcium binding protein A11S100A11 −3.098 0.0431
HG2815-HT2931_at 475 −3.039 0.0469
U20350_at 610 Chemokine (C-X3-C motif) receptor 1CX3CR1 2.891 0.0567
U08815_at 772 Splicing factor 3a, subunit 3, 60 kDaSF3A3 2.743 0.0683

Rank: gene-ranking is computed using P values.

4. Discussion

The proposed SVM-t method is a multivariate feature selection approach in which sample heterogeneity is accommodated during a sequential backward elimination process. The method could benefit from the use of support vectors on the hyperplane of SVM and select a group of informative genes with aids in improving classification performance. The use of support vectors for feature selection is intuitive because they play an important role for building the decision function in SVM. For the linear SVM, the feature dimensions corresponding to the support vectors are also critical for classification by the property of linearity. Therefore, we consider trimmed t-statistics which compare group mean values after removing the nonsupport vectors and show that these have higher power to identify relevant features than the other two SVM-based approaches in case of high noise. The result for simulation studies show that all three feature selection methods of SVMRFE, RSVM, and the proposed method have good performance when the distribution of differentially expressed genes is much different from noise genes. The influences could be attributed to two sources: the magnitude of difference and variance of gene expression levels between two classes. When fixing the magnitude of difference, the result is greatly affected by the variance. With identical variance of informative and noninformative genes, RSVM performs better result since it gives more weights to informative genes (i.e., genes with higher difference between the two classes become more important in this case). On the other hand, when considering distinct variances of informative and noninformative genes, the proposed method outperforms RSVM and SVMRFE. Moreover, the proposed method is able not only to select more informative genes with lower test error rate than RSVM and SVMRFE but also to accommodate heterogeneity within the complex clinical samples. In applications, we use two published microarray datasets to evaluate the performance of the three methods via leave-one-out cross-validated accuracy. We can find that proposed method can consistently select a smaller subset of informative genes with good prediction accuracy. Both the proposed method and RSVM outperform SVMRFE for applications on the two datasets. All SVM-based feature selection methods under comparisons select features based on the support vectors in the linear SVM. Ideally, SVMs can use a proper kernel function to map the data into a separating hyperplane in the feature space when the data points are not linearly separable. However, the corresponding property between support vectors and feature dimensions in the nonlinear SVM is far from being clear and some further work is required in order to fully understand this issue. In addition, it is well known that applying proper kernel functions and parameters to a specific real database remains challenging, and the number of support vectors that compose the decision function increases dramatically when the decision manifold becomes complicated. Therefore, our approach is limited to the use of linear SVM here and considers it as an alternative approximation to the real dataset. However, if the objective is to select relevant features for better classification performance, rather than to provide insights into relative importance of features in the feature space, then our approach provides an alternative solution. Further work will be focused on the impact of our trimmed t-statistics on the use of nonlinear SVMs for classification of nonlinear datasets.
  11 in total

1.  Support vector machine classification and validation of cancer tissue samples using microarray expression data.

Authors:  T S Furey; N Cristianini; N Duffy; D W Bednarski; M Schummer; D Haussler
Journal:  Bioinformatics       Date:  2000-10       Impact factor: 6.937

2.  Knowledge-based analysis of microarray gene expression data by using support vector machines.

Authors:  M P Brown; W N Grundy; D Lin; N Cristianini; C W Sugnet; T S Furey; M Ares; D Haussler
Journal:  Proc Natl Acad Sci U S A       Date:  2000-01-04       Impact factor: 11.205

3.  A comparative study of feature selection and multiclass classification methods for tissue classification based on gene expression.

Authors:  Tao Li; Chengliang Zhang; Mitsunori Ogihara
Journal:  Bioinformatics       Date:  2004-04-15       Impact factor: 6.937

4.  A top-r feature selection algorithm for microarray gene expression data.

Authors:  Alok Sharma; Seiya Imoto; Satoru Miyano
Journal:  IEEE/ACM Trans Comput Biol Bioinform       Date:  2012 May-Jun       Impact factor: 3.710

5.  Multiple SVM-RFE for gene selection in cancer classification with expression data.

Authors:  Kai-Bo Duan; Jagath C Rajapakse; Haiying Wang; Francisco Azuaje
Journal:  IEEE Trans Nanobioscience       Date:  2005-09       Impact factor: 2.935

Review 6.  A review of feature selection techniques in bioinformatics.

Authors:  Yvan Saeys; Iñaki Inza; Pedro Larrañaga
Journal:  Bioinformatics       Date:  2007-08-24       Impact factor: 6.937

7.  Predicting the clinical status of human breast cancer by using gene expression profiles.

Authors:  M West; C Blanchette; H Dressman; E Huang; S Ishida; R Spang; H Zuzan; J A Olson; J R Marks; J R Nevins
Journal:  Proc Natl Acad Sci U S A       Date:  2001-09-18       Impact factor: 11.205

8.  An expression profile for diagnosis of lymph node metastases from primary head and neck squamous cell carcinomas.

Authors:  Paul Roepman; Lodewyk F A Wessels; Nienke Kettelarij; Patrick Kemmeren; Antony J Miles; Philip Lijnzaad; Marcel G J Tilanus; Ronald Koole; Gert-Jan Hordijk; Peter C van der Vliet; Marcel J T Reinders; Piet J Slootweg; Frank C P Holstege
Journal:  Nat Genet       Date:  2005-01-09       Impact factor: 38.330

9.  Gene-expression profiles predict survival of patients with lung adenocarcinoma.

Authors:  David G Beer; Sharon L R Kardia; Chiang-Ching Huang; Thomas J Giordano; Albert M Levin; David E Misek; Lin Lin; Guoan Chen; Tarek G Gharib; Dafydd G Thomas; Michelle L Lizyness; Rork Kuick; Satoru Hayasaka; Jeremy M G Taylor; Mark D Iannettoni; Mark B Orringer; Samir Hanash
Journal:  Nat Med       Date:  2002-07-15       Impact factor: 53.440

10.  Recursive SVM feature selection and sample classification for mass-spectrometry and microarray data.

Authors:  Xuegong Zhang; Xin Lu; Qian Shi; Xiu-Qin Xu; Hon-Chiu E Leung; Lyndsay N Harris; James D Iglehart; Alexander Miron; Jun S Liu; Wing H Wong
Journal:  BMC Bioinformatics       Date:  2006-04-10       Impact factor: 3.169

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.