Literature DB >> 18973864

A DSRPCL-SVM approach to informative gene analysis.

Wei Xiong1, Zhibin Cai, Jinwen Ma.   

Abstract

Microarray data based tumor diagnosis is a very interesting topic in bioinformatics. One of the key problems is the discovery and analysis of informative genes of a tumor. Although there are many elaborate approaches to this problem, it is still difficult to select a reasonable set of informative genes for tumor diagnosis only with microarray data. In this paper, we classify the genes expressed through microarray data into a number of clusters via the distance sensitive rival penalized competitive learning (DSRPCL) algorithm and then detect the informative gene cluster or set with the help of support vector machine (SVM). Moreover, the critical or powerful informative genes can be found through further classifications and detections on the obtained informative gene clusters. It is well demonstrated by experiments on the colon, leukemia, and breast cancer datasets that our proposed DSRPCL-SVM approach leads to a reasonable selection of informative genes for tumor diagnosis.

Entities:  

Mesh:

Year:  2008        PMID: 18973864      PMCID: PMC5054100          DOI: 10.1016/S1672-0229(08)60023-6

Source DB:  PubMed          Journal:  Genomics Proteomics Bioinformatics        ISSN: 1672-0229            Impact factor:   7.691


Introduction

With the rapid development of DNA microarray technology, we can now get the expression levels of thousands of genes via one single experiment with a relative cheap cost 1., 2.. These microarray data are essential in analyzing health situation of the human body and recognizing symptoms of human illnesses. Mathematically, they can be expressed as a gene expression matrix X = (x), where each row represents a gene, while each column represents a sample or a patient for tumor diagnosis. That is, the numerical value x denotes the expression level of a specific gene i at a particular sample j. As a matter of fact, there are many microarray datasets available on the web. For medical diagnosis and treatment, it is very important to select or discover informative genes of a tumor via the analysis of microarray data, since the informative genes can not only provide valuable information for discovering the crucial reasons of the tumor as well as the treatment methods, but also support to construct an efficient tumor diagnosis system from their expression levels directly without any influence of the other irrelevant genes. Actually, there have already been many elaborate methods for informative gene selection. However, most of them are based on ranking genes according to a kind of criterion, such as t, F, rank sum and χ2 test statistics 3., 4., 5., 6., 7. and the information criterion (. Recently, the up-and down-regulation probabilities for each gene were defined and then the informative genes can be successfully selected according to the decrease rank of the absolute difference values between the two regulation probabilities 9., 10.. Generally, these statistical and information methods just select a number of top genes (the number is fixed or determined by the threshold given to the criterion). In this way, informative genes are selected through individual gene evaluations and thus the relations among the genes are neglected, which may lead to an incomplete selection of informative genes for tumor analysis and diagnosis. The relations or structures of genes can be discovered through unsupervised classification or clustering. In fact, some typical clustering methods have been already applied to the analysis of informative genes and tumor diagnosis, such as hierarchical clustering 4., 11., 12., k-means ( and self-organizing map (SOM) (. However, as the number of genes is large, hierarchical clustering is very difficult to be implemented and cannot determine the informative genes by itself. Moreover, the other classical clustering methods like k-means and SOM require predetermination of the number of gene clusters. However, we usually do not know the number of gene clusters since it depends on the structures of genes that are implied in the microarray data. In 1993, a new kind of unsupervised clustering method, called the rival penalized competitive learning (RPCL) algorithm 14., 15., was proposed to automatically determine the number of clusters during the clustering or competitive learning on the sample data. For each input sample, the basic idea is that not only the weight vector of the winner unit is modified to adapt to the input, but also the weight vector of its rival (the 2nd winner) is de-learned by a smaller learning rate. In this way, as the learning and de-learning rates are properly selected and the number of units or weight vectors is larger than the number of actual clusters in the sample data, the RPCL algorithm can automatically allocate an appropriate number of weight vectors for a sample dataset, with the other extra weight vectors being driven far away from the sample data. Recently, the RPCL algorithm was generalized to the distance sensitive rival penalized competitive learning (DSRPCL) algorithm through a cost function theory (. Actually, the implementation of the DSRPCL algorithm becomes more efficient and easier since the learning rate can be easily set. Thus, we can use the DSRPCL algorithm to analyze gene clusters from microarray data without knowing the actual number of gene clusters. In fact, the RPCL and DSRPCL algorithms have already been used to analyze microarray data 17., 18.. In Nair et al. (, the RPCL algorithm was applied to classify the genes into a number of clusters that could have certain functional meanings. Moreover, the DSRPCL algorithm was utilized to classify the genes into a number of clusters from which a compact set of informative genes could be established via a post-filtering gene selection method (. Similarly to the DSRPCL algorithm, the cooperative competition clustering algorithm ( was established to classify the genes into an appropriate number of clusters and then the informative genes can be selected from each of the obtained clusters. As the DSRPCL algorithm can automatically divide the genes into a number of functional clusters, we may wonder whether there exists a cluster that can be served as a set of informative genes directly. Actually, we can utilize support vector machine (SVM) ( to check which gene cluster contributes best to the tumor diagnosis and whether it can be used as an informative gene set for a tumor. That is, we can use SVM to train a tumor diagnosis system with the m expression profiles on the genes in each cluster and to see which tumor diagnosis system gets the best prediction accuracy. If a tumor diagnosis system gets the best prediction accuracy that is high enough, we consider the corresponding gene cluster is just the set of informative genes for the tumor. Moreover, since the distributions of gene expression levels on the samples should have the similar structures at the critical or powerful informative genes, that is, the informative genes that can strongly discriminate the tumor from the normal via their expression levels on the samples, we can consider these critical or powerful informative genes belong to a sub-cluster in this informative gene set. Thus, we can find these critical or powerful informative genes of the tumor by further clustering and checking on the sub-clusters of this informative gene cluster or set. As for the critical informative gene selection, Guyon et al. have already proposed an SVM-based method (. They used a linear SVM to train a tumor diagnosis system and selected the genes with higher weights in the final discriminate function as the critical or powerful genes. Recently, Zhang et al. improved this SVM-based method by adding the smoothly clipped absolute deviation penalty on the original objective function of the SVM (. However, the major disadvantage of the SVM-based method is that the number of informative genes is still determined by the pre-assumed threshold value to the weights, not by the structure of the genes. In this paper, in light of the above ideas, we propose a DSRPCL-SVM approach where the DSRPCL algorithm is firstly utilized to classify the genes expressed through the microarray data into a number of clusters and the informative gene cluster or set is then detected with the help of SVM. Moreover, the critical or powerful informative genes is found through further classifications and detections on the informative gene clusters. The performance of this method is demonstrated by experiments on the colon, leukemia, and breast cancer datasets.

Method

The DSRPCL algorithm for gene clustering

Given a microarray dataset X = (x) with n genes and m samples, we let , where X = [x, x, …, x] represents the μ-th gene through its expression levels over all the m samples. Suppose that X is just an input to a simple competitive learning network, that is, a layer of k competitive units. These competitive units are dominated by the corresponding weight vectors W = [w, w, …, w] for i = 1, 2, …, k. All the weight vectors can be represented by a big vector W = vec[W1, W2, …, W]. For each input X, the basic idea of the DSRPCL algorithm is that not only the weight vector of the winner unit (the closest weight vector to the sample) is modified to adapt to the input, but also the weight vectors of the rivals or losers (the other weight vectors) are punished to keep away from the input. As a weight vector diverges to infinity, the corresponding cluster becomes empty and can be canceled. Therefore, we can automatically obtain the number of gene clusters as well as the centers of these clusters assuming k is larger than the true number of the actual gene clusters. As a result, the genes are automatically divided into several clusters by classifying each gene into the cluster whose center is closest to it. Theoretically, the DSRPCL algorithm can be realized by minimizing the cost function in Equation 1, where c(μ) is the index of the winner unit for the μ-th gene, W is the nearest weight vector for X, and P is a positive constant. Ma and Wang ( obtained the derivatives of E(W) with respect to w as in Equation 2, where δ is the Kronecker function. With these derivatives, the DSRPCL algorithm is designed as a kind of gradient-descent algorithm. Table 1 summarizes the details of the DSRPCL algorithm and its variants, where we denote it as the batch DSRPCL algorithm. The DSRPCL1 algorithm is the adaptive DSRPCL algorithm, and the DSRPCL2 algorithm modifies only the rival weight vector (the second winner) so that E2(W) is only affected by the largest term with r(μ), which is consistent with the original RPCL algorithm 14., 15.. The other variant of the DSRPCL algorithm is the simulated annealing rival penalized competitive learning (SARPCL) by applying the simulated annealing mechanism to the DSRPCL1 algorithm. The stopping threshold value ε is a pre-fixed small positive number. Parameters k0, k1, c0 and c1 are positive constant numbers that can be selected by experience. Since these DSRPCL algorithms have the similar functions in clustering analysis, we typically use the DSRPCL1 algorithm in our experiments.
Table 1

The DSRPCL algorithm and its variants

Randomly initialize the vector W1(0),,Wk(0), and let T = 0.

Update Wi with a learning rate η (0 ≤ η ≤ 1):

Batch DSRPCL:ΔWi=ηE(W)Wi={ημ(XμWi),ifi=c(μ),ημXμWiP2(XμWi),otherwise.

DSRPCL1:ΔWi={η(XμWi),ifi=c(μ),ηXμWiP2(XμWi),otherwise.

DSRPCL2:ΔWi={η(XμWi),ifi=c(μ),ηXμWiP2(XμWi),ifi=r(μ),0,otherwise.

SARPCL:

Let λ=e(k1Tk0), η = η0/(c1 T + c0) and t = 0.

Randomly select Xμ from S = {X1, …, Xn}, and take ξ ∼ Uniform[0, 1].

ΔWi={η(XμWi),ifi=c(μ),ηXμWiP2(XμWi),otherwise.

If ξλ, let ΔWi = −ΔWi.

If t < M, let t = t + 1 and return to STEP b).

If λ < ε, stop.

If |E(W)(T+1)E(W)(T)| > ε1, let T = T + 1, and return to STEP 2; otherwise, stop.

According to the properties of the DSRPCL algorithm shown in Ma and Wang (, when k is selected to be large enough, the DSRPCL algorithm can detect the number of gene clusters during the clustering. From the obtained gene clusters, we can check them with SVM and find the informative gene cluster or set.

The DSRPCL-SVM approach to informative gene analysis

We further consider the informative gene analysis through the DSRPCL algorithm and SVM. In order to do so, we can implement the DSRPCL algorithm directly on the sample data from the microarray data with respect to a tumor. Usually, we can overestimate the number of the clusters in S and set it to be k. As a result of the implementation, the DSRPCL algorithm divides the n genes (represented by X) into a number of functional gene clusters among which there is an informative gene cluster contributing the best to the recognition or diagnosis of the tumor. If we think this informative gene cluster is too large, that is, it contains too many genes for tumor analysis, or we just want to get some more critical or powerful informative genes for the tumor, we can implement the DSRPCL algorithm on the previously obtained informative gene cluster or set (the obtained subset of S). Again, the DSRPCL algorithm divides the selected informative genes into a number of gene clusters, among which we can find the most powerful sub-cluster and genes. In such a way, we can finally find a small number of critical or powerful genes for the tumor. On the other hand, the above informative gene search can be only done with the help of a supervised learning system for checking which gene cluster or sub-cluster contributes the best to the recognition or diagnosis of the tumor. Actually, suppose is a divided gene cluster or sub-cluster obtained by the DSRPCL algorithm, if the genes in this cluster (or sub-cluster) are informative (or powerful informative) to the tumor, then a supervised learning system on the sample expression profiles of these genes, , and their corresponding diagnosis results (if the sample is tumorous, the diagnosis result is 1; otherwise the diagnosis result is 0) will lead to the highest prediction accuracy or the lowest average error. In this way, we can find the informative or powerful informative genes through a supervised learning system. SVM ( is a powerful learning machine for the supervised learning or classification. In fact, many researchers have used it to analyze microarray data and demonstrated its advantages 24., 25., 26.. Therefore, we also utilize SVM to check the informative or powerful informative gene cluster. For comparison, we use the MATLAB toolbox OSUSVM 3.0 to set up three kinds of SVMs, including linear SVM (no kernel), 3-poly SVM (cubic polynomial kernel), and radial basis function SVM (RBF kernel).

Evaluation

To test the effectiveness of our proposed DSRPCL-SVM approach for informative gene analysis, we conducted experiments on three real-world microarray datasets: Colon cancer dataset. It contains expression profiles of 2, 000 genes in 22 normal tissues and 40 colon tumor tissues, which can be retrieved from the web at http://microarray.princeton.edu/oncology/affydata/index.html. In our experiment, we used the training set (22 normal and 22 tumorous tissues) and the test set (18 tumorous tissues) provided at the website. Leukemia dataset. It contains expression profiles of 7, 129 genes in 47 acute lymphoblastic leukemia (ALL) and 25 acute myeloid leukemia (AML) samples, which can be retrieved from the web at http://www.genome.wi.mit.edu/MPR/data_set_ALL_AML.htm. In our experiments, we used the training set (27 ALL and 11 AML) and the test set (20 ALL and 14 AML) provided at the website. Breast cancer dataset. It contains expression profiles of 5, 776 genes in 14 normal tissues and 13 breast tumorous tissues, which can be retrieved from the web at http://genome-www.stanford.edu/breast_cancer/sbcmp/data.shtml. In our experiment, we used the 14 normal and 13 tumorous samples provided at the website as both the training set and the test set since the number of the samples is very limited. For the three kinds of SVMs, there are two parameters in the last two SVMs, γ and C, and their selection affects the performance of SVM. In our experiments, by experience we set γ = 0.02, C = 0.05 on colon cancer data and γ = 0.002, C = 10 on leukemia data and breast cancer data. To improve the efficiency of the DSRPCL1 algorithm, we set the de-learning rate in the update rule, that is, the corresponding learning rate for the opposite or minus direction learning, to attenuate to zero with the number of iterations. In this way, the algorithm could make the convergent weight vectors converge to the centers of the actual clusters in the sample data without any deviation, keeping the extra weight vectors being driven far away from the sample data. For the convenience of implementation of the two algorithms, we filtered out 10% obviously irrelevant genes using the cosine method. That is, we computed the cosine value of the vector of the expressions of each gene at all the samples with the reference vector in which each element is 1, and then ranked these genes with the decrease of the cosine value and finally filtered out 10% genes from the last. Moreover, we normalized the expression vectors of the remaining genes for our analysis. First, we ran the DSRPCL1 algorithm on the three microarray datasets and obtained five, four, and nine gene clusters (except the empty clusters represented by the weight vectors of the DSRPCL1 algorithm), respectively, where the number of the weight vectors or clusters was always set initially by 10 (the number of clusters in each microarray data was assumed to be no more than 10). According to these gene clusters obtained by the DSRPCL1 algorithm, we trained the three kinds of SVMs with the training data on each gene cluster of the three datasets. Then we obtained the prediction or classification accuracy of each trained SVM with the test data. The experiment results on the three datasets are summarized in Table 2.
Table 2

Experimental results of the first DSRPCL-SVM procedure on three datasets

DatasetGene clusterNo. of genesClassfication accuracy
Linear SVMPoly SVMRBF SVM
Colon cancer13810.88890.94440.3889
21820.77780.55560.3889
33850.88890.88890.5000
4 (optimal)4350.94440.88891
54180.83330.94440.8889

Leukemia12,7690.95450.95450.8182
21,9390.81820.77270.7273
3 (optimal)1,7080.95450.95450.9545
410.181800

Breast cancer130.66670.55560.5185
220.44440.55560.5185
3 (optimal)1,580111
4926111
5771111
6486111
7175111
8672111
9584111
From Table 2, we can find that the DSRPCL1 algorithm not only discovers a good information gene set to a tumor, but also improves the accuracy of tumor diagnosis or classification. Actually, in some cases, the SVM on the selected informative gene set (cluster) can even reach 100% prediction accuracy. This means that the optimal gene cluster obtained from a DSRPCL-SVM procedure on the microarray data is just the informative gene set to the tumor. Next, we repeated the above DSRPCL-SVM procedure (DSRPCL clustering and SVM checking) on the optimal or informative gene cluster. That is, we just considered the genes in the optimal cluster and neglected all the other genes. In this way, we clearly obtained a smaller optimal gene cluster or sub-cluster. If the SVM on this smaller optimal gene cluster still has a high prediction accuracy, we could consider the genes in this cluster are more powerful. In other words, they are powerful informative genes to the tumor. From the second DSRPCL-SVM procedure, we also obtained a number of sub-clusters of genes and their prediction accuracies on the three datasets (Table 3).
Table 3

Experimental results of the second DSRPCL-SVM procedure on three datasets

DatasetGene clusterNo. of genesClassfication accuracy
Linear SVMPoly SVMRBF SVM
Colon cancer11570.55560.05560.8333
2 (optimal)1450.88890.72220.8889
3370.611100.8333
4960.94440.22220.8333

Leukemia1 (optimal)9590.95450.95450.9545
27470.90910.86360.9091
310.36360.31820.3182
410.86360.77270.7727

Breast cancer17010.92590.5185
2152110.6296
340.59260.51850.5185
493110.6296
5187110.5926
682110.5185
7 (optimal)632110.8148
862110.5185
9298110.6296
From Table 3, we can find that, after the second DSRPCL-SVM procedure on each microarray dataset, a smaller informative gene cluster was obtained, while the best prediction accuracy of SVM on this gene cluster was still rather high. In this way, we could repeat the procedures and ultimately find a set of most powerful genes to the tumor. Table 4 gives the size (the number of genes) of the optimal gene cluster and the corresponding highest prediction accuracy of SVM after each round of the DSRPCL-SVM procedure.
Table 4

Experimental results of the successive DSRPCL-SVM procedures on three datasets

DatasetSubdivisionHighest accuracySize of optimal gene cluster
Colon cancer11435
20.8889145
30.777861
40.833324
50.72226
60.72226

Leukemia10.95451,708
20.9545959
30.9545479
40.9545479
50.9545271
60.9091104
70.909131
80.909131
90.86365
100.86365

Breast cancer111,580
21632
3194
40.851925
50.703711
60.703711
70.62964
80.62964
From Table 4, we can find that the division of the optimal gene cluster would stop after several DSRPCL-SVM procedures. As we successively divided the optimal gene cluster, the prediction accuracy of SVM would drop slightly. This means that some informative genes were left, while the remaining genes became more powerful. As the division of the DSRPCL-SVM procedure on the optimal gene cluster finally stopped, we obtained the most powerful informative genes to the tumor in this indivisible cluster. Obviously, these genes are keys to the diagnosis and treatment of the tumor. Table 5 is a list of the identity numbers of the powerful genes obtained from the experiments on the three datasets, respectively. Although we do not know the biological meanings of these genes, we are sure that they are critical to the corresponding tumors for the medical diagnosis and treatment.
Table 5

Identity numbers of the powerful genes for the three datasets

DatasetPowerful gene ID No.
Colon cancer211, 1215, 1394, 1621, 1858, 1865
Leukemia331, 569, 787, 2281, 4586
Breast cancer383, 385, 5294, 5797
From the further experiments, we can find that the DSRPCL-SVM procedure can always find a set of several powerful genes to a tumor or some biological phenotypes. Clearly, this result is very significant for the medical analysis and treatment. However, the experiment result is not very stable. That is, the powerful genes may be changed greatly with the different initial values of the parameters in the DSRPCL algorithm. The experiment results given above are some typical examples. We think that the reasons of the instability or sensitivity to the initial parameter values may be two-fold. First, the DSRPCL algorithm may be sensitive with the initial values of the parameters when there are just a small number of samples in the dataset. Second, the real powerful genes for a tumor my be dependent and our DSRPCL-SVM method can only find a set of powerful genes on which an SVM diagnosis system can still be made efficiently.

Conclusion

We investigated the problem of informative gene discovery and analysis from the perspective of the newly established unsupervised clustering method—DSRPCL algorithm. Since the DSRPCL algorithm can detect the number of clusters automatically in a dataset, we apply it to dividing the genes expressed through microarray data into a number of functional gene clusters and use SVM to check which cluster is the set of informative genes to the tumor. Moreover, this DSRPCL-SVM procedure can be further implemented on the informative gene cluster successively and find out the critical or powerful informative genes of the tumor. Our experiments on the colon, leukemia, and breast cancer datasets demonstrated that this DSRPCL-SVM method is really efficient for discovering the informative gene set as well as the powerful informative genes for a tumor through the microarray data.

Authors’ contributions

WX and ZC participated in the method design, performed the experimental work and drafted the manuscript. JM conceived the idea of using this approach, guided the research and completed the manuscript. All authors read and approved the final manuscript.

Competing interests

The authors have declared that no competing interests exist.
  13 in total

1.  From global expression data to gene networks.

Authors:  D Thieffry
Journal:  Bioessays       Date:  1999-11       Impact factor: 4.345

2.  Support vector machine classification and validation of cancer tissue samples using microarray expression data.

Authors:  T S Furey; N Cristianini; N Duffy; D W Bednarski; M Schummer; D Haussler
Journal:  Bioinformatics       Date:  2000-10       Impact factor: 6.937

3.  Systematic determination of genetic network architecture.

Authors:  S Tavazoie; J D Hughes; M J Campbell; R J Cho; G M Church
Journal:  Nat Genet       Date:  1999-07       Impact factor: 38.330

4.  Knowledge-based analysis of microarray gene expression data by using support vector machines.

Authors:  M P Brown; W N Grundy; D Lin; N Cristianini; C W Sugnet; T S Furey; M Ares; D Haussler
Journal:  Proc Natl Acad Sci U S A       Date:  2000-01-04       Impact factor: 11.205

Review 5.  Options available--from start to finish--for obtaining data from DNA microarrays II.

Authors:  Andrew J Holloway; Ryan K van Laar; Richard W Tothill; David D L Bowtell
Journal:  Nat Genet       Date:  2002-12       Impact factor: 38.330

6.  A cost-function approach to rival penalized competitive learning (RPCL).

Authors:  Jinwen Ma; Taijun Wang
Journal:  IEEE Trans Syst Man Cybern B Cybern       Date:  2006-08

7.  Rival penalized competitive learning for clustering analysis, RBF net, and curve detection.

Authors:  L Xu; A Krzyzak; E Oja
Journal:  IEEE Trans Neural Netw       Date:  1993

8.  Broad patterns of gene expression revealed by clustering analysis of tumor and normal colon tissues probed by oligonucleotide arrays.

Authors:  U Alon; N Barkai; D A Notterman; K Gish; S Ybarra; D Mack; A J Levine
Journal:  Proc Natl Acad Sci U S A       Date:  1999-06-08       Impact factor: 11.205

9.  Gene selection using support vector machines with non-convex penalty.

Authors:  Hao Helen Zhang; Jeongyoun Ahn; Xiaodong Lin; Cheolwoo Park
Journal:  Bioinformatics       Date:  2005-10-25       Impact factor: 6.937

10.  Cluster analysis and display of genome-wide expression patterns.

Authors:  M B Eisen; P T Spellman; P O Brown; D Botstein
Journal:  Proc Natl Acad Sci U S A       Date:  1998-12-08       Impact factor: 11.205

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.