Literature DB >> 25562040

Predicting 5-Year Survival Status of Patients with Breast Cancer based on Supervised Wavelet Method.

Maryam Farhadian1, Hossein Mahjub2, Jalal Poorolajal3, Abbas Moghimbeigi4, Muharram Mansoorizadeh5.   

Abstract

OBJECTIVES: Classification of breast cancer patients into different risk classes is very important in clinical applications. It is estimated that the advent of high-dimensional gene expression data could improve patient classification. In this study, a new method for transforming the high-dimensional gene expression data in a low-dimensional space based on wavelet transform (WT) is presented.
METHODS: The proposed method was applied to three publicly available microarray data sets. After dimensionality reduction using supervised wavelet, a predictive support vector machine (SVM) model was built upon the reduced dimensional space. In addition, the proposed method was compared with the supervised principal component analysis (PCA).
RESULTS: The performance of supervised wavelet and supervised PCA based on selected genes were better than the signature genes identified in the other studies. Furthermore, the supervised wavelet method generally performed better than the supervised PCA for predicting the 5-year survival status of patients with breast cancer based on microarray data. In addition, the proposed method had a relatively acceptable performance compared with the other studies.
CONCLUSION: The results suggest the possibility of developing a new tool using wavelets for the dimension reduction of microarray data sets in the classification framework.

Entities:  

Keywords:  breast cancer; microarray data; supervised wavelet; support vector machine

Year:  2014        PMID: 25562040      PMCID: PMC4281603          DOI: 10.1016/j.phrp.2014.09.002

Source DB:  PubMed          Journal:  Osong Public Health Res Perspect        ISSN: 2210-9099


Introduction

Metastatic breast cancer is a stage of breast cancer where the disease has spread to distant organs or tissues. Treatments against metastasis exist, but usually further treatments after surgery can have serious side effects and involve high medical costs [1]. An important task to optimize the adjuvant chemotherapy of metastasis related to breast cancer is to diagnose the risk of metastasis accurately [2-4]. Classification of cancer patients into different risk classes is very important in clinical applications. Traditional methods for patient classification were mainly based on a series of clinical and histological features [3]. It is estimated that the advent of high-dimensional gene expression data could improve patient classification [5]. Gene expression profiles of breast tumor samples could be used to predict relapse and metastatic patterns in breast cancer patients that could be potential candidate targets for new treatments [4]. It is reasonable to assume that any difference between the two tumors should be represented by some difference in gene expression. However, in microarray studies, the number of samples is relatively small compared to the number of genes per sample. Furthermore, from the biological aspect, only a small portion of genes have predicted the power for phenotypes. If all or most of the genes are considered in the predictive model, they can induce substantial noise and thereby lead to poor predictive performance [6]. Thus, in order to obtain good classification accuracy, a crucial step towards the application of microarray data is the dimensional reduction from the gene expression profiles. In recent years, both feature selection and feature extraction methods have been widely used for classifying gene expression data [7]. Bair and Tibshirani [8] and Bair et al. [9] explored the use of supervised principal component analysis (PCA), which is similar to conventional PCA except that it uses a subset of the predictors selected based on their association with the outcome. Wavelet-based methods have also been used to solve the dimension reduction problem. The primary intuition for applying wavelets in the case of gene expression is that genes are often coexpressed in groups. Therefore, it would be useful to treat the group as a single variable, akin to the motivation behind methods such as PCA [10]. One-dimensional discrete wavelet transform (DWT) is frequently used for feature extraction in the analysis of high-dimensional biomedical data [11]. Studies showed that this method has an acceptable performance in the field of feature extraction in the classification framework [11-15]. The current study aimed to introduce a dimension reduction strategy for transforming the high-dimensional gene expression data in a low-dimensional space based on wavelet transform (WT) in order to predict metastasis of breast cancer. Accordingly, a predictive support vector machine (SVM) model was built upon the reduced dimensional space. Then, the proposed novel supervised wavelet method of feature extraction was compared with the supervised PCA.

Materials and methods

The proposed method was applied to three publicly available microarray data sets related to breast cancer.

Data

Breast cancer data from van't Veer (NKI_97)

The first data set is reported by van't Veer et al [2] and referred to as NKI_97. The original van't Veer data consists of gene expression profiles and clinical information for 97 samples of primary breast cancer tumors, and each case is described by the expression levels of 24,481 genes. Fifty-one patients remained free from metastasis for at least 5 years and were metastasis-negative, and 46 cancer patients developed metastasis within 5 years and were metastasis-positive. All patients were <55 years old and were lymph node-negative. They had no tumor cells in local lymph nodes [2]. The data used in this study is a filtered version of the van't Veer data including gene expression values of 4948 genes in 97 tumor samples [2]. The data are publicly available at the “cancer data” R package (http://www.bioconductor.org/packages/release/data/experiment/html/cancerdata.html).

Breast cancer data from van de Vijver (NKI_295)

The second data set is reported by van de Vijver et al [4] and referred to as NKI_295. The data set provides the gene information for 295 primary breast cancer patients, of which 234 patients were new and the remaining 61 patients were involved in the first data set. Of the total 295 patients, 194 patients were metastasis-negative and 101 patients were metastasis-positive. Of the 234 new patients, 164 patients were metastasis-negative and 70 patients were metastasis-positive. Of the 61 patients involved in the first data set, 30 were metastasis-negative and 31 patients were metastasis-positive. The data is a filtered version of the van de Vijver data including gene expression values of 4948 genes in 295 tumor samples [4]. The data are publicly available at the “cancer data” R package.

Breast cancer data from the Wang study (VDX_286)

The last data set, reported by Wang et al [16] and referred to as VDX_286, contains 286 lymph node-negative breast cancer patients who had not received any adjuvant systemic treatment [16]. Among them, 106 patients had distant metastasis within 5 years of follow up and were considered as metastatic patients, while the rest were considered as nonmetastatic patients. A set of 22,283 genes is available for this data set. The data are publicly available at the “breast cancer VDX” R package.

Wavelet Transform

A wavelet is a “small wave”, which has its energy concentrated in time. In signal processing, a transformation technique is used to transfer data in another domain where hidden information can be extracted. Wavelets have a nice feature of local description and separation of signal characteristics, and provide a tool for the analysis of transient or time-varying signals [11]. A wavelet is a set of orthonormal basis functions generated from dilation and translation of a single scaling function or father wavelet () and a mother wavelet (). WTs are classified into two different categories: the continuous WT and the DWT. The DWT is a linear operation that operates on a data vector, transforming it into a wavelet coefficient. The idea underlying DWT is to express any function in terms of ö (t) and ø (t) as follows:where , , and represent the scaling function, mother wavelet function, scaling coefficients (approximation coefficients) at scale zero, and detail coefficients at scale j, respectively. The variable k is the translation coefficient for the localization of gene expression data. The scales denote the different (low to high) scale bands. The variable symbol is the scale (level) number selected [10]. One-dimensional DWT decomposes a signal as a sum of wavelets at different time shifts and scales (frequencies) using DWT. For this purpose, the signal is passed through a series of high-pass and low-pass filters in order to analyze low as well as high frequencies in the signal as follows:where and are the low-pass filters and high-pass filters, respectively. At each level, the high-pass filter produces detail coefficients (wavelet coefficients) d1, while the low-pass filter associated with the scaling function produces approximation coefficients (scaling coefficients) c1. Subsequently, the approximation coefficients c1 are split into two parts by using the same algorithm and are replaced by c2 and d2, and so on. This decomposition process is repeated until the required level is reached. The coefficient vectors are produced by down sampling and are only half the length of the signal or the coefficient vector at the previous level [12]. The main advantage of the WT is that each basis function is localized jointly in both the time and frequency domains. From a viewpoint of time-frequency, the approximation coefficients correspond to the larger-scale low-frequency components, and the detail coefficients correspond to the small-scale high-frequency components. Generally, the former can be used to approximate the original signal, and the latter represents some local details of the original signal [14,15]. There are different families of wavelets: symlets, coiflets, Daubechies, and biorthogonal wavelets. They vary in the various basic properties of wavelets, such as compactness. Haar wavelets, belonging to Daubechies wavelet family, are the most commonly used wavelets in database literature because they are easy to comprehend and fast to be computed.

Q-value

It is usual to simultaneously test many hundreds or thousands of genes in microarray studies to determine which are differentially expressed. Each of these tests will produce a p value. One main challenge in those studies is to find suitable multiple testing procedures that provide an accurate control of the error rates. Whereas the p value is a measure of significance in terms of the false positive rate, the q value is an approach used to measure statistical significance based on the concept of the false discovery rate. Similar to the p value, the q value gives each feature its own individual measure of significance [17].

Supervised WT

Firstly, any patients who remain free from metastasis for at least 5 years are placed into Class 1, otherwise into Class 2. The proposed DWT-based feature selection method consists of the following steps: (1) A t test is taken as the measure to identify differently expressed genes and a list of q values is derived. All the genes are ranked according to their corresponding q value and the required numbers of genes are selected from the list; and (2) in each step the top number of genes based on the q value are picked out. Then, this reduced set of genes is modeled by the one-dimensional DWT using Haar mother wavelet and finally, the wavelet approximation coefficients in the first and second levels of decomposition are used in the SVM model, respectively.

Supervised PCA

Bair and Tibshirani [8] and Bair et al [9] proposed supervised principal components regression. This procedure first picks out a subset of the gene expressions that correlates with response by using univariate selection, and then applies PCA to this subset. In our analysis, we pick out the top number of genes based on q values. We then apply PCA to this subset of genes, and in each step include the top numbers of principal components into a SVM model. The top numbers of principal components that will be comprised of at least 75% of the total variance are included in the SVM model.

SVM

The SVM model proposed by Vapnik [18] is a supervised learning method that is widely used in microarray data classification. Unlike many modeling techniques which aim to minimize the objective function (such as mean square error) for all instances, SVM attempts to find the hyperplanes that produce the largest separation between the decision function values for the instances located on the borderline between the two classes. The optimally identified hyperplane in the feature space corresponds to a nonlinear decision boundary in the input space. The SVM takes a set of input data with corresponding class labels and predicts a new input which belongs to the classes. In the binary classification mode, given a training set of instance-label pairs () where and SVM can be regarded as the solution of the following quadratic optimization problem:where the training data are mapped to a higher dimensional space by the function and C is a user-defined penalty parameter on the training error that controls the trade-off between classification errors and the complexity of the model. By solving the optimization problem (1) by finding the parameters w and b for a given training set, a decision hyperplane over an n-dimensional input space that produces the maximal margin in the space is designed. Thus, the decision function can be formulated as follows: SVM can derive the optimal hyperplane for nonlinearly separated data by mapping the impute data into the n-dimensional space using kernel function []. There are four basic kernels: linear, polynomial, radial basic function, and sigmoid [18,19]. In this study, the goal of SVM modeling was to classify patients who had a high risk of breast cancer recurrence. The predictive performance of the SVM-classifier was reported based on sensitivity, specificity, accuracy, and the area under the receiver operating characteristic curve (AUC). These criteria are defined as follows: (TP = true positive; TN = true negative; FN = false negative; and FP = false positive): Accuracy: ACC = Sensitivity: SN = Specificity: SP = The method is implemented using MATLAB r2012a software (MATLAB Release 2012a, the MathWorks, Inc., Natick, Massachusetts, United States) and R statistical package (e1071, q value).

Cross data set comparison

To avoid over fitting and to provide a realistic evaluation, the cross data method was used. In this method, features obtained from one data set were used to construct classifiers for the other data set. In this regard, common patients in the NKI_295 and NKI_97 data were removed and the remaining data (NKI_234) were used as a test data set. This method was implemented using genes selected from NKI_234 breast cancer data as input in the supervised wavelet method in the NKI_61 data.

Results

The t test statistics were used to identify discriminative genes in each data set. After selecting the top ranked genes based on q values, one-dimensional WT in the first and second levels was applied to these preselected genes. SVMs with three types of kernels—linear, sigmoid, and radial, were used based on wavelet approximation coefficients in the first and the second levels of decomposition. For further assessment of the reported subsets of 70 genes selected by van't Veer et al [2] (for NKI_97 and NKI_295) and 76 signature genes selected by Wang et al [16] (for VDX_286), the supervised wavelet method and supervised PCA were applied. The predictive performance of SVM models was tested by cross-validation, consisting of 10 times 10-folding experiments. The results of supervised wavelet and supervised PCA for the three data sets are shown in Tables 1–3, respectively. In the NKI_97 data set, the results showed that the SVM with radial kernels based on wavelet approximation coefficients in the first level extracted from 58 preselected genes had the best performance in terms of the evaluation criteria with regard to accuracy (83.11) as well as AUC (83.45). In addition, the SVM with radial kernel based on the first supervised PCA computed based on 84 preselected genes had the best performance in terms of accuracy (79.22) as well as specificity (83.25), sensitivity (75.22), and AUC (79.24). In both methods (supervised wavelet and supervised PCA), the classifier performance based on the 70 genes selected by q values was better than the 70 gene signature from the van't Veer study (Table 1).
Table 1

Results for supervised wavelet and supervised principal component analysis (PCA): NKI_97, 10 times 10-fold cross-validation.

MethodNo. of preselected genes.MethodAccuracySensitivitySpecificityAUC
SVM (linear)70 genes (van't Veer)Wavelet (Db1.1)77.1178.3076.1577.22
Wavelet (Db1.2)69.1164.4773.0068.74
Supervised PCA73.7775.7271.8473.78
SVM (radial)70 genes (van't Veer)Wavelet (Db1.1)77.5582.2873.2477.76
Wavelet (Db1.2)75.6682.2069.7675.98
Supervised PCA71.7771.2572.2171.73
SVM (sigmoid)70 genes (van't Veer)Wavelet (Db1.1)78.8878.5779.1878.87
Wavelet (Db1.2)71.8874.8269.2672.04
Supervised PCA68.7767.5869.7368.66
SVM (linear)70 genesWavelet (Db1.1)72.3367.5576.3871.97
Wavelet (Db1.2)76.4475.5377.2476.38
Supervised PCA74.0072.5175.3173.91
SVM (radial)70 genesWavelet (Db1.1)82.7790.1474.4682.30
Wavelet (Db1.2)82.0088.4776.2182.34
Supervised PCA75.8875.2276.5275.87
SVM (sigmoid)70 genesWavelet (Db1.1)77.4486.7468.9377.84
Wavelet (Db1.2)77.0082.8671.7277.29
Supervised PCA78.2276.8379.4578.14
SVM (linear)q < 0.02 (84 genes)Wavelet (Db1.1)71.0068.4073.0970.75
Wavelet (Db1.2)72.8872.0973.6772.88
Supervised PCA78.0078.0177.9878.00
SVM (radial)q < 0.02 (84 genes)Wavelet (Db1.1)82.5587.5578.2182.88
Wavelet (Db1.2)81.6684.4779.0081.73
Supervised PCA79.2283.2575.2279.24
SVM (sigmoid)q < 0.02 (84 genes)Wavelet (Db1.1)79.8888.1772.5380.35
Wavelet (Db1.2)78.8886.6270.9478.78
Supervised PCA75.5580.0071.4875.74
SVM (linear)q < 0.01 (58 genes)Wavelet (Db1.1)73.7776.6271.3473.98
Wavelet (Db1.2)70.8867.7873.9570.86
Supervised PCA76.6679.3674.0776.71
SVM (radial)q < 0.01 (58 genes)Wavelet (Db1.1)83.1188.2778.6383.45
Wavelet (Db1.2)82.3385.1179.5582.33
Supervised PCA77.3382.4372.7277.58
SVM (sigmoid)q < 0.01 (58 genes)Wavelet (Db1.1)80.6689.6972.5181.10
Wavelet (Db1.2)80.7785.7776.0780.92
Supervised PCA76.0080.8771.8676.37

AUC = area under the receiver operating characteristic curve; SVM = support vector machine.

In the NKI_295 data set (Table 2), the results showed that the SVM with radial kernels based on wavelet approximation coefficients in the first level extracted from 91 preselected genes had the best performance in terms of the evaluation criteria, with the highest accuracy (75.37) as well as AUC (70.03). In addition, the SVM with linear kernel based on the first supervised PCA computed based on 91 preselected genes had the best performance in terms of accuracy (73.03) as well as AUC (66.63). In both methods (supervised wavelet and supervised PCA), the classifier performance based on the 70 genes selected by q values was better than the 70 gene signature from the van't Veer study.
Table 2

Results for supervised wavelet and supervised principal component analysis (PCA): NKI_295, 10 times 10-fold cross-validation.

MethodNo. of preselected genesMethodAccuracySensitivitySpecificityAUC
SVM (linear)70 genes (van't Veer)Wavelet (Db1.1)65.1038.3277.8258.07
Wavelet (Db1.2)66.1329.7184.3357.02
Supervised PCA67.0028.5587.3857.97
SVM (radial)70 genes (van't Veer)Wavelet (Db1.1)70.9632.8290.3761.59
Wavelet (Db1.2)67.9626.3788.6457.50
Supervised PCA65.7218.3691.1454.75
SVM (sigmoid)70 genes (van't Veer)Wavelet (Db1.1)63.1724.7081.8253.26
Wavelet (Db1.2)64.5519.2588.1053.67
Supervised PCA66.2723.7389.0456.39
SVM (linear)70 genesWavelet (Db1.1)70.2048.6881.2964.98
Wavelet (Db1.2)72.6553.0882.5267.80
Supervised PCA69.3745.8381.7163.77
SVM (radial)70 genesWavelet (Db1.1)71.1336.9888.7662.87
Wavelet (Db1.2)70.0639.9286.2263.07
Supervised PCA70.1034.4189.3761.89
SVM (sigmoid)70 genesWavelet (Db1.1)65.7943.0377.0860.06
Wavelet (Db1.2)63.4444.5073.7259.11
Supervised PCA68.8633.9287.5560.74
SVM (linear)q < 0.001 (56 genes)Wavelet (Db1.1)69.6848.6580.8764.76
Wavelet (Db1.2)67.2041.1280.8760.99
Supervised PCA71.6846.8184.5665.68
SVM (radial)q < 0.001 (56 genes)Wavelet (Db1.1)70.3733.9089.4061.65
Wavelet (Db1.2)65.7228.3086.4857.39
Supervised PCA70.8240.5486.6263.58
SVM (sigmoid)q < 0.001 (56 genes)Wavelet (Db1.1)65.7944.6876.5360.60
Wavelet (Db1.2)66.3741.3879.4960.43
Supervised PCA71.1045.4684.2164.83
SVM (linear)q < 0.002 (91 genes)Wavelet (Db1.1)72.3746.5086.0066.25
Wavelet (Db1.2)70.4380.9757.0067.24
Supervised PCA73.0346.5186.7666.63
SVM (radial)q < 0.002 (91 genes)Wavelet (Db1.1)75.3752.8587.2170.03
Wavelet (Db1.2)74.5849.1886.4867.83
Supervised PCA71.0639.5688.0563.81
SVM (sigmoid)q < 0.002 (91 genes)Wavelet (Db1.1)72.4442.3688.0165.19
Wavelet (Db1.2)74.3447.2188.3867.80
Supervised PCA69.1049.4778.6364.05

AUC = area under the receiver operating characteristic curve; SVM = support vector machine.

In the VDX_286 data set (Table 3), the results showed that the SVM with linear kernels based on wavelet approximation coefficients in the second level extracted from 67 preselected genes had the best performance with the highest accuracy (79.21) as well as AUC (76.04). In addition, the SVM with linear kernel based on the first supervised PCA computed based on 67 preselected genes had the best performance in terms of accuracy (76.00) as well as AUC (74.71). In both methods (supervised wavelet and supervised PCA), the classifier performance based on the selected 76 genes using t statistics was better than the 76 gene signature identified in the Wang study.
Table 3

Results for supervised wavelet and supervised principal component analysis (PCA): VDX_286, 10 times 10-fold cross-validation.

MethodNo. of preselected genesMethodAccuracySensitivitySpecificityAUC
SVM (linear)76 genes (Wang)Wavelet (Db1.1)64.4244.4276.2560.33
Wavelet (Db1.2)66.3944.8679.1361.99
Supervised PCA68.1739.1385.8262.47
SVM (radial)76 genes (Wang)Wavelet (Db1.1)63.8935.7479.7757.75
Wavelet (Db1.2)65.1028.9787.4558.21
Supervised PCA67.8233.9787.8860.92
SVM (sigmoid)76 genes (Wang)Wavelet (Db1.1)66.9245.4979.6662.58
Wavelet (Db1.2)65.6443.4279.1161.27
Supervised PCA67.3943.5481.2862.41
SVM (linear)76 genesWavelet (Db1.1)75.1761.9783.0272.50
Wavelet (Db1.2)76.3559.9485.9972.96
Supervised PCA67.9642.0483.6562.85
SVM (radial)76 genesWavelet (Db1.1)76.0760.8084.8672.83
Wavelet (Db1.2)77.2556.4889.2372.86
Supervised PCA67.3237.1785.3761.27
SVM (sigmoid)76 genesWavelet (Db1.1)77.2162.4186.1074.26
Wavelet (Db1.2)71.5761.7977.3469.56
Supervised PCA68.1042.8582.7762.81
SVM (linear)q < 0.04 (67 genes)Wavelet (Db1.1)78.2167.0584.6075.83
Wavelet (Db1.2)79.2164.4687.6176.04
Supervised PCA76.0068.7680.6674.71
SVM (radial)q < 0.04 (67 genes)Wavelet (Db1.1)77.0058.6587.5673.10
Wavelet (Db1.2)75.1754.4188.3371.37
Supervised PCA75.0060.9783.6872.33
SVM (sigmoid)q < 0.04 (67 genes)Wavelet (Db1.1)77.0365.7583.5474.65
Wavelet (Db1.2)78.5066.7985.5976.19
Supervised PCA75.2164.9681.6373.30
SVM (linear)q < 0.05 (86 genes)Wavelet (Db1.1)77.0067.0483.0275.03
Wavelet (Db1.2)78.1765.5785.6275.60
Supervised PCA75.9666.1482.1174.12
SVM (radial)q < 0.05 (86 genes)Wavelet (Db1.1)75.9655.1588.2071.68
Wavelet (Db1.2)76.1753.5789.4571.51
Supervised PCA75.5763.5082.9873.24
SVM (sigmoid)q < 0.05 (86 genes)Wavelet (Db1.1)77.3266.1883.9175.04
Wavelet (Db1.2)74.6759.4083.3671.38
Supervised PCA74.2865.6179.1972.40

AUC = area under the receiver operating characteristic curve; SVM = support vector machine.

To evaluate the reproducibility of the proposed method, a cross data-set comparison was also performed. As shown in Table 4, the results confirmed that the supervised wavelet method also had an acceptable performance, although the improvements were not as high as in the inner data set comparison. The results of other studies based on the same data sets are shown in Table 5. It can be seen that the proposed method had a higher capability for the prediction of metastasis than the other studies [20-29].
Table 4

External validation for supervised wavelet: NKI_234_61, 10 times 10-fold cross-validation.

MethodNo. of preselected genesWaveletAccuracySensitivitySpecificityAUC
SVM (linear)70 genesDb1. Level 167.8375.6359.1567.39
Db1. Level 264.3369.4558.8264.13
SVM (radial)70 genesDb1. Level 164.5072.4754.9463.71
Db1. Level 267.6667.9467.3667.65
SVM (sigmoid)70 genesDb1. Level 165.6672.9358.2465.59
Db1. Level 262.1656.0668.4762.27
SVM (linear)q < 0.00 (13 genes)Db1. Level 164.0068.8159.3464.07
Db1. Level 261.5053.9669.8261.89
SVM (radial)q < 0.003 (13 genes)Db1. Level 171.8378.3365.3371.83
Db1. Level 269.0070.1667.8669.01
SVM (sigmoid)q < 0.003 (13 genes)Db1. Level 170.6665.0676.7370.90
Db1. Level 268.8367.8969.7668.83

AUC = area under the receiver operating characteristic curve; SVM = support vector machine.

Table 5

Previously published analyses for the breast cancer data.

No. of samplesFeature selectionClassifierMeasureValidation method
Current study97Supervised waveletSVM radial kernelAccuracy: 83.11CV
Supervised PCASVM radial kernelAccuracy: 79.22
295Supervised waveletSVM radial kernelAccuracy: 75.37
Supervised PCASVM linear kernelAccuracy: 73.03
286Supervised waveletSVM linear kernelAccuracy: 79.21
Supervised PCASVM linear kernelAccuracy: 76.00
Michiels et al (2005) [20]97CorrelationNearest-centroidAccuracy: 68.00CV
Peng (2005) [23]97Signal to noise ratioSVMAccuracy: 75.00Leave-one-out CV
Signal to noise ratioBagg & Boost SVMAccuracy: 77.00
SubsamplingEnsemble SVMAccuracy: 81.00
Pochet et al (2004) [24]78+19*NoneLS-SVM linear kernelAccuracy: 69.00Leave-one-out CV
NoneSVM RBF kernelAccuracy: 69.00
NoneSVM linear kernelAccuracy: 52.00
Alexe et al (2006) [22]78+19Support set identified by logical analysis of dataSVM linear kernelAccuracy: 77.00CV
Artificial NNAccuracy: 79.00
Logistic regressionAccuracy: 78.00
Nearest neighborsAccuracy: 76.00
Decision trees (C4.5)Accuracy: 67.00
Jahid et al (2012) [26]295Steiner tree based methodSVMAccuracy: 62.00CV
286Accuracy: 61.00
Chuang et al (2007) [25]295Subnetwork markerSVMAccuracy: 72.00CV
286Accuracy: 62.00
van Vliet et al (2012) [21]295Filtering approach (t test)Nearest mean classifierAUC: 73.80CV
Dehnavi et al (2013) [27]286Rough-set theoryNeuro-fuzzy SystemAccuracy: 78.0010-fold CV
Lee et al (2011) [28]286Modules with condition responsive correlationsNaïve Bayesian classifierAUC: 0.62Leave-one-out CV
Jahid et al (2014) [29]295Patient–patient co-expression networksPC-classifierAUC: 0.78Leave-one-out CV
DaggingAUC: 0.72
AdaBoostAUC: 0.66
286PC-classifierAUC: 0.68
DaggingAUC: 0.61
AdaBoostAUC: 0.55

AUC = area under the receiver operating characteristic curve; CV = cross validation; PCA = principal component analysis; RBF = radial basic function; SVM = support vector machine.

Discussion

This study proposed a new method based on WT to develop a novel predictive model for the prediction of breast cancer metastasis. Furthermore, the performance of this method was compared with supervised PCA. The main purpose of the feature extraction method using WT is that the approximation coefficients usually comprise the majority of the important information [11]. In addition, the powerful capability of the DWT to compress the signal energy makes it a good candidate for feature extraction applications. The DWT compresses most of the energy from the input signal and concentrates it in a few high-magnitude coefficients in the transformed matrix. The wavelet feature extraction method does not depend on the training data set to obtain the basis of feature space compared to the PCA method. Therefore, the wavelet feature extraction method dramatically reduces the computation load compared to PCA [11,12]. Considering the fact that most genes are irrelevant to patients' metastasis, we analyzed the reduced data set given by selecting genes that were significantly related to metastasis based on the t test statistics. If the WT is performed directly by using all of the genes in a data set, there is no guarantee that the resulting wavelet coefficients will be related to metastasis. Thus, this study introduced a supervised form of WT that can be considered as a supervised wavelet. After extracting supervised wavelet approximation coefficients using discrete Haar WT, these coefficients had higher predictive performances than the first three principal components. Therefore, our results suggested that the wavelet coefficients are the efficient way to characterize the features of high-dimensional microarray data. Because the performance of the proposed supervised wavelet method is likely to be improvable compared to some other studies, we conclude that this method is worth further investigation as a tool for cancer patient classification based on gene expression data. For example, to achieve optimal classification performance, a suitable combination of the classifier and the gene selection method needs to be specifically selected for a given data set. Some studies reported misclassification rates that were obtained by the application of their classifier to a one splitting of the test and training set. For example, van't Veer et al [2] developed a 70-gene classifier predicting a distant metastasis of breast cancer. In the training set, the classifier predicted the class of 65/78 cases correctly (i.e., with an accuracy of 83.3%, corresponding to a weighted accuracy of 83.6%), whereas in the test set it predicted the class of 17/19 cases correctly (i.e., with an accuracy of 89.5%, corresponding to a weighted accuracy of 88.7%). However, in the present study, in order to avoid the over fitting problem, we followed the 10 times 10-fold cross-validation for evaluating the SVM classifier. The evaluation of the classifier based on one test set is very impressed with the data splitting process. Future investigations can focus on different ways of preselecting genes in the first stage of the proposed method. For example, rather than ranking genes based on their t test scores, one would use a different metric to measure the association between a given gene and metastasis occurrence. By contrast, another mother wavelet and a different level of decomposition can be studied. In this study, gene expression data were employed as predictors. However, prediction performance may be improved by adding other covariates such as age, lymph node status, tumor size, and histological grade. It is likely that the classification performances could be improved with the use of some other classifiers. This study confirmed that the SVM model based on the supervised wavelet feature extraction method was superior with regards to predictive performance than the supervised PCA and some other studies. Gene expression profiling can help to distinguish between patients at high risk and those at low risk for developing distant metastases, therefore, this technology and other high-throughput techniques are helping to alter our view of breast cancer and provide us with new tools for molecular diagnoses. These results exhibit the possibility of developing a new tool using wavelets for the dimension reduction of microarray data sets in the classification framework and therefore, the use of this method in similar classification problems is recommended.

Conflicts of interest

The authors have no conflicts of interest to declare.
  19 in total

1.  Statistical significance for genomewide studies.

Authors:  John D Storey; Robert Tibshirani
Journal:  Proc Natl Acad Sci U S A       Date:  2003-07-25       Impact factor: 11.205

2.  Predicting survival from microarray data--a comparative study.

Authors:  H M Bøvelstad; S Nygård; H L Størvold; M Aldrin; Ø Borgan; A Frigessi; O C Lingjaerde
Journal:  Bioinformatics       Date:  2007-06-06       Impact factor: 6.937

3.  Feature extraction and dimensionality reduction for mass spectrometry data.

Authors:  Yihui Liu
Journal:  Comput Biol Med       Date:  2009-07-30       Impact factor: 4.589

4.  Predicting disease phenotypes based on the molecular networks with condition-responsive correlation.

Authors:  Sejoon Lee; Eunjung Lee; Kwang H Lee; Doheon Lee
Journal:  Int J Data Min Bioinform       Date:  2011       Impact factor: 0.667

5.  Gene expression profiling predicts clinical outcome of breast cancer.

Authors:  Laura J van 't Veer; Hongyue Dai; Marc J van de Vijver; Yudong D He; Augustinus A M Hart; Mao Mao; Hans L Peterse; Karin van der Kooy; Matthew J Marton; Anke T Witteveen; George J Schreiber; Ron M Kerkhoven; Chris Roberts; Peter S Linsley; René Bernards; Stephen H Friend
Journal:  Nature       Date:  2002-01-31       Impact factor: 49.962

6.  Dimension reduction methods for microarrays with application to censored survival data.

Authors:  Lexin Li; Hongzhe Li
Journal:  Bioinformatics       Date:  2004-07-15       Impact factor: 6.937

7.  A Steiner tree-based method for biomarker discovery and classification in breast cancer metastasis.

Authors:  Md Jamiul Jahid; Jianhua Ruan
Journal:  BMC Genomics       Date:  2012-10-26       Impact factor: 3.969

8.  Integration of clinical and gene expression data has a synergetic effect on predicting breast cancer outcome.

Authors:  Martin H van Vliet; Hugo M Horlings; Marc J van de Vijver; Marcel J T Reinders; Lodewyk F A Wessels
Journal:  PLoS One       Date:  2012-07-11       Impact factor: 3.240

9.  Semi-supervised methods to predict patient survival from gene expression data.

Authors:  Eric Bair; Robert Tibshirani
Journal:  PLoS Biol       Date:  2004-04-13       Impact factor: 8.029

10.  Network-based classification of breast cancer metastasis.

Authors:  Han-Yu Chuang; Eunjung Lee; Yu-Tsueng Liu; Doheon Lee; Trey Ideker
Journal:  Mol Syst Biol       Date:  2007-10-16       Impact factor: 11.429

View more
  2 in total

1.  Diagnosing thyroid disorders: Comparison of logistic regression and neural network models.

Authors:  Shiva Borzouei; Hossein Mahjub; Negar Asaad Sajadi; Maryam Farhadian
Journal:  J Family Med Prim Care       Date:  2020-03-26

2.  Application of Censored Quantile Regression to Determine Overall Survival Related Factors in Breast Cancer.

Authors:  Javad Faradmal; Ghodratollah Roshanaei; Maryam Mafi; Abdolazim Sadighi-Pashaki; Manoochehr Karami
Journal:  J Res Health Sci       Date:  2016
  2 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.