| Literature DB >> 27433543 |
Mukesh Kumar1, Santanu Kumar Rath1.
Abstract
The DNA microarray classification technique has gained more popularity in both research and practice. In real data analysis, such as microarray data, the dataset contains a huge number of insignificant and irrelevant features that tend to lose useful information. Classes with high relevance and feature sets with high significance are generally referred for the selected features, which determine the samples classification into their respective classes. In this paper, kernel fuzzy inference system (K-FIS) algorithm is applied to classify the microarray data (leukemia) using t-test as a feature selection method. Kernel functions are used to map original data points into a higher-dimensional (possibly infinite-dimensional) feature space defined by a (usually nonlinear) function ϕ through a mathematical process called the kernel trick. This paper also presents a comparative study for classification using K-FIS along with support vector machine (SVM) for different set of features (genes). Performance parameters available in the literature such as precision, recall, specificity, F-measure, ROC curve, and accuracy are considered to analyze the efficiency of the classification model. From the proposed approach, it is apparent that K-FIS model obtains similar results when compared with SVM model. This is an indication that the proposed approach relies on kernel function.Entities:
Year: 2014 PMID: 27433543 PMCID: PMC4897118 DOI: 10.1155/2014/769159
Source DB: PubMed Journal: Int Sch Res Notices ISSN: 2356-7872
Relevant works on cancer classification using microarray (leukemia) dataset.
| Author | Feature selection/extraction method | Classifier used | Accuracy (%) |
|---|---|---|---|
| Cho et al. [ | Kernel fisher feature discriminant analysis (KFDA) | 73.53 | |
|
| |||
|
Deb and Raji Reddy [ | NSGA-II | 100 | |
|
| |||
| Lee et al. [ | Bayesian model | Artificial neural network (ANN), KNN, and SVM | 97.05 |
|
| |||
| Ye et al. [ | Uncorrelated linear discriminant analysis (ULDA) | KNN ( | 97.5 |
|
| |||
| Cho et al. [ | SVM-RFE | Kernel KFDA | 94.12 |
|
| |||
|
Paul and Iba [ | Probabilistic model building genetic algorithm (PMBGA) | Naive-Bayes (NB), weighted voting classifier | 90 |
|
| |||
|
D | Random forest | 95 | |
|
| |||
| Peng et al. [ | Fisher ratio | NB, decision tree J4.8, and SVM | 100, 95.83, and 98.6 |
|
| |||
| Pang et al. [ | Bootstrapping consistency gene selection | KNN | 94.1 |
|
| |||
| Hernandez et al. [ | Genetic algorithm (GA) | SVM | 91.5 |
|
| |||
| Zhang and Deng [ | Based Bayes error filter (BBF) | Support vector machine (SVM), | 100, 98.61 |
|
| |||
| Bharathi and Natarajan [ | ANOVA | SVM | 97.91 |
|
| |||
|
Tang et al. [ | ANOVA | Discriminant Kernel partial least square (Kernel-PLS) | 100 |
|
| |||
|
Mundra and Rajapakse [ |
| SVM | 96.88, 98.12, 97.88, and 98.41 |
|
| |||
|
Lee and Leu [ |
| Hybrid with GA + KNN and SVM | 100 |
|
| |||
|
Salem et al. [ | Multiple scoring gene selection technique (MGS-CM) | SVM, KNN, and linear discriminant analysis (LDA) | 90.97 |
Figure 1Proposed work for microarray classification.
Performance parameters.
| Performance parameters | Description |
|---|---|
| Precision = TP/(FP + TP) | It is the degree to which the repeated measurements under unchanged conditions show the same results |
| Recall = TP/(FN + TP) | It indicates that the number of the relevant items are to be identified |
|
| It combines the “precision” and “recall” numeric values to give a single score, which is defined as the harmonic mean of the precision and recall |
| Specificity = TN/(FP + TN) | It focuses on how effectively a classifier identifies negative labels |
| Accuracy = (TP + TN)/(FP + FN + TP + TN) | It measures the percentage of inputs in the test set that the classifier correctly labeled |
| Receive operating characteristic (ROC) curve | ROC curve is a graphical plot which illustrates that the performance of a binary classifier system as its discrimination threshold is varied. It investigates and employs the relationship between “true positive rate (sensitivity)” and “false positive rate (1 − specificity)” of a classifier |
Classification matrix.
| NO | YES | |
|---|---|---|
| NO | True Negative (TN) | False Positive (FP) |
| YES | False Negative (FN) | True Positive (TP) |
Figure 2Empirical cumulative distribution function (CDF) of the P values.
Figure 3Framework of kernel fuzzy inference system (K-FIS).
Algorithm 1Kernel subtractive clustering.
Classification matrix before classification.
| ALL(0) | AML(1) | |
|---|---|---|
| ALL(0) | 47 | 0 |
| AML(1) | 25 | 0 |
Selected features with “P value” in descending order.
| Number of features | Notation | Selected features with gene ID. |
|---|---|---|
| 5 | F5 | { |
| 10 | F10 | F5 ∪{ |
| 15 | F15 | F10 ∪{ |
| 20 | F20 | F15 ∪{ |
| 25 | F25 | F20 ∪{ |
| 30 | F30 | F25 ∪{ |
Parameters of K-FIS model.
| Parameters used | Range | Value used |
|---|---|---|
| Squash factor ( | [1, 2] | 1.25 |
| Accept ratio ( | (0, 1] | 0.75 |
| Reject ratio ( | (0, 1] | 0.15 |
| Cluster radius ( | (0, 1] | — |
Algorithm 2F-fold cross-validation.
Figure 9Training accuracy in each fold with different set of features using K-FIS with linear kernel.
Figure 10Testing accuracy in each fold with different set of features using K-FIS with linear kernel.
Figure 4ROC curve for FIS with different set of features.
Performance analysis of FIS with different set of features with best suitable cluster radius (r in small bracket).
| Models ( | Accuracy | Precision | Recall | Specificity |
|
|---|---|---|---|---|---|
| F5 (0.5) | 0.9306 | 0.9200 | 0.8846 | 0.9565 | 0.9020 |
| F10 (0.2) | 0.9722 | 0.9600 | 0.9600 | 0.9787 | 0.9600 |
| F15 (0.4) | 0.9167 | 0.9200 | 0.8519 | 0.9556 | 0.8846 |
| F20 (0.45) | 0.9444 | 1.0000 | 0.8621 | 1.0000 | 0.9259 |
| F25 (0.2) | 0.9583 | 0.9600 | 0.9231 | 0.9783 | 0.9412 |
| F30 (0.4) | 0.9583 | 0.9600 | 0.9231 | 0.9783 | 0.9412 |
Figure 11Training accuracy in each fold with different set of features using K-FIS with polynomial kernel.
Figure 12Testing accuracy in each fold with different set of features using K-FIS with polynomial kernel.
Figure 5ROC curve for K-FIS using polynomial kernel (γ = 1, c = 0.5, and d = 3) with various feature sets.
Performance analysis of K-FIS using polynomial kernel (γ = 1, c = 0.5, d = 3) with different set of features with best suitable cluster radius (r in small bracket).
| Models ( | Accuracy | Precision | Recall | Specificity |
|
|---|---|---|---|---|---|
| F5 (0.2) | 0.9861 | 1.0000 | 0.9615 | 1.0000 | 0.9804 |
| F10 (0.2) | 0.9722 | 0.9600 | 0.9600 | 0.9787 | 0.9600 |
| F15 (0.3) | 0.9583 | 0.9600 | 0.9231 | 0.9783 | 0.9412 |
| F20 (0.2) | 0.9583 | 0.9600 | 0.9231 | 0.9783 | 0.9412 |
| F25 (0.2) | 0.9583 | 0.9600 | 0.9231 | 0.9783 | 0.9412 |
| F30 (0.4) | 0.9583 | 0.9600 | 0.9231 | 0.9783 | 0.9412 |
Figure 13Training accuracy in each fold with different set of features using K-FIS with RBF kernel.
Figure 14Testing accuracy in each fold with different set of features using K-FIS with RBF kernel.
Figure 6ROC curve for K-FIS using RBF kernel (γ = 0.5) with various feature sets.
Performance analysis of K-FIS using RBF kernel (γ = 0.5) with different set of features with best suitable cluster radius (r in small bracket).
| Models | Accuracy | Precision | Recall | Specificity |
|
|---|---|---|---|---|---|
| F5 (0.4) | 0.9722 | 1.0000 | 0.9259 | 1.0000 | 0.9615 |
| F10 (0.2) | 0.9583 | 0.9600 | 0.9231 | 0.9783 | 0.9412 |
| F15 (0.3) | 0.9722 | 1.0000 | 0.9259 | 1.0000 | 0.9615 |
| F20 (0.4) | 0.9306 | 1.0000 | 0.8333 | 1.0000 | 0.9091 |
| F25 (0.6) | 0.9306 | 1.0000 | 0.8333 | 1.0000 | 0.9091 |
| F30 (0.6) | 0.8750 | 0.9200 | 0.7667 | 0.9524 | 0.8364 |
Figure 15Training accuracy in each fold with different set of features using K-FIS with tansig kernel.
Figure 16Testing accuracy in each fold with different set of features using K-FIS with tansig kernel.
Figure 7ROC curve for K-FIS using tansig kernel (γ = 0.5, c = 0.1) with various feature sets.
Performance analysis of K-FIS using tansig kernel (γ = 0.5, c = 0.1) with different set of features with best suitable cluster radius (r in small bracket).
| Models ( | Accuracy | Precision | Recall | Specificity |
|
|---|---|---|---|---|---|
| F5 (0.2) | 0.9861 | 1.0000 | 0.9615 | 1.0000 | 0.9804 |
| F10 (0.2) | 0.9722 | 0.9600 | 0.9600 | 0.9787 | 0.9600 |
| F15 (0.2) | 0.9583 | 0.9600 | 0.9231 | 0.9783 | 0.9412 |
| F20 (0.2) | 0.9444 | 0.9200 | 0.9200 | 0.9575 | 0.9200 |
| F25 (0.2) | 0.9683 | 0.9600 | 0.9231 | 0.9783 | 0.9412 |
| F30 (0.2) | 0.9683 | 0.9600 | 0.9231 | 0.9783 | 0.9412 |
Average training, average testing accuracy, and CPU time (in seconds) with different models.
| Models∖number of Features | F5 | F10 | F15 | F20 | F25 | F30 | ||||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|
| Train Acc. | Test Acc. | Train Acc. | Test Acc. | Train Acc. | Test Acc. | Train Acc. | Test Acc. | Train Acc. | Test Acc. | Train Acc. | Test Acc. | |
| KFIS (linear kernel) | 95.71 | 93.06 (2.9) | 97.81 | 97.22 (7.6) | 96.86 | 91.66 (14.7) | 94.5 | 94.44 (24.6) | 96.88 | 95.83 (30.4) | 95.98 | 95.83 (37.1) |
| KFIS (poly kernel) | 98.55 | 98.61 (44.3) | 97.19 | 97.22 (52.1) | 97.83 | 95.83 (60.2) | 94.31 | 95.83 (79.5) | 96.79 | 95.83 (81.5) | 96.76 | 95.83 (80.7) |
| KFIS (RBF kernel) | 99.24 | 97.22 (5.5) | 95.71 | 95.83 (13.4) | 96.55 | 97.22 (18.8) | 92.12 | 93.05 (26.1) | 92.07 | 93.05 (31.3) | 89.36 | 87.50 (36.8) |
| KFIS (tansig kernel) | 98.71 | 98.61 (41.7) | 97.19 | 97.22 (53.4) | 97.5 | 95.83 (69.9) | 93.88 | 94.44 (81.1) | 96.92 | 96.83 (80.7) | 96.62 | 96.83 (84.2) |
|
| ||||||||||||
| SVM (linear Kernel) | 97.22 | 97.22 (3) | 97.37 | 97.22 (3.5) | 96.61 | 94.44 (3.6) | 97.22 | 95.83 (3.8) | 97.22 | 95.83 (4) | 97.84 | 97.22 (4.2) |
| SVM (poly Kernel) | 96.75 | 91.67 (1.5) | 96.14 | 94.44 (1.6) | 96.76 | 93.06 (1.7) | 95.83 | 93.06 (2) | 97.22 | 97.22 (2.2) | 97.22 | 97.22 (2.3) |
| SVM (RBF Kernel) | 97.68 | 94.44 (2) | 97.84 | 97.22 (2.3) | 99.38 | 100.00 (2.7) | 98.00 | 95.83 (3.2) | 98.61 | 98.61 (3.7) | 98.15 | 98.61 (4.7) |
| SVM (tansig Kernel) | 98.00 | 97.22 (3.1) | 98.30 | 98.61 (3.3) | 98.15 | 95.83 (3.5) | 97.69 | 94.44 (3.7) | 97.22 | 95.83 (4) | 97.84 | 97.22 (4.7) |
Figure 8Comparison of testing accuracy of K-FIS using different feature set.
(a) F5
| 0 | 1 | |
| 0 | 44 | 3 |
| 1 | 2 | 23 |
(b) F10
| 0 | 1 | |
| 0 | 46 | 1 |
| 1 | 1 | 24 |
(c) F15
| 0 | 1 | |
| 0 | 43 | 4 |
| 1 | 2 | 23 |
(d) F20
| 0 | 1 | |
| 0 | 44 | 4 |
| 1 | 0 | 25 |
(e) F25
| 0 | 1 | |
| 0 | 45 | 2 |
| 1 | 1 | 24 |
(f) F30
| 1 | 0 | |
| 0 | 45 | 2 |
| 1 | 1 | 24 |
(a) F5
| 0 | 1 | |
| 0 | 46 | 1 |
| 1 | 0 | 25 |
(b) F10
| 0 | 1 | |
| 0 | 46 | 1 |
| 1 | 1 | 24 |
(c) F15
| 0 | 1 | |
| 0 | 45 | 2 |
| 1 | 1 | 24 |
(d) F20
| 0 | 1 | |
| 0 | 45 | 2 |
| 1 | 1 | 24 |
(e) F25
| 0 | 1 | |
| 0 | 45 | 2 |
| 1 | 1 | 24 |
(f) F30
| 0 | 1 | |
| 0 | 45 | 2 |
| 1 | 1 | 24 |
(a) F5
| 0 | 1 | |
| 0 | 45 | 2 |
| 1 | 0 | 25 |
(b) F10
| 0 | 1 | |
| 0 | 45 | 2 |
| 1 | 1 | 24 |
(c) F15
| 0 | 1 | |
| 0 | 45 | 2 |
| 1 | 0 | 25 |
(d) F20
| 0 | 1 | |
| 0 | 42 | 5 |
| 1 | 0 | 25 |
(e) F25
| 0 | 1 | |
| 0 | 42 | 5 |
| 1 | 0 | 25 |
(f) F30
| 0 | 1 | |
| 0 | 40 | 7 |
| 1 | 2 | 23 |
(a) F5
| 0 | 1 | |
| 0 | 46 | 1 |
| 1 | 0 | 25 |
(b) F10
| 0 | 1 | |
| 0 | 46 | 1 |
| 1 | 1 | 24 |
(c) F15
| 0 | 1 | |
| 0 | 45 | 2 |
| 1 | 1 | 24 |
(d) F20
| 0 | 1 | |
| 0 | 45 | 2 |
| 1 | 2 | 23 |
(e) F25
| 0 | 1 | |
| 0 | 45 | 2 |
| 1 | 1 | 24 |
(f) F30
| 0 | 1 | |
| 0 | 45 | 2 |
| 1 | 1 | 24 |