Literature DB >> 30108438

Co-ABC: Correlation artificial bee colony algorithm for biomarker gene discovery using gene expression profile.

Hala Mohammed Alshamlan1.   

Abstract

In this paper, we propose a new hybrid method based on Correlation-based feature selection method and Artificial Bee Colony algorithm,namely Co-ABC to select a small number of relevant genes for accurate classification of gene expression profile. The Co-ABC consists of three stages which are fully cooperated: The first stage aims to filter noisy and redundant genes in high dimensionality domains by applying Correlation-based feature Selection (CFS) filter method. In the second stage, Artificial Bee Colony (ABC) algorithm is used to select the informative and meaningful genes. In the third stage, we adopt a Support Vector Machine (SVM) algorithm as classifier using the preselected genes form second stage. The overall performance of our proposed Co-ABC algorithm was evaluated using six gene expression profile for binary and multi-class cancer datasets. In addition, in order to proof the efficiency of our proposed Co-ABC algorithm, we compare it with previously known related methods. Two of these methods was re-implemented for the sake of a fair comparison using the same parameters. These two methods are: Co-GA, which is CFS combined with a genetic algorithm GA. The second one named Co-PSO, which is CFS combined with a particle swarm optimization algorithm PSO. The experimental results shows that the proposed Co-ABC algorithm acquire the accurate classification performance using small number of predictive genes. This proofs that Co-ABC is a efficient approach for biomarker gene discovery using cancer gene expression profile.

Entities:  

Keywords:  ABC; Artificial bee colony; CFS; Cancer classification; Correlation-based feature selection; Gene expression profile; Gene selection method

Year:  2018        PMID: 30108438      PMCID: PMC6088113          DOI: 10.1016/j.sjbs.2017.12.012

Source DB:  PubMed          Journal:  Saudi J Biol Sci        ISSN: 1319-562X            Impact factor:   4.219


Introduction

Gene expression profile or microarray data can be used to identify which genes are being expressed in a given cell type at a particular time and under particular conditions. This very helpful to compare the gene expression in two different cell types or tissue samples. Therefore, we can select the most informative and predictive genes that are responsible for causing a specific disease or cancer (Alshamlan et al., 2013, Alba et al., 2007). However, gene expression profile is conceded as high dimensional dataset. In other word, it suffers from the curse of dimensionality, the small number of samples, and the level of irrelevant and noise genes, all of which makes the classification task for a given sample more challenging (Alshamlan et al., 2013, Ghorai et al., 2010, Sheng-Bo et al., 2006). In this paper, we developed a new gene selection method to select the smallest subset of informative genes that are most predictive to its relative class using a classification model. In addition our new algorithm aim to determine the genes that contribute the most to cancer diagnosis, which would assist in drug discovery and early diagnosis, and increase the classifier’s ability to classify new samples accurately. The artificial bee colony (ABC) algorithm is an effective meta-heuristic algorithm that was invented in 2005 by Karaboga (2005). ABC algorithm was inspired by the social life of bees and is used to look for an optimal solution in numerical optimization problems (Karaboga, 2005). Becouse its simplicity and ease of implementation, the ABC algorithm; it has been widely applied in many optimization applications such as protein tertiary structures (Bahamish et al., 2009), digital IIR filters (Karaboga, 2009), artificial neural networks (Karaboga and Akay, 2005) and others. However, the ABC algorithm suffer from major critical problems, which is shared and similar to other evolutionary algorithms. Especially in computational efficiency, when it is applied to high dimensional dataset such as gene expression profile. In our previous researches, we applied ABC algorithm with support vector machine (SVM) classifier in order to generate new wrapper feature selection method, which is named ABC-SVM (Alshamlan et al., 2016). Also, we developed a new hybrid feature selection algorithm based on ABC, called mRMR-ABC (Alshamlan et al., 2015). In mRMR -ABC, we successfully combined minimum Redundancy Maximum Relevance (mRMR) filter algorithm with ABC algorithm in order to select informative genes that are minimum redundancy for other genes and maximum relevancy for specific cancer classe. This hybrid mRMR-ABC gene selection algorithm offers a good tradeoff between computationally effective and feature dependencies. However, we did not get high classification accuracy with some of microarray dataset. In order to solve this problem and further improve the performance of the ABC algorithm. In this paper, we propose combined a Correlation-based Feature Selection (CFS) filtering method, as a preprocessing stage, with ABC algorithm. It worth mentioning that CFS can be effectively combined with other feature selectors, such as wrappers. This can be done to find a very compact subset from candidate features at lower expense. Chuang et al. (2011) proposed a novel hybrid gene selection method by combining the CFS and TGA methods. The experimental results for both binary and multi-class cancer microarray datasets show that the proposed method reduced the dimensionality of microarray datasets by illuminating the redundant genes and achieved high classification accuracy. In addition, Yang et al. (2008) proposed an improved binary particle swarm optimization (IBPSO), which further developed the standard BPSO. The filter methods applied in this study were information gain (IG) and correlation-based feature selection (CFS). The authors used the Weka software package (Quinlan, 1986) to determine the information value of each feature and to sort the features in accordance with their information gain value. The wrapper method that was adopted in this study was IBPSO. The authors evaluated the performance of these hybrid methods using a leukemia dataset. The experimental results indicated that CFS with IBPSO achieved a minimum number of selected genes, while IG with IBPSO produced high classification accuracy. In our proposed algorithm, which is named Co-ABC, we select the genes that have maximal correlation between genes and class and have minimal correlation between gene to gene. This step will reduced the dimensional of microarray dataset, because we will identify and select the relevant and informative genes only. After that, we applied ABC algorithm for those selected genes in order to select small number of predictive genes. Then, we will measure the efficiency of the selected genes using a support vector machine (SVM) as a classifier. We used an SVM classifier because its displayed substantial benefits when compared to other classification approaches (Alshamlan et al., 2014). In addition, it addresses this problem by mapping the input space into a high-dimensional feature (gene) space. After that, it generates a linear classification decision to classify the initial dataset (microarray dataset) with a maximum margin hyperplane. An SVM is more efficient, very accurate, and faster than other machine learning methods, such as Neural Networks (NN) and k-Nearest Neighbour (K-NN) classifiers when they applied with gene expression profile (Wang and Gotoh, 2009). The Co-ABC algorithm is tested using six gene expression profile for binary and multi-class cancer datasets. Also, it compared with our previous proposed algorithms ABC-SVM (Alshamlan et al., 2016), and mRMR-ABC (Alshamlan et al., 2015). In addition, in order to proof the efficiency of our proposed Co-ABC algorithm, we compare it with previously known related methods. Two of these methods was re-implemented for the sake of a fair comparison using the same parameters. These two methods are: Co-GA, which is CFS combined with a genetic algorithm GA. The second one named Co-PSO, which is CFS combined with a particle swarm optimization algorithm PSO. Furthermore, Co-ABC was compared with other related and recently published algorithms. The experimental results show improvements in both the number of selected informative genes and cancer classification accuracy. The rest of this paper is organized as follows: Section 2 provides a description of proposed Co-ABC algorithm. Section 3 outlines the experimental setup and provides results. Finally, Section 4 concludes our paper.

The proposed Co-ABC algorithm

In this section, we present the proposed Co-ABC algorithm for elect the highly informative genes from the cancer gene expression profile. As shown in Fig. 1, Co-ABC consists of three main phases: preprocessing phase, gene selection phase, and classification phase. In the following sub section, we introduce the function of each phase.
Fig. 1

The main phases and steps of the proposed Co-ABC algorithm.

The main phases and steps of the proposed Co-ABC algorithm.

Preprocessing phase: Correlation-based Feature Selection (CFS) filter method

Correlation-based feature selection (CFS) scores (and ranks) the worth of subsets of features according to a correlation-based heuristic evaluation function, rather than scoring (and ranking) individual features (Yvan et al., 2007). As microarray feature (genes) space is usually huge, CFS uses a best-first-search heuristic that takes into account the usefulness of individual features for predicting the class. Therefore, CFS selects the subset that has maximal correlation to the class, and minimal correlation between features (Yvan et al., 2007). CFS first calculates a matrix of (feature to class) and (feature to feature) correlations from the training data. Then, a score for the subset of features assigned by the heuristic is calculated using Eq. (1).where is the heuristic merit of a feature subset S containing k features, is the average correlation between features and class, and is the average correlation between features. CFS starts from the empty set of features and then the subset with the highest Merit found during the search will be selected. In our problem, genes which have correlation with specific cancer classes should be selected using the CFS method. As microarray feature (genes) space is usually huge, CFS uses a best-first-search heuristic that takes into account the usefulness of individual features for predicting the class. The main purpose of applying the CFS gene selection method is to find the highly correlated subset of genes from initial microarray dataset. As illustrate in Fig. 2, the initial gene expression profile is preprocessed using the CFS filtering method. Each gene is evaluated and sorted based one CFS criteria as explained previously in this section. The highly correlated genes that give high classification accuracy with an SVM classifier will selected to create a new subset named the CFS dataset. Suppose the initial microarray dataset contains S genes, as shown in Fig. 2. After applying CFS filter method, the number of genes will be reduced to m genes that have the maximal correlation between genes to class and minimal correlation between gene to gene.
Fig. 2

Co-ABC dataset which is contains the highly correlated genes m selected by the CFS filter approach.

Co-ABC dataset which is contains the highly correlated genes m selected by the CFS filter approach. Our main objective is to maximize the classification accuracy and reduce the number of informative genes. Hence, we adopted the CFS filter methods as a preprocessing step for ABC algorithm to enhance the speed and classification accuracy performance of the search. In addition, in order to eliminate the irrelevant genes and filter the noisy genes, which is reduces the computational complexity for the ABC algorithm and SVM classifier as well.

Gene selection phase: Artificial Bee Colony (ABC) algorithm

In this phase, we apply ABC algorithm to elect and identify the most informative and predictive genes from an CFS dataset that achieve the highest classification accuracy with an SVM classifier.The ABC algorithm is a meta-heuristic evolutionary algorithm that simulates the search for food in a group of bees. An ABC algorithm is implemented as presented in our prevues research article (Alshamlan et al., 2015). This section will present briefly the main step of ABC algorithm and how can we apply them in gene selection for gene expression data analysis, which is already explained in more detail in our prevues research article (Alshamlan et al., 2015).

Employed bee step

In this step, the employee bees looking around the solutions (food resources) at in order to find the best genes index at the new location . We determined the new gene index using following equation (Xiang and An, 2013):where is the new gene index (the location vector in the artificial bees), is the current gene index (the location vector of the bee), is a right random number in , and the SN is the number of artificial bees, which considers as a solution in our problem. is a random number that is distributed randomly between . The selection of random numbers from the CFS dataset index is done using the following equation (Xiang and An, 2013):where and are the high (up) limit and the low (down) limit of the variable respectively, , and . While, is the random selected numbers between . When the new selected indes of the gene is determined, the optimization of it must be computed based on the fitness function. In this problem, our fitness value identified based on the solution classification accuracy using an SVM classifier. If the new fitness value is greaterr than the fitness value acquired thus far, then the bee moves to the new solution (i.e,food source) leaving the old one; otherwise it stay in the old one.

Onlooker bee step

After the employed bees complete there task, looking for the best solutions, they shared the information with onlooker bees. Then, an onlooker bee selects the genes depending on their winning probability value that is very similar to roulette wheel selection in genetic algorithm (GA) as follows: the possibility of selecting the particular solution (food source) by the onlooker bees is calculated using the following equation:

Scout bee step

Each employee or onlooker bee will search for best genes (solution) for a certain number and limited number of cycles. If the fitness value for any employee or looked bee does not improve, then that particular bee becomes a scout bee. In other word, a solution which could not be improved through “limit” number of trials becomes a scout bee. A scout bee select an index of genes randomly from CFS dataset (search space). It is worth mentioning that the ABC algorithm suffer from some critical issues, especially in computational and speed efficiency, when it is applied on huge and high dimensional dataset such as a cancer gene expression profile. This motivates us to solve these difficulties and further enhance the performance of the ABC algorithm by proposing a hybrid technique between the ABC algorithm and mRMR filter gene selection approach, namely, the mRMR-ABC algorithm. In the following subsection, we explain the mRMR algorithm when applied to our problem.

Classification phase: Support Vector Machine (SVM) classifier

Support vector machines (SVM) is a classification algorithm that belongs to a new generation of learning system based on recent advances in statistical learning theory (Vapnik, 1998). A SVM is very effective classification algorithm that showed a good performance in a variety of computational biological classification tasks. It worth mentioning that SVM-based classifiers are becoming increasingly popular classifiers for gene expression profile. Support vector machine is very helpful in cancer diagnostic models, where the number of feature, which is genes in our problem, is so large in relative to the sample size. Because, it has the ability to be fitted with all genes and at the same time with stable performance when using the full set of genes (Alonso et al., 2012, Huerta et al., 2006, Lee and Leu, 2011, Mukherjee, 2003). Also, SVM is aim is to identify the hyperplane that is separating the feature with the largest margin (distance between itself and the closest samples from each classes). Generally, the effective SVM classifier seeks to a trade-off between increasing the margin and decreasing the number of errors. In our proposed Co-ABC algorithm, we use the informative and predictive genes that are predicted from the second phase to train the SVM classifier. Then, the SVM is applied again to classify the testing gene expression dataset and restore the classification accuracy. The main steps for the proposed Co-ABC algorithm illustrated in Fig. 1. In addition, the pseudo code for the proposed Co-ABC algorithm is demonstrated in Algorithm 1. Co-ABC algorithm

Experimental setup and results

Experiential setup

In this section, we evaluated the overall performance of the Co-ABC algorithm using six more useful benchmark binary and multi-class microarray cancer datasets, which we used to evaluate our previously proposed algorithms ABC-SVM (Alshamlan et al., 2016), and mRMR-ABC (Alshamlan et al., 2015). The binary-class microarray datasets are: colon (Alon et al., 1999), leukemia (Golub et al., 1999), and lung (Beer et al., 2002). Where the multi-class microarray datasets are: SRBCT (Khan et al., 2001), lymphoma (Alizadeh et al., 2000), and leukemia (Armstrong et al., 2001). Table 2 shows a detailed description of these six popular cancer microarray datasets with illustration of number of classes, number of samples, number of genes, and a brief description of each dataset construction.
Table 2

The cancer microarray datasets statistical values.

Microarray datasetsNo of classesNo of samplesNo of genesDescription
Colon Alon et al. (1999)262200040 cancer samples and 22 normal samples
Leukemia1 Golub et al. (1999)272712925 AML samples and 47 ALL samples
Lung Beer et al. (2002)296712986 cancer samples and 10 normal samples
SRBCT Khan et al. (2001)483230829 EWS cancer samples, 18 NB cancer samples, 11 BL cancer samples, and 25 RMS cancer samples
Lymphoma Alizadeh et al. (2000)362402642 DLBCL cancer samples, 9 FL cancer samples, and 11 B-CLL cancer samples
Leukemia2 Armstrong et al. (2001)372712928 AML sample, 24 ALL sample, and 20 MLL samples
In order to make fair evaluation and comparison, we applied same control parameters that have been used for our previous proposed algorithms. Table 1 shows the control parameters for the Co-ABC algorithm applied in this research. The first control parameter is the bee colony size or population size. The second one is the maximum cycle or maximum number of generations. The third control parameter is the number of runs, we use it a stopping criteria. The fourth control parameter is the non-improved limit, which is means the number of iterations allowed when the food source is not enhanced (i.e, exhausted). If the food source(bee) exceeds this limit value, it will become as scout bee.
Table 1

The control parameters for Co-ABC algorithm.

Control parameterValue
Beecolonysize80
Maximumnumberofcycle100
Numberofruns30
Nonimprovedlimit5
The control parameters for Co-ABC algorithm. The cancer microarray datasets statistical values. In this research paper, we tested the performance of the proposed Co-ABC algorithm by comparing it with our previously proposed algorithms ABC-SVM (Alshamlan et al., 2016), and mRMR-ABC (Alshamlan et al., 2015) using two parameters: the first one is the classification accuracy and the second is the number of predictive and informative genes that have been applied in cancer classification task. Moreover, we adopt leave-one-out cross-validation (LOOCV) (Ng et al., 1997) in order to evaluate the performance of Co-ABC and the existing methods in the literature. We applied LOOCV in this study, because it is more suitable to our research problem, and it has the capability to prevent and reduce the “overfitting” problem (Ng et al., 1997).

Experimental results

In this section, we illustrate and evaluate the results that are generated from Co-ABC algorithm. First, we applied CFS filter method to select the highly correlated genes that acquire accurate classification result using SVM classifier. From Table 3, we can observe that the most correlated 80 genes from leukemia1 dataset achieve 100% accuracy. For colon cancer dataset, we can generate 91.94% classification accuracy by 25 genes. While, in the lung cancer dataset, we got 100% with 71 genes and 110 correlated genes to achieve the same performance percentage for the SRBCT dataset. Also, by selecting 184 highly correlated genes from the lymphoma cancer dataset and 103 correlated genes from the leukemia2 cancer dataset, we generated 100% as classification performance.
Table 3

The CFS with an SVM classification performance.

Microarray datasetsNumber of genesClassification accuracy
Colon2591.94%
Leukemia180100%
Lung71100%
SRBCT110100%
Lymphoma184100%
Leukemia2103100%
The CFS with an SVM classification performance. When we compared the classification accuracy performance of CFS with mRMR (Alshamlan et al., 2015) using SVM classifier. We noted that the CFS achieved same classification accuracy used small number of genes exempt the colon and lymphoma datasets. In colon dataset, CFS achieved lass classification accuracy than mRMR, which is 91.94%. While in lymphoma dataset, CFS used more genes to generate 100% classification accuracy. After employing CFS filter method, we used these highly correlated genes as input in the ABC algorithm to select the predictive and informative genes from these correlated ones. In addition, we compare the performance of the proposed Co-ABC algorithm with the previously proposed ABC-based algorithm, ABC-SVM (Alshamlan et al., 2016), and mRMR-ABC (Alshamlan et al., 2015), with the similar number of selected genes for all six benchmark gene expression microarray datasets. The performance comparison for binary-class cancer dataset, which are colon, leukemia1, and lung are presented in Table 4, Table 5, Table 6, respectively. While, Table 7, Table 8, Table 9, respectively, show the performance comparison for multi-class cancer datasets, which are SRBCT, lymphoma, and leukemia2. As demonstrated on these tables, we noted that Co-ABC algorithm performs better than mRMR-ABC algorithm and ABC-SVM algorithm for all cancer datasets (binary or multi- classes) with different number of predictive genes. In addition, in order to make our experiments statistically validated, we apply each experiment 30 times for all cancer dataset. Furthermore, the best, worst, and average classification accuracies results of the 30 independent runs are computed to evaluate the performance of Co-ABC algorithm. The values obtained for classification accuracy are further analyzed using the Kruskal–Wallis test (Kruskal and Wallis, 1952) for statistical significance. The p-values obtained for classification accuracy and discovered informative genes is 0.0063, which indicates the statistical significance of the results.
Table 4

The classification performance of the Co-ABC algorithm with comparison to mRMR-ABC algorithm and ABC-SVM algorithm for Colon cancer dataset.

Number of genesClassification accuracy
Co-ABC
mRMR-ABC
ABC-SVM
BestMeanWorstBestMeanWorstBestMeanWorst
390.32%90.16%88.71%88.71%87.50%85.48%87.10%85.91%83.87%
491.94%91.34%90.32%90.23%88.27%87.10%87.10%86.71%85.48%
591.94%91.94%91.94%91.94%89.50%87.10%90.32%87.98%85.48%
693.55%92.42%91.94%91.94%90.12%87.10%90.32%88.44%85.48%
795.16%93.55%91.94%93.55%91.64%88.81%91.94%90.20%88.81%
895.16%94.25%93.55%93.55%91.80%88.81%91.94%90.61%88.81%
996.77%94.62%93.55%93.55%92.11%90.16%91.94%90.95%88.81%
1096.77%94.68%93.55%93.55%92.74%90.16%93.55%91.31%88.81%
1595.16%94.95%93.55%96.77%93.60%91.93%93.55%91.38%90.32%
2095.16%93.44%91.94%96.77%94.17%91.93%95.61%92.44%90.32%
Table 5

The classification performance of the Co-ABC algorithm with comparison to mRMR-ABC algorithm and ABC-SVM algorithm for Leukemia1 cancer dataset.

Number of genesClassification accuracy
Co-ABC
mRMR-ABC
ABC-SVM
BestMeanWorstBestMeanWorstBestMeanWorst
297.22%97.22%97.22%91.66%89.63%81.94%87.5%86.45%81.94%
3100%99.58%98.61%93.05%90.37%83.33%88.88%89.82%83.33%
4100%100%100%94.44%91.29%86.11%88.8%91.15%83.33%
14100%100%100%100%95.83%93.05%93.05%92.51%88.88%
Table 6

The classification performance of the Co-ABC algorithm with comparison to mRMR-ABC algorithm and ABC-SVM algorithm for Lung cancer dataset.

Number of genesClassification accuracy
Co-ABC
mRMR-ABC
ABC-SVM
BestMeanWorstBestMeanWorstBestMeanWorst
297.91%97.91%97.91%96.87%95.83%93.75%88.54%87.5%84.37%
3100%100%100%97.91%96.31%93.75%89.58%88.54%84.37%
4100%100%100%98.95%97.91%96.87%91.66%89.58%87.5%
8100%100%100%100%98.95%96.87%97.91%93.75%91.66%
Table 7

The classification performance of the Co-ABC algorithm with comparison to mRMR-ABC algorithm and ABC-SVM algorithm for SRBCT cancer dataset.

Number of genesClassification accuracy
Co-ABC
mRMR-ABC
ABC-SVM
BestMeanWorstBestMeanWorstBestMeanWorst
277.11%77.03 %75.90 %75.90%71.08%68.67%72.28%69.87%67.46%
389.16%86.51%83.13%85.54%79.51%71.08%73.34%71.08%68.67%
4100%95.82%92.77%87.95%84.33%77.10%84.33%81.92%77.10%
5100%98.43%96.38%91.56%86.74%84.33%87.95%84.33%77.10%
10100%98.43%96.38%100%96.30%92.77%95.36%91.56%89.15%
Table 8

The classification performance of the Co-ABC algorithm with comparison to mRMR-ABC algorithm and ABC-SVM algorithm for Lymphoma cancer dataset.

Number of genesClassification accuracy
Co-ABC
mRMR-ABC
ABC-SVM
BestMeanWorstBestMeanWorstBestMeanWorst
2100%99.1%98.48%86.36%86.36%86.36%86.36%86.36%86.36%
3100%100%100%93.93%90.90%86.36%89.39%87.87%86.36%
5100%100%100%100%96.96%93.93%96.96%92.42%90.90%
Table 9

The classification performance of the Co-ABC algorithm with comparison to mRMR-ABC algorithm and ABC-SVM algorithm for Leukaemia2 cancer dataset.

Number of genesClassification accuracy
Co-ABC
mRMR-ABC
ABC-SVM
BestMeanWorstBestMeanWorstBestMeanWorst
287.5%87.5%87.5%84.72%84.72%84.72%84.72%84.72%84.72%
397.22%96.32%95.83%87.5%86.11%84.72%86.11%85.23%84.72%
498.61%96.81%95.83%90.27%87.5%84.72%87.5%86.11%84.72%
598.61%98.19%97.22%90.27%88.88%86.11%87.5%86.45%84.72%
6100%99.21%98.61%94.44%90.27%87.5%90.27%88.88%86.11%
20100%99.21%98.61%100%96.12%95.83%97.22%93.15%91.66%
The classification performance of the Co-ABC algorithm with comparison to mRMR-ABC algorithm and ABC-SVM algorithm for Colon cancer dataset. The classification performance of the Co-ABC algorithm with comparison to mRMR-ABC algorithm and ABC-SVM algorithm for Leukemia1 cancer dataset. The classification performance of the Co-ABC algorithm with comparison to mRMR-ABC algorithm and ABC-SVM algorithm for Lung cancer dataset. The classification performance of the Co-ABC algorithm with comparison to mRMR-ABC algorithm and ABC-SVM algorithm for SRBCT cancer dataset. The classification performance of the Co-ABC algorithm with comparison to mRMR-ABC algorithm and ABC-SVM algorithm for Lymphoma cancer dataset. The classification performance of the Co-ABC algorithm with comparison to mRMR-ABC algorithm and ABC-SVM algorithm for Leukaemia2 cancer dataset. In order to sake of a fair comparison using the same parameters. In this research, we re-implement two related evolutionary based algorithm: Co-GA, which is CFS combined with a genetic algorithm GA. The second one named Co-PSO, which is CFS combined with a particle swarm optimization algorithm PSO. Furthermore, we compared Co-ABC with recently published related algorithms. Notably, all these algorithms that are under comparison have been combined with the support vector machine (SVM) for classification task. Table 10 present the performance comparison results of the Co-ABC algorithm and other related algorithms. Compared with the mRMR-ABC algorithm, the mRMR-ABC algorithm selected 15 genes to achieve 96.77% classification accuracy. In contrast, the Co-ABC algorithm achieves same classification accuracy using only 9 genes.
Table 10

The classification performance of the related algorithms under comparison for six cancer gene expression profile Numbers between parentheses means the numbers of informative genes that has been used in classification task.

AlgorithmsColonLeukemia1LungSRBCTLymphomaLeukemia2
Co-ABC96.77(9)100(3)100(2)100(4)100(2)100(6)
CFS-GA90.32(8)100(24)100(20)100(38)100(17)100(36)
CFS-PSO91.94(7)100(15)100(5)100(35)
mRMR-ABC Alshamlan et al. (2015)96.77(15)100(14)100(8)100(10)100(5)100(20)
ABC-SVM Alshamlan et al. (2016)95.61(20)93.05(14)97.91(8)95.36(10)96.96(5)97.22(20)
PSO Qi et al. (2007)85.48(20)94.44(23)
PSO Javad and Giveki (2013)87.01(2000)93.06 (7129)
mRMR-PSOJavad et al. (2012)90.32(10)100(18)
GADPLee and Leu (2011)100(6)
mRMR-GA Amine et al. (2009)100(15)95(5)
ESVM Huang and Chang (2007)95.75(7)98.75(6)
MLHD-GA Huang et al. (2007)97.1(10)100(11)100(6)100(9)
CFS-IBPSO Yang et al. (2008)100(6)98.57(41)
GA Peng et al. (2003)93.55(12)
mAnt Yu et al. (2009)91.5(8)100(7)
The classification performance of the related algorithms under comparison for six cancer gene expression profile Numbers between parentheses means the numbers of informative genes that has been used in classification task. For the Leukemia1 dataset, Co-ABC generate very accurate calcification, which is 100% using only three informative genes. As observed in Table 10, there are three algorithms acquired 100% accuracy result, however, their informative genes are greater. The mRMR-ABC algorithm got 100% classification accuracy using 14 genes. Also, the mRMR-PSO algorithm proposed by Javad et al. (2012) acquired 100% classification percentage with 18 informative genes. And, the Co-GA algorithm select 24 genes to get 100% classification accuracy. For the Lung dataset, the Co-ABC algorithm make a superior improvement by select only two predictive genes to achieve 100% classification result. This accuracy was achieved by other algorithms, which are CFS-GA, mRMR-ABC, and mRMR-GA algorithm proposed by Amine et al. (2009), however they using greater number of selected genes. In SRBCT dataset, the MLHD-GA algorithm developed by Huang et al. (2007) got 100% classification result with 11 predictive genes. Also, the mRMR-ABC algorithm used 10 predictive genes and acquires 100% classification accuracy. In addition, the CFS-GA and CFS-PSO achieve 100% classification accuracy using greater number of selected genes. By contrast, our proposed algorithm select only four informative genes to achieve 100% classification accuracy. For the Lymphoma cancer dataset, there are many proposed algorithms achieve 100%, the Co-ABC algorithm identified a fewer number of informative genes. The Co-ABC selected only two genes to achieve 100% classification accuracy, which is obviously improved the other existing result so far. For the Leukemia2 dataset, the MLHD-GA algorithm developed by Huang et al. (2007), select 9 genes to achieve 100% classification accuracy. While our previous mRMR-ABC algorithm used 20 genes to got 100% classification accuracy. In contrast, our proposed Co-ABC achieve a greatest improvement in number of selected genes demand. It select only 6 genes to acquire 100% classification performance. Computational complexity is an important aspect in algorithm assessment. Therefore, we tested and compared the time and memory complexity of Co-ABC with proposed bio-inspired meta-heuristics algorithms, ABC-SVM, and mRMR-ABC. Table 11 shows the average runtime in seconds for the Cot-ABC algorithm and other algorithms under comparison; the Co-ABC algorithm has a faster execution time. Table 12 shows the average memory space in virtual machine in giga bites for the Cot-ABC algorithm and other algorithms under comparison. The Co-ABC algorithm used approximately similar memory space as mRMR-ABC algorithm, and less than ABC-SVM algorithm.
Table 11

Average runtime (in s) for the Co-ABC algorithm and other classification algorithms.

AlgorithmsPreprocessing timeAverage classification timeTotal
Co-ABC with SVM24.37 s40.22 s64.59 s
mRMR-ABC with SVM25.17 s72.13 s97.3 s
ABC with SVM0.0 s134.74 s134.74 s
Table 12

Average memory space (in GB) for the Co-ABC algorithm and other classification algorithms.

AlgorithmsMemory space
Co-ABC with SVM1.56 GB
mRMR-ABC with SVM1.58 GB
ABC with SVM1.89 GB
Average runtime (in s) for the Co-ABC algorithm and other classification algorithms. Average memory space (in GB) for the Co-ABC algorithm and other classification algorithms. It is worth noted that the Co-ABC algorithm generally outperforms the previously reported results, it generates the highest classification accuracy and the lowest average of selected genes when evaluated using all six cancer datasets, as compared to the original ABC algorithm under the same cross-validation approach. In comparison between CFS and mRMR filter methods when combined with ABC algorithm, the CFS elects less number of genes than mRMR with relatively improved in classification accuracy. This implies that CFS method is able to improve the classification accuracy and it is an effective tool for identifying the highly correlated genes and omitting the non-relevant and noisy genes. Therefore, we can conclude that Co-ABC is a promising method for biomarker gene discovery using gene expression profile. The explanation of the highly correlative and informative genes that achieve highest classification accuracy for all microarray datasets using Co-ABC algorithm have been reported in Table 13.
Table 13

The highly correlative and informative genes that achieve highest classification accuracy for six cancer gene expression profile using Co-ABC algorithm.

DatasetsPredictive genesAccuracy
ColonGene 625, Gene1562, Gene576, Gene1328, Gene1917, Gene 1772, Gene682, Gene1200, Gene167196.77%
Leukemia1S50223_at, U05259_rna1_at, M23197_at100%
LungX64559_at, U19247_rna1_s_at100%
SRBCTGene123, Gene742, Gene1954, Gene1003100%
LymphomaGene2403X, Gene3519X100%
Leukemia2L47738_at, X00274_at, X58072_at, X95735_at, D63880_at, U48251_at100%
The highly correlative and informative genes that achieve highest classification accuracy for six cancer gene expression profile using Co-ABC algorithm. Furthermore, in order to evaluate the performance of Co-ABC algorithm result more accurately. Table 14 present the precision and sensitivity of Co-ABC algorithm with selected informative genes. Precision is the positive predictive value or the fraction of the positive predictions that are actually positive. While the specificity is the true negative rate or the proportion of negatives that are correctly identified.
Table 14

The precision and sensitivity of Co-ABC algorithm using selected informative genes.

DatasetsPrecisionSensitivity
Colon97.04%96.77%
Leukemia1100%100%
Lung100%100%
SRBCT100%100%
Lymphoma100%100%
Leukemia2100%100%
The precision and sensitivity of Co-ABC algorithm using selected informative genes. It worth noting, that one of the CFS filters method limitation, it selects only the highly correlated genes. In some dataset such as colon datasets, the gene expression not correlated with each other. For this reason, the CFS select only few number of genes and loos some predictive and important genes. This is definitely effect the classification accuracy. While in other datasets such as lymphoma dataset, there are many correlated genes. In this case, the CFS retrieves many related genes as presented in Table 3. Therefore, we recommend addressing this limitation in feature work.

Conclusion

In this research, we developed a new hybrid feature selection approach, called Co-ABC algorithm. In our proposed Co-ABC algorithm, we adopted the CFS filter method as a preprocessing step to the ABC algorithm to enhance the search speed and classification performance. In addition, in order to eliminate the unimportant and filter the noisy genes and reduces the computational complexity for the ABC algorithm with SVM as classifier. The major aim for adopting the CFS filter method is to find the highly correlated subset of genes from initial microarray dataset. Extensive experiments and comparisons were conducted using six binary and multi-class microarray cancer gene expression profile. The results showed that the proposed Co-ABC algorithm perfumes better than related algorithms. Moreover, the Co-ABC algorithm generates greater classification accuracy performance with fewer average of informative genes when tested using all six cancer datasets as compared to the original ABC algorithm under the same cross-validation technique (LOOCV). Also, when we compared the performance of CFS and mRMR when they combined with ABC algorithm, the CFS identify less number of predictive genes than mRMR with relatively greater classification accuracy performance. This implies that CFS method is able to improve the classification accuracy and it is an effective tool for identifying the highly correlated genes and omitting the non-relevant and noisy genes. Therefore, we can conclude that Co-ABC is a promising method for biomarker gene discovery using gene expression profile
1: Select the maximum relevant genes subset using CFS filter method that achieve highly classification accuracy with SVM Classifier from initial microarray dataset.
2: ABC parameters setting, include maximum cycles, bee colony size and limited trail.
3: Initialize ABC food sources randomly.
4: Food sources quality evaluation using fitness calculation, which is SVM classification accuracy.
5: Cycle1
6: WhileCycle<MaximumCyclesDo
7:  Generate new employed bees (new candidate solutions)
8:  New solution quality evaluation using fitness calculation.
9:  Adopt greedy selection approche.
10: Determine the probability values by using fitness values.
11: Generate new onlooker bees (new candidate solutions) using the probability of food source.
12: New solution quality evaluation using fitness calculation.
13: Adopt greedy selection process.
14: Identify abandoned solutions and produce new solutions randomly using scout bee.
15: Identify and save the best solution found so far.
16: CycleCycle+1
17: End While
18: Generate and return best solution (predictive and biomarker genes).
19: Train the SVM classifier algorithm using generated biomarker genes.
20: Classify gene expression profile using SVM classifier.
21: Calculate the classification accuracy
  16 in total

1.  Selecting a minimal number of relevant genes from microarray data to design accurate tissue classifiers.

Authors:  Hui-Ling Huang; Chong-Cheng Lee; Shinn-Ying Ho
Journal:  Biosystems       Date:  2006-07-10       Impact factor: 1.973

2.  ESVM: evolutionary support vector machine for automatic feature selection and classification of microarray data.

Authors:  Hui-Ling Huang; Fang-Lin Chang
Journal:  Biosystems       Date:  2006-12-16       Impact factor: 1.973

Review 3.  A review of feature selection techniques in bioinformatics.

Authors:  Yvan Saeys; Iñaki Inza; Pedro Larrañaga
Journal:  Bioinformatics       Date:  2007-08-24       Impact factor: 6.937

4.  A combination of modified particle swarm optimization algorithm and support vector machine for gene selection and tumor classification.

Authors:  Qi Shen; Wei-Min Shi; Wei Kong; Bao-Xian Ye
Journal:  Talanta       Date:  2006-09-01       Impact factor: 6.057

5.  A hybrid feature selection method for DNA microarray data.

Authors:  Li-Yeh Chuang; Cheng-Huei Yang; Kuo-Chuan Wu; Cheng-Hong Yang
Journal:  Comput Biol Med       Date:  2011-03-03       Impact factor: 4.589

6.  Broad patterns of gene expression revealed by clustering analysis of tumor and normal colon tissues probed by oligonucleotide arrays.

Authors:  U Alon; N Barkai; D A Notterman; K Gish; S Ybarra; D Mack; A J Levine
Journal:  Proc Natl Acad Sci U S A       Date:  1999-06-08       Impact factor: 11.205

7.  Classification and diagnostic prediction of cancers using gene expression profiling and artificial neural networks.

Authors:  J Khan; J S Wei; M Ringnér; L H Saal; M Ladanyi; F Westermann; F Berthold; M Schwab; C R Antonescu; C Peterson; P S Meltzer
Journal:  Nat Med       Date:  2001-06       Impact factor: 53.440

8.  Gene-expression profiles predict survival of patients with lung adenocarcinoma.

Authors:  David G Beer; Sharon L R Kardia; Chiang-Ching Huang; Thomas J Giordano; Albert M Levin; David E Misek; Lin Lin; Guoan Chen; Tarek G Gharib; Dafydd G Thomas; Michelle L Lizyness; Rork Kuick; Satoru Hayasaka; Jeremy M G Taylor; Mark D Iannettoni; Mark B Orringer; Samir Hanash
Journal:  Nat Med       Date:  2002-07-15       Impact factor: 53.440

9.  Microarray-based cancer prediction using soft computing approach.

Authors:  Xiaosheng Wang; Osamu Gotoh
Journal:  Cancer Inform       Date:  2009-05-26

10.  A novel weighted support vector machine based on particle swarm optimization for gene selection and tumor classification.

Authors:  Mohammad Javad Abdi; Seyed Mohammad Hosseini; Mansoor Rezghi
Journal:  Comput Math Methods Med       Date:  2012-07-26       Impact factor: 2.238

View more
  2 in total

1.  Melittin inhibits proliferation, migration and invasion of bladder cancer cells by regulating key genes based on bioinformatics and experimental assays.

Authors:  Jie Yao; Zhan Zhang; Sheng Li; Bai Li; Xing-Huan Wang
Journal:  J Cell Mol Med       Date:  2019-11-05       Impact factor: 5.295

2.  Cancer Categorization Using Genetic Algorithm to Identify Biomarker Genes.

Authors:  M Sathya; M Jeyaselvi; Shubham Joshi; Ekta Pandey; Piyush Kumar Pareek; Sajjad Shaukat Jamal; Vinay Kumar; Henry Kwame Atiglah
Journal:  J Healthc Eng       Date:  2022-02-22       Impact factor: 2.682

  2 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.