Literature DB >> 30155265

Evolutionary sequential genetic search technique-based cancer classification using fuzzy rough nearest neighbour classifier.

Loganathan Meenachi1, Srinivasan Ramakrishnan1.   

Abstract

Cancer is one of the deadly diseases of human life. The patient may likely to survive if the disease is diagnosed in its early stages. In this Letter, the authors propose a genetic search fuzzy rough (GSFR) feature selection algorithm, which is hybridised using the evolutionary sequential genetic search technique and fuzzy rough set to select features. The genetic operator's selection, crossover and mutation are applied to generate the subset of features from dataset. The generated subset is subjected to the evaluation with the modified dependency function of the fuzzy rough set using positive and boundary regions, which act as a fitness function. The generation and evaluation of the subset of features continue until the best subset is arrived at to develop the classification model. Selected features are applied to the different classifiers, from the classifiers fuzzy-rough nearest neighbour (FRNN) classifier, which outperforms in terms of classification accuracy and computation time. Hence, the FRNN is applied for performance analysis of existing feature selection algorithms against the proposed GSFR feature selection algorithm. The result generated from the proposed GSFR feature selection algorithm proved to be precise when compared to other feature selection algorithms.

Entities:  

Keywords:  FRNN classifier; GSFR feature selection algorithm; boundary regions; cancer; classification accuracy; computation time; deadly diseases; evolutionary sequential genetic search technique-based cancer classification; feature selection; fitness function; fuzzy rough nearest neighbour classifier; fuzzy rough set; fuzzy set theory; generated subset; genetic algorithms; genetic operator; genetic search fuzzy rough feature selection algorithm; mathematical operators; medical computing; modified dependency function; pattern classification; positive regions; rough set theory; search problems

Year:  2018        PMID: 30155265      PMCID: PMC6103784          DOI: 10.1049/htl.2018.5041

Source DB:  PubMed          Journal:  Healthc Technol Lett        ISSN: 2053-3713


Introduction

The computerised classification of data for diagnosis of diseases plays a pivotal role in biomedical field. The cancer is one of the deadly diseases. For every disease or ailment for that matter, early the diagnosis, lesser the mortality rate. Various computational diagnostic techniques are available to classify the malady. The clinical dataset or the dataset made known from machine learning repository form the basis to predict the cancer data and the predicted data for early clinical decision with higher reliability. From the large dataset, the classification of cancer stage of a patient is diagnosed by taking into account the informative features [1]. The feature selection is the process where the irrelevant, redundant features are removed from the original features [2]. The feature selection methods are categorised as filter, wrapper and hybrid methods [3]. In filter method, the features are considered individually by ranking and selection of features, which is either by forward selection or by backward elimination. It never receives feedback from any of the learning algorithm. The next method is wrapper method wherein the subset of features are taken and evaluated through the induction of the learning algorithm. The last method is hybrid method, which is a combination of both. The model generated from the classification produces better accuracy with lesser computation time through the informative features from the dataset [1-4]. There are numerous research done in cancer data classification. Onan [5] proposed an approach of fuzzy-rough nearest neighbour (FRNN) classifier which is combined with consistency based subset evaluation and fuzzy rough instance selection algorithms with a view to generating the model with minimum features and instances that provide accurate classification similar to that of the whole dataset. Jain et al. [6] proposed a hybrid model for cancer classification by hybridising correlation-based feature selection (CFS) and improved-binary particle swarm optimisation. This hybrid model selects the features to classify the binary and multi-class cancers using Naive–Bayes classifier to gain better accuracy. Korfiatis et al. [7] proposed the diagnosis method based on new wrapper feature selection algorithmic scheme for the primary and secondary polycythaemia. The novel wrapper feature selection algorithm is used from the local maximisation which acts as an initial dataset. Further, it refines and produces the subset which is applied for classification to achieve better accuracy. Narendar Reddy and Ravi [8] proposed two classification techniques, one is based on kernel principal component analysis and another based on quartile regression for effective classification. Chuang et al. [9] have proposed the K-nearest neighbour (K-NN) classification with the lowest error rate by applying the reduced features. This is carried out through the hybrid method of feature selection by involving correlation-based feature selection and the Taguchi chaotic binary particle swarm optimisation. Baitharua and Panib [10] proposed an approach for discovering the hidden patterns using the decision algorithm J48 to classify the diseases in a different scenario with better accuracy and lesser computation time in the early stage. Kumar and Rath [11] proposed a method for classifying the microarray dataset by reducing the features using the t-test as a feature selection method. The reduced feature is classified with KNN and fuzzy KNN algorithms. Comparing the performance measures, the fuzzy K-NN model entailed a better model than the K-NN. Zhou and Dickerson [12] proposed a novel class-dependent feature selection technique to select the desirable features. Support vector machine (SVM) combined with fuzzy KNN helps to classify the cancer using class-dependent features. Qu et al. [13] presented the FRNN classifiers and its merits. This paper highlights the accuracy of the FRNN based on the similarity between the objects in the training and test dataset. Wei et al. [14] proposed a rough set theory based classification of different types of cancers by analysing implicit hypercuboids. Maa and Xia [15] have proposed a method where the whole dataset is divided into multiple tribes and generated an optimal subset to produce the accurate pattern classification. Jensen and Cornelis [16] proposed the nearest neighbour (NN) algorithm using both lower and upper approximations from the fuzzy rough set to classify the uncertainty in both fuzzy and rough set. While comparing the other NN classifiers it outperforms the other classification in terms of results. Based on the related works discussed, the feature selection plays a major role in the data classification. The existing techniques also perform feature selection, even then few limitations occur as follows: the classification cannot provide a significant computation time when a different variety of dataset is used, the highly correlated features which are selected for classification results fail to bring about the important feature for classification, as it produces a higher error rate with a limited classification accuracy [5-8]. The above said limitations are addressed in this Letter by using metaheuristic search for probing different characteristic dataset. The function used for subset evaluation addresses inadequate features to reduce the error rate and to obtain the stopping condition for subset search. The contributions of this Letter are The proposed Letter is structured as follows: Section 2 presents the synopsis of rough set theory, fuzzy rough set, and evolutionary sequential genetic algorithm, FRNN algorithm, dataset description and evaluation parameters. Section 3 talks about the proposed method. Section 4 interprets the result and discussions. Finally, the conclusion part of this proposed method is presented in Section 5. Evolutionary sequential genetic search technique and fuzzy rough set are hybridised in this Letter to devise the genetic search fuzzy rough (GSFR) feature selection algorithm. Considering both positive and boundary regions of the fuzzy rough set, modified dependency function is recommended. The devised modified dependency function is applied as a fitness function in the proposed GSFR feature selection algorithm.

Theoretical background

Rough set theory

The rough set theory is a tool used to determine the data dependencies and to reduce the dimensions of the dataset [14]. The dataset with discrete attribute values is used to find out the subset of attributes with minimum loss of information. Let I be the Information system it is represented as where U is the non-empty set of universe and A is the non-empty finite set of attributes such that a: U→Va for every a∈A. Va set of values for the attribute a. The attribute set is separated into two non-empty disjoint subsets C and D, where C is the conditional set and D is the decision set. The indiscernibility relation IND (P) with P⊆A [17], the equivalence relation is defined as follows: where P is the subset of attribute set, x and y are objects. The set of equivalence classes is denoted by U/IND (P). The information in P is approximated by constructing lower and upper approximations of P [14] The positive region can be defined as where P and Q are the equivalence relations on U. Once the positive region is derived, the degree of dependency can be obtained The attributes are reduced by the equivalence relations generated by the set of attributes. The prediction of the reduced attribute remains the same.

Fuzzy rough set

It denotes the generalisation of rough set and which is the approximation of a fuzzy set over the crisp dataset [9]. Fuzzy rough set operates on the real valued attributes, whereas the rough set operates on discrete values only. The fuzziness is integrated into the rough set so that the membership value of fuzzy set removes the roughness of the boundary region, which ranges from 0 to 1. The fuzzy rough set approach operates on the real valued attribute in order to exercise the fuzziness as vagueness exists in the real valued dataset. The fuzzy sets F = {F1, F2, …, F} which is expressed from an equivalence class. F is a fuzzy set where i∈{1, 2, …, n}, is the membership where x belongs to the fuzzy set X executed on the universe U. The lower and upper approximations are given as The domain F is defined as where is the positive region and U is a non-empty finite set of object. The boundary region of F is defined as where is the boundary region, is the lower approximation, is the upper approximation.

Evolutionary sequential genetic search algorithm

A genetic algorithm is a random evolutionary search technique used to find out vector (string). The evolutionary sequential genetic algorithm generates the new population of string from the existing population. The generated string is an encoded version of a tentative solution. The evaluation function is used as the fitness measure in every string by indicating its suitability to the problem. In order to generate new stings, the genetic algorithm applies genetic operators such as selection, crossover and mutation on random population [15].

FRNN classifier

It captures the fuzzy and rough uncertainties in the dataset [16]. The uncertainty of fuzzy and rough set is due to intersecting classes and inadequate features. As labels of vagueness and approximation are present, fuzziness in the neighbour's closeness and the fuzzy rough set to form the FRNN. The advantages of using FRNN include insignificance of optimal value, better classification ability, non-dependence on prior structural details on the training data information, and better robust classification result.

Dataset description

The breast cancer (breast), diffuse large B cell lymphoma (DLBCL), small round blue-cell tumours (SRBCT), Leukemia datasets are used in the proposed system are extracted from the works [14]. Electroencephalography (EEG) and Gisette dataset are extracted from http://epileptologiebonn.de/cms/front_content.php?idcat=193&lang=3 and UCI machine learning repository also classified using the proposed system like cancer dataset. These datasets are used to measure the effectiveness of the classification model. The description of the dataset is shown in Table 1.
Table 1

Description of the datasets

BreastDLBCLSRBCTLeukaemiaEEGGisette
features92174027230912,58340975000
instances5458635750013,500
classes564352
Description of the datasets

Evaluation parameters

The performance of the selected features is evaluated based on classification accuracy, sensitivity, specificity, precision and F-measure parameters. Classification accuracy is measured based on the proportion of number of correctly classified instances against the total number of instances. Sensitivity is gauged based on the proportion of number of correct positive predictions against the total number of positives, which is otherwise known as recall or true positive rate. Specificity is looked at based on the proportion of number of correct negative predictions against the total number of negatives. Precision is arrived at based on the proportion of number of positive predictions against the total number of positives. F-measure is the harmonic mean of precision and recall.

Proposed model

The proposed classification model is conceived with the reduced feature is described in Fig. 1. The cancer datasets are collected from the repository and the data reduction is performed to select the contributing attribute for the classification [18, 19]. To support the functionality, a subset of features are selected by the proposed GSFR feature selection algorithm, which hybridises the property of evolutionary sequential genetic search technique and fuzzy rough set. Thereafter the reduced dataset model is generated by applying the FRNN classification. The model generated from the reduced features from the dataset foresees the cancer data with higher classification accuracy than the complete attributes without reduction [15].
Fig. 1

Block diagram denotes the classification model involving the feature selection by GSFR algorithm using genetic search technique and fuzzy rough set

Block diagram denotes the classification model involving the feature selection by GSFR algorithm using genetic search technique and fuzzy rough set Fuzzy rough feature selection identifies the appropriate set of features by eliminating the irrelevant features to improve the performance of the classifier [5]. The attribute subset is generated by the evolutionary sequential genetic search algorithm based on crossover and mutation probabilities. The steps involved in the generation and evaluation of subset are as follows: The initial subset for evaluation is randomly generated from the dataset and the generated subset is evaluated by the modified dependency function in the fuzzy rough set. The modified dependency function depends on the fuzzy rough degree of dependency by taking into account the positive region value and boundary region value (see Fig. 2).
Fig. 2

Algorithm: proposed GSFR algorithm

Randomly select the population (subset) from the dataset. Fitness function is constructed and evaluated to find the fitness of the population. Parent chromosome (features) is selected based on the precise value. Crossover probability of 0.80 is applied to crossover the parent chromosomes to generate the offspring (generate new subset). Mutation probability of 0.05 is to create the best population (subset). Fitness function is applied until it reaches the final shape. Algorithm: proposed GSFR algorithm The fuzzy rough dependencies of the attributes in (9) and (10) act as the fitness function in a genetic search algorithm to evaluate the subset of attribute. The evaluated subset is considered as the final one when it reaches the stopping criteria or else the crossover and mutation probabilities are applied to generate a new subset and evaluated with the fitness function. The same process goes on until it reaches the stopping criterion. FRNN classification is the combination of fuzzy rough approximation and fuzzy nearest neighbour (FNN) algorithm [5, 13]. If the lower approximation is high, almost all y objects neighbour belongs to the class. If the upper approximation is high, at least one y objects neighbour belongs to the class. The membership of test data objects is based on the average of upper and lower approximations [16]. Then the average is compared with the higher existing value (τ). If the average value is higher, it needs to be replaced with an existing value. If it is otherwise, the process should continue (see Fig. 3).
Fig. 3

Algorithm: FRNN algorithm

Algorithm: FRNN algorithm The FRNN generate the model by including the vague quantifiers in the definition of upper and lower approximations. Then it is applied to non-exhausting 10-fold cross-validation where the complete dataset is involved where the model is tested in the training phase. In 10-fold cross-validation, the dataset is split into 10 equal sized samples i.e. k = 1, …, 10. Ten rounds of validation are performed in the first round, wherein one subset acts as test set and the remaining nine subsets are involved in model generation. In the next round, the subset involved in testing moves to training and one subset in training set involves in testing [10]. Similarly, remaining rounds of validation are performed, where all the subsets are involved in training and testing. The average of ten rounds of validation entails the final result.

Results and discussions

In order to analyse the performance of proposed GSFR feature selection algorithm the experiments are conducted using Weka tool with system configuration of Inte1 core i5-4590 processor, 3.00 GHz and 4 GB RAM.

Percentage of feature selection

In order to compare the number of features selected in the proposed GSFR feature selection algorithm with existing feature selection algorithms, experiments are conducted and results are presented in Table 2.
Table 2

Datasets with reduced features considering different feature selection techniques

DatasetTotal number of featuresNumber of features selected with proposed and other feature selection techniques (% of features selected)
CFS [6]CSE [5]PCA [8]WSE [7]Proposed GSFR
breast92172677 (29%)1928 (21%)1458 (16%)1337 (15%)1329 (14%)
DLBCL40271014 (25%)872 (22%)543 (13%)313 (8%)306 (8%)
SRBCT2309254 (11%)167 (7%)96 (4%)53 (2%)47 (2%)
Leukaemia12,5831239 (10%)767 (6%)593 (5%)193 (2%)190 (2%)
EEG40971075 (26%)912 (22%)527 (13%)298 (7%)155 (4%)
Gisette50001202 (24%)923 (18%)519 (10%)392 (8%)65 (1%)
Datasets with reduced features considering different feature selection techniques The number and percentage of features selected using various feature selection algorithms namely CFS [6], consistency subset evaluation-based feature selection (CSE) [5], principal component analysis-based feature selection (PCA) [8], Wrapper subset evaluation-based feature selection (WSE) [7] and the proposed GSFR feature selection algorithm for the different dataset are presented in Table 2. As the SRBCT and Leukaemia datasets have more redundant features, the feature selection techniques produce less features, when compared to other datasets. The difference in a number of features selected by the proposed GSFR feature selection algorithm and the average number of features selected by the existing algorithms, which read breast at 6%, DLBCL at 9%, SRBCT and Leukaemia at 4%, EEG at 13% and Gisette at 14%. This indicates that the proposed GSFR feature selection algorithm selects less number of features compared to other algorithms. This is due to the fact that the proposed GSFR feature selection algorithm deploys metaheuristic search property of genetic algorithm and applies modified dependency function in the fuzzy rough set.

Identification of suitable classifier

In order to identify the suitable classifier related experiments are conducted with proposed GSFR feature selection algorithm, without applying the proposed algorithm by employing various classifiers namely nearest neighbour (NN) [9], J48 [10], FNN [11], Random forest (RandF) [20], Gradient Booting (GradBoost) [21], SVM [12] and FRNN. The outcome, in terms of classification accuracy and computation time, is presented in Table 3, wherein it is quite evident that in FRNN classifier, the classification accuracy is improved upon with lesser computation time (A). For instance, the difference between the classification accuracy of FRNN classifier and the average of other existing classifier for DLBCL dataset works out to 13.57% and the computation time differs by 0.05 s. Similarly, it is apparent that the FRNN produces improved classification accuracy with lesser computation time against other datasets. This is due to the fact that the modified dependency function of the fuzzy rough set is adopted as a fitness function in the proposed GSFR feature selection algorithm.
Table 3

Classification accuracy and computation time for different classifier with proposed GSFR feature selection algorithm A and without using proposed GSFR feature selection algorithm B

DatasetAccuracy, % and time, sNN [9]J48 [10]FNN [11]RandF [20]GradBoost [21]SVM [12]FRNN [13]
BABABABABABABA
breastacc.60.3766.6770.3773.1966.6770.3762.9672.2273.7873.9073.5474.0474.0274.07
time0.370.101.300.070.250.060.340.184.880.651.340.060.260.04
DLBCLacc.72.7677.5968.8072.4080.7682.7575.8681.0387.9389.6690.0094.4593.1096.55
time0.120.080.440.060.120.040.090.012.440.180.120.050.110.02
SRBCTacc.66.1974.6072.5474.6067.3076.1972.5475.7175.7178.6578.8979.4179.1580.95
time0.080.040.180.100.060.010.060.011.380.020.060.040.010.01
Leukaemiaacc.67.7273.6861.1965.7856.7457.9073.6875.4470.7081.2372.4882.9870.1787.71
time0.370.120.710.120.360.050.840.023.770.060.440.010.540.05
EEGacc.60.0861.8074.2088.8073.2089.8075.8089.8978.0084.2378.2087.2078.8090.80
time1.180.051.280.361.750.051.410.121.170.181.420.391.820.05
Gisetteacc.69.0876.0370.0278.7670.9479.0972.5080.0073.2782.6075.0087.8075.2590.60
time1.340.121.460.191.120.090.900.051.020.060.500.040.240.01
Classification accuracy and computation time for different classifier with proposed GSFR feature selection algorithm A and without using proposed GSFR feature selection algorithm B Since the FRNN classifier outperforms all other classifiers, the performance analysis in terms of sensitivity, specificity, precision, recall, F-measure and accuracy for the existing feature selection algorithm and proposed GSFR feature selection algorithm is analysed thereon by using FRNN classifier. The results are shown in Table 4.
Table 4

Performance analysis of FRNN with different feature selection techniques

Feature selection techniquesDatasetSensitivitySpecificityPrecisionF-measureAccuracy
CFS [6]Breast0.6530.8890.7090.66066.67
DLBCL0.8250.9480.8630.82977.59
SRBCT0.7090.8840.7110.70971.43
Leukaemia0.7350.8480.7210.72673.68
EEG0.8320.9520.8190.81683.2
Gisette0.8180.7860.8130.79681.82
CSE [5]Breast0.6870.8920.6750.66668.52
DLBCL0.8460.9520.8750.84679.31
SRBCT0.7320.8920.7280.72673.02
Leukaemia0.7470.8610.7430.74375.44
EEG0.850.9570.8440.83885
Gisette0.8360.8020.8340.81483.6
PCA [8]Breast0.6850.9050.7930.70670.37
DLBCL0.8560.9560.8840.85981.03
SRBCT0.7470.9010.7490.74474.60
Leukaemia0.7610.8720.7610.75877.19
EEG0.8660.9620.8650.85486.6
Gisette0.8430.8220.8470.83084.26
WSE [7]Breast0.7390.9120.7960.72872.22
DLBCL0.8780.960.8770.87382.76
SRBCT0.7670.9070.7880.76076.19
Leukaemia0.8560.9230.8640.85877.78
EEG0.8820.9670.8770.86988.2
Gisette0.8710.8460.8780.85787
proposed GSFRBreast0.7410.9480.8260.78174.07
DLBCL0.9420.9850.9040.92396.55
SRBCT0.8490.940.8410.84580.95
Leukaemia0.7050.8560.7460.72587.71
EEG0.9080.9770.9160.90890.8
Gisette0.9060.850.910.90490.6
Performance analysis of FRNN with different feature selection techniques

Performance comparison

The selected features from the proposed GSFR feature selection algorithm achieve improved classification accuracy with lesser computation time which aligns with FRNN classifier. Thereafter, it is applied to measure the performance of the existing and proposed feature selection algorithm as shown in Table 4. To support this claim, the difference between classification accuracy of the proposed GSFR feature selection algorithm and the average of other existing feature selection algorithm for the SRBCT dataset produces the improved accuracy of 7.14%. It implies that the proposed GSFR feature selection algorithm selects the contributing features for classification.

Conclusion

This Letter prescribes the cancer classification model developed from FRNN classifier using the reduced feature from proposed GSFR feature selection algorithm based on the hybridisation of evolutionary sequential genetic search algorithm and fuzzy rough set. The process aims to reduce the number of features from the datasets and thereafter a model is developed and validated by 10-fold cross-validation technique. The proposed GSFR feature selection algorithm and classification model offers proven results compared to other techniques in terms of a number of features, classification accuracy, precision, recall, F-measure and computation time. As it applies lesser features, the time consumed to develop the proposed model is minimal. The proposed GSFR feature selection algorithm selects ideal features which result in improved accuracy with lesser computational time. This Letter opens new areas of further study with specific reference to hybridise fuzzy rough set with both particle swarm optimisation and ant colony optimisation to improve upon still further.

Funding and declaration of interests

None declared.
  5 in total

1.  A novel class dependent feature selection method for cancer biomarker discovery.

Authors:  Wengang Zhou; Julie A Dickerson
Journal:  Comput Biol Med       Date:  2014-02-06       Impact factor: 4.589

2.  Predicting adherence of patients with HF through machine learning techniques.

Authors:  Georgia Spiridon Karanasiou; Evanthia Eleftherios Tripoliti; Theofilos Grigorios Papadopoulos; Fanis Georgios Kalatzis; Yorgos Goletsis; Katerina Kyriakos Naka; Aris Bechlioulis; Abdelhamid Errachid; Dimitrios Ioannis Fotiadis
Journal:  Healthc Technol Lett       Date:  2016-09-27

3.  A random version of principal component analysis in data clustering.

Authors:  Luigi Leonardo Palese
Journal:  Comput Biol Chem       Date:  2018-02-02       Impact factor: 2.877

4.  Symptom severity classification with gradient tree boosting.

Authors:  Yang Liu; Yu Gu; John Chu Nguyen; Haodan Li; Jiawei Zhang; Yuan Gao; Yang Huang
Journal:  J Biomed Inform       Date:  2017-05-22       Impact factor: 6.317

5.  A classification system based on a new wrapper feature selection algorithm for the diagnosis of primary and secondary polycythemia.

Authors:  Vasileios Ch Korfiatis; Pantelis A Asvestas; Konstantinos K Delibasis; George K Matsopoulos
Journal:  Comput Biol Med       Date:  2013-09-28       Impact factor: 4.589

  5 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.