Literature DB >> 34373776

A Holistic Performance Comparison for Lung Cancer Classification Using Swarm Intelligence Techniques.

Sunil Kumar Prabhakar1, Harikumar Rajaguru2, Dong-Ok Won3.   

Abstract

In the field of bioinformatics, feature selection in classification of cancer is a primary area of research and utilized to select the most informative genes from thousands of genes in the microarray. Microarray data is generally noisy, is highly redundant, and has an extremely asymmetric dimensionality, as the majority of the genes present here are believed to be uninformative. The paper adopts a methodology of classification of high dimensional lung cancer microarray data utilizing feature selection and optimization techniques. The methodology is divided into two stages; firstly, the ranking of each gene is done based on the standard gene selection techniques like Information Gain, Relief-F test, Chi-square statistic, and T-statistic test. As a result, the gathering of top scored genes is assimilated, and a new feature subset is obtained. In the second stage, the new feature subset is further optimized by using swarm intelligence techniques like Grasshopper Optimization (GO), Moth Flame Optimization (MFO), Bacterial Foraging Optimization (BFO), Krill Herd Optimization (KHO), and Artificial Fish Swarm Optimization (AFSO), and finally, an optimized subset is utilized. The selected genes are used for classification, and the classifiers used here are Naïve Bayesian Classifier (NBC), Decision Trees (DT), Support Vector Machines (SVM), and K-Nearest Neighbour (KNN). The best results are shown when Relief-F test is computed with AFSO and classified with Decision Trees classifier for hundred genes, and the highest classification accuracy of 99.10% is obtained.
Copyright © 2021 Sunil Kumar Prabhakar et al.

Entities:  

Mesh:

Year:  2021        PMID: 34373776      PMCID: PMC8349254          DOI: 10.1155/2021/6680424

Source DB:  PubMed          Journal:  J Healthc Eng        ISSN: 2040-2295            Impact factor:   2.682


1. Introduction

The number of patients who are diagnosed with cancer is steadily increasing in a rapid manner [1]. With the help of biopsies, image processing techniques, and blood analysis, the diagnosis of cancer is done presently. When damaged cells are excessively accumulated in human body, it leads to cancer [2]. For every patient, the behavior of cancer differs, and by examining deeply into the origin of it, it can be well understood. The cancer originates in the cells and to every individual, the structure of the cell is quite unique. Therefore, to cure cancer permanently, there is not a single specific vaccine available [3]. Understanding the relation between the gene and its products is a contribution to the genetic approach to cancer diagnosis, so that the identification of biomarker genes for targeting drug therapies can be understood well [4]. With this approach, the effects of genes on some cell signaling pathways can be well understood [5]. The information about active levels of a gene is provided by the gene expression. For gene expression, one of the widely used measurement technique is microarray [6]. In the cancer diagnosis and cancer classification types, the gene expression values obtained by microarrays can be utilized. In many studies, the microarray datasets are employed for these purposes. For the selection of biomarker gene subsets, various feature selection algorithms are employed [7]. To this microarray dataset, statistic machine learning techniques are implemented with or without feature selection [8]. Biomarker genes are to classify cancer types, with a highest classification accuracy being identified by the biomarker genes. In recent years, a new dimension to cancer research has been encompassed by the advent of microarray technology. For the classification, analysis, diagnosis, and treatment of cancer, a proficient method has been emerged by the microarray-based gene expression data [9]. Thousands of features termed as genes are found in the microarray gene expression dataset. Such data has records or instances from a few patients only and due to this limited availability of samples in comparison to larger number of genes, it is termed as curse of dimensionality problem [10]. Due to this, (a) the training time during the classification process is increased, and (b) there is a reduction in classification accuracy [11]. Therefore, the extraction of useful information from the dataset is hindered due to these challenging issues. So, the number of genes has to be reduced, and then, the highly informative genes should be selected, so that classification accuracy is increased, and it is a significant step in the microarray data analysis [12]. Feature selection/gene selection in the microarray data classification aim is to select a small subset of features from the original huge feature space [13]. By removing redundant and irrelevant features, feature selection can be done, so that the classification accuracy is increased, and the classification time is reduced. The feature selection technique proposed in the literature includes hybrid method, embedded technique, filter, and wrapper methods [14]. In this study, the primary aim is to classify and select the optimal gene subsets for lung cancer. Then, feature selection is implemented along with optimization techniques and finally classified. Some of the prominent works in the lung cancer classification using microarray gene analysis are explained as follows. For the molecular classification of lung cancer, a cross study comparison of gene expression study was done by Parmigiani et al. [15]. Using the significance analysis of Microarray-Gene set reduction algorithm, the classification of non-small cell lung cancer was performed by Zhang et al. [16]. For multiclass classification of lung cancer, an adaptive multinomial regression with overlapping groups is performed by Li et al. [17]. The lung cancer prediction from microarray data by gene expression programming was done by Azzawi et al. [18]. A support vector machine-based classification method for lung cancer gene expression data base analysis was done by Guan et al. [19]. Some progresses in the techniques and integrated analysis related to the image processing techniques and the development of advanced devices for tissue engineering approach as a potential solution to treat lung diseases too have been discussed in the literature [20, 21]. As far as the microarray gene selection techniques using optimization and classification are concerned, self-organizing maps [22], ensemble classification techniques [23], Taguchi chaotic binary Particle Swarm Optimization (PSO) [24], adaptive wrapper approach combined with SVM [25], kernel based methods [26], pattern classification methods [27], Convolution Neural Networks (CNN) [28], fuzzy approaches [29], Analysis of Variance (ANOVA), and K-Nearest Neighbour (KNN) [30] were proposed in the literature. Using ant colony optimization, a hybrid gene selection approach was proposed by Sharbaf et al. [31]. For the cancer classification data on gene expression data, PSO and DT classifiers were implemented by Chen et al. [32]. For gene selections, the various techniques reported in literature are utilizing multiobjective algorithms [33], a hybrid binary Imperialist Competition Algorithm (ICA), and tabu search approach [34], a binary differential evolution algorithm [35], a simplified swarm optimization using a Social Spider Optimization (SSO) algorithm [36], Artificial Bee Colony (ABC) [37], Binary PSO [38], novel rule-based algorithm [39], and Shuffled Leap Frog Algorithm (SLFA) [40], and it has been well explored. However, in this paper, other suitable swarm intelligence techniques have been explored and analyzed comprehensively. The organization of the paper is as follows. In Section 2, the materials and methods followed by the gene selection techniques are explained. In Section 3, the optimization techniques for gene selection are explained, and in Section 4, the classification techniques are explained followed by the results and discussion in Section 5 and conclusion in Section 6.

2. Materials and Methods

For the lung cancer classification, a lung Harvard 2 dataset was utilized, which is publicly available online [41]. The dataset has 181 samples with 150 Adenocarcinoma (ADCA) and 31 Malignant Pleural Mesothelioma (MPM). The dataset is tabulated in Table 1.
Table 1

Dataset details.

DatasetNumber of genesClass 1 (ADCA)Class 2 (MPM)Total samples
Lung Harvard 21253315031181
The pictorial representation of the work is shown in Figure 1.
Figure 1

Pictorial representation of the work.

2.1. Gene Selection Techniques

The gene selection techniques utilized in this paper are Information Gain, Relief-F, Chi-square statistic, and T-statistic. The discretization of the attribute values is done before using chi-square, information gain, and other feature selection methods. The main intention of utilizing the gene selection techniques is to select the most important genes from 12,533 genes. Here, in our work, we have selected 1000 important genes after the gene selection process through the following techniques.

2.1.1. Information Gain

It is used generally as an attribute selection criteria while dealing with decision trees; hence, it is used as a gene selection technique too [7]. Assume the class set S={S}, where x=1,2,…, l. For every feature Y, the Information Gain is expressed aswhere H(S)=−∑p(s)log2  p(s) and Only for discrete features, Information Gain is used widely, and therefore prior to computing Information Gain, the discretization of numeric features should be done. Depending on the large values of information gain, the selection of features are done.

2.1.2. Relief–F

For dealing with multiclass, noisy, and incomplete datasets, Relief-F is introduced, and it is an extension of Relief algorithm [7]. To each feature, a relevance weight is assigned. The selection of a random sample instance I is done from n sample instances. Based on the basic differences between the selected instance I and its neighboring instance represented as Q and termed as hit and different class termed as nearest miss represented by N(S), the updating of the relevance features is done. The features that discriminate the instance from various neighbors of the surrounding classes are given more weight. By analyzing the average contribution of neighboring nearest misses N(S), the updating of the weights is done. The prior probability of each class is considered by the average contribution. The updating of the weight of jth feature Y is as follows:where the distance between sample instances (I) and the nearest hit (Q) or nearest miss N(S) is calculated by the function Ψ(YI, Q).

2.1.3. Chi-Square Statistic

With respect to the classes, for each feature, the value of χ2 statistic is computed [7]. Before computing χ2 statistic, the discretization of the numeric attributes is done. For every feature, Y, χ2 statistic is computed aswhere n( represent the number of samples in Y for class s whose value is y. The definition of expected frequency is expressed aswhere the number of samples in Y with value y is denoted by n; n indicates the number of samples of class s. The total number of samples is expressed by n. Based on the sorted value of χ2 statistic, the selection of features is done.

2.1.4. T-Statistic

This is a famous gene selection technique and quite popular in two-class problems [7]. Every sample can be classified into either class S1 or class S2. For every feature Y, the computation of t-statistic is expressed aswhere μ indicates the mean of the jth feature for class S. The “k” indicates the class index, i.e., k=1 or k=2. Once the t-statistic value for each feature is computed, then it is sorted out in a descending order, so that the important features can be selected.

3. Optimization Techniques

The shortlisted 1000 genes will undergo again a secondary feature selection methodology to select the best 50 genes, 100 genes, and 200 genes by means of utilization of optimization techniques. The second level feature selection is done using the five optimization algorithms as follows.

3.1. Grasshopper Optimization Algorithm

In many engineering optimization problems, this algorithm is widely used. Based on the biggest swarms of all creatures, one of the recently proposed naturally inspired algorithms is GO [42]. Severe damage to the crops is caused by the herbivores grasshopper. The grasshopper has a swarming behavior, and it depends on both adults and nymphs. Soft plants and succulents are fed by the nymph, which rolls on the ground continuously. In search of food, the adult grasshopper can jump to a very high extent, and so, it will have a very large area to explore. The observation of both types of movement such as slow movement and abrupt movement has been achieved, which indicates that exploitation and exploration are possible. For the grasshopper, the swarming behavior is represented mathematically aswhere Q represents the position of the jth grasshopper, A represents the social interaction, F represents the gravity force in the jth grasshopper, and B represents the wind advection. The representation of social interaction A is given aswhere d=|q − q| represents the distance between the jth and kth grasshopper and represents a unit vector from the jth grasshopper to the kth grasshopper. The social forces are expressed by the function “a” and are expressed aswhere the intensity of the attraction is represented as g and the attractive length scale is expressed by l. In terms of social interaction, three types of regions are created by the grasshoppers in search of food, that is, attraction zone, comfort zone, and repulsion region. Strong forces cannot be applied by the function “a” when the distance is large between grasshoppers. To resolve this, the F component in (7) is expressed aswhere the f represents the gravitational constant and indicates a unity vector progressing towards the Earth center. The computation of B component is as follows:where v represents the constant drift and represents a unity vector in the wind direction. If the values of a, F and B are substituted in (7), thenwhere a(s) is given by (9) and the number of grasshoppers is represented by N. To solve optimization problems, a revised version of this formula is used aswhere v represents the upper bound and l represents the lower bound in the Dth dimension. In the target, the value of the Dth dimension is represented by . To shrink the three worms, the decreasing coefficient is “c.” Only towards a target the wind direction is progressed always. While the food is a searched form, adults start jumping in the air, and nymphs move on rolling in the ground creating both cases of exploration and exploitation. By reducing the parameter c in the below equation, one can balance both these two in proportion to the total number of iterations. Its computation is done aswhere the maximum value is represented as cmax, minimum value is represented as cmin, i denotes the current iteration, and I represents the maximum number of iterations.

3.2. Moth Flame Optimization Algorithm

Based on the simulation of moth behavior for this special movement method during nighttime, Moth flame optimization algorithm was developed [43]. For the purpose of navigation or movement, a mechanism termed as transverse orientation is utilized. By maintaining a standard angle with reference to the moon, moth flies, which is a very effective methodology for travelling by distances in a straight path, as the distance between the moon and the moth is very far away. This kind of methodology is adopted, so that moth flies along a very straight path at nighttime. It is a general observation that the moths fly around the lights in a spiral manner. The artificial lights can easily trick the moths to exhibit such behavior. As the light lies with close proximity to the moon, a spiral fly of moths is caused due to the maintenance of a similar angle to the light source. In this algorithm, the representation of the set of moths is done as a matrix A. For the storage of all the corresponding fitness values, there is an array OA for all the moths. In this algorithm, the second key component is the flames. Now, again, a matrix B similar to the moth matrix is considered. For the storage of all the corresponding fitness values, there is an array OB for all the flames. The global optimal of the optimization problem is approximated by the MFO algorithm by a three-tuple process as follows: A random population of moths with its corresponding fitness value is denoted by a function C. In this function, the methodical model is expressed as The movement of the moths around the search space is determined by the function D which is the primary function. The matrix of A is received by this function and eventually returns the updated one as The K function remains true if termination criterion is satisfied and false if the termination criterion is not satisfied.With C, D, and K, the general framework of the MFO is expressed as follows: A=C()( ); while K(A) is equal to false A=D(A); end Until the K function returns true, the D function is run iteratively after the initialization. To simulate the moth behavior mathematically, the updating of the position of every moth is updated with respect to a flame using the following equation:where A indicates the cth moth, B specifies the gth flame, and F represents the spiral function. Subject to the following conditions, the utilization of any type of spiral can be done using the three conditions as follows: The initial point of the spiral should begin from the moth The final point of the spiral should be the flame position The range of spiral fluctuation should not exceed the search space For the MFO algorithm, the logarithmic spiral is defined aswhere J specifies the distance of the cth moth for the gth flame, h denotes a constant for defining the shape of a logarithmic spiral, and k is a random number in [−1, 1]. The computation of J is done as follows:where A indicates the cth moth, B specifies the gth flame, and J specifies the distance of the cth moth for the gth flame. The spiral flying path of the moth is expressed by (20). From this equation, with respect to a flame, the next position of a moth is explained. In the spiral equation, the k parameters defer the next position of the moth with reference to its proximity or closeness to the flame. While the position is updated, it only regains a moth to progress towards a flame; thereby, it may be trapped in local optima fastly. Each moth is obliged to update its position using only one of the flames to prevent such situations. The position updating of moths with respect to “n” various locations in the search space may sometimes denote the exploitation of most promising solutions.where I denotes the current number of iterations, N denotes the maximum number of flames, and K specifies the maximum number of iterations. To balance the exploration and exploitation of the search space, there is a gradual decrease in the number of flames. The general steps of the D function are described in Algorithm 1.
Algorithm 1

D function execution and termination.

As projected in the algorithm, unless the K function returns true, the D function is executed. Once the D function is terminated, the best moth returns, as it is shown as the best attained optimum approximation value.

3.3. Bacterial Foraging Optimization Algorithm

The three main mechanisms are present in the classical BFO, that is, chemotaxis process, reproduction process, and elimination-dispersal process [44].

3.3.1. Chemotaxis Process

Here, a tumble indicates a unit walk with random direction, and a run indicates a unit walk with the similar direction in the last step. Assuming θ(b, c, d) indicates the bacterium at bth chemotactic, cth reproductive, and dth elimination-dispersal method. R(a) is considered as the run-length unit parameter is the chemotactic step size during every tumble or run. The movement of the ath bacterium in every computational chemotactic step is expressed aswhere Δ(a) represents the direction vector of the bth chemotactic step. Δ(a) is the same as the final chemotactic step if the bacterial movement is run; or else Δ(a) becomes a random vector, where specific elements lie in the range of [−1, 1]. A step fitness indicated as B(a, b, c, d) is evaluated with the activity of both run or tumble assumed and considered at each step during the chemotaxis process.

3.3.2. Reproduction Process

During its lifetime, the sum of the step fitness is calculated as the health status of each bacterium as ∑B(a, b, c, d), where N represents the maximum step in a chemotaxis process. Based on the health status, the sorting of the bacteria is done in a reverse order. Only the first half of population lives/survives in the reproductive step. The living bacterium divides into two identical ones, and they are kept in the same places, and so the population of bacteria keeps constant.

3.3.3. Elimination and Dispersal Process

A basis for local search is provided by the chemotaxis, and the convergence is sped up by the reproduction process. Using the classical BFO, this situation has been simulated to a large extent. For searching of global optima, only chemotaxis and reproduction are not enough. Around the local optima, the bacteria may get stuck and to eliminate the accidents of being trapped into local optima easily and gradually, the diversity of the BFO changes. Only after a certain number of reproductive processes, the dispersion event happens. Then, based on a probability P, some bacteria are chosen to be killed and shifted to another position within a particular environment. The step by step procedure is explained in Algorithm 2.
Algorithm 2

BFO.

3.4. Krill Herd Optimization Algorithm

Based on the simulation of the herding of krill swarms, a famous metaheuristic algorithm for solving optimal problems is KH optimization algorithm [45]. The herding of the skill swarms is usually in response to a certain environmental and biological process. In a 2D space, the time-dependent position of an individual krill is decided by 3 primary actions, that is, Movement which influences or influenced by other krill individuals. Foraging actions Random diffusion In a d-dimensional decision space, the following Lagrangian model is adopted by the KH algorithm aswhere M is the motion led by other krill individuals, G is the foraging motion, and D is the physical diffusion of the jth krill individual. The krill individuals affecting the other movement are represented by the direction of motion α and it is computed by the target swarm density, a local swarm density, and a repulsive swarm density. The movement for a krill individual is defined as follows: The maximum induced speed is represented by Mmax, the inertia weight of the motion induced in [0, 1] is represented as v and the latest motion induced is represented by Mold. With the help of two main components, the estimation of foraging motion is done. The first one is the food location, and the second one is the basic knowledge about the food location. The motion is approximately formulated for the jth krill individual as follows:wherewhere the foraging speed is represented by W. The inertia weight of the foraging motion between 0 and 1 is represented as v and Gold is the last foraging action. A random process is modelled to the random diffusion of the krill individuals. In terms of both a random directional vector and maximum diffusion speed, the description of the motion can be done. It is represented as follows:where the maximum diffusion speed is Dmax, the random directional vector is δ and its arrays are random values in the range of [−1, 1]. Utilizing various motion parameters during the time and based on the above-mentioned movements, using the following equation, the position vector of a krill individual from the time interval t to t+Δt is given asΔt is regarded as the most important term and based on the specific type of optimization, the parameters can be fine-tuned. The scalar factor of the speed vector is assumed because of Δt parameter.

3.5. Artificial Fish Swarm Optimization Algorithm

It is a famous Swarm Intelligence technique, which is helping to solve the optimization problem by utilizing the behavior of artificial fishes like imitating swarming process, chasing process, and preying behaviors [46]. Assume A is the current position of one artificial fish and A is the viewpoint of artificial fish at one specific moment. The visual scope of every individual is expressed as Vis; therefore, within Vis of A be the fishes A and A. The largest step of artificial fish is assumed as step and the congestion factor of the fish swarm is expressed as δ. The food concentration factor is highly proportional to the fitness function f(A). In the fish swarm, the behavior patterns are expressed as follows:

3.5.1. Swarming Behavior

If f(A) > f(A), then A is the central point inside the Vis of the point A and so the execution of swarming behavior is done easily. Assume A as A and so, the fish at A will progress towards the point A.

3.5.2. Chasing Behavior

The point (expressed by Amax), which has the best objective function value, is present inside the visual satisfying the criterion f(Amax) > f(A) and if there is less crowd in the visual of A , then the execution of chasing behavior is done. Consider Amax as A and so, the fish at Awill progress towards the point Amax.

3.5.3. Preying Behavior

Under the following situations, preying behavior is tried. f(A) > f(A), f(Amax) < f(A) and there is less or no crowd in the Vis Alternatively, if the visual is crowded, then the random selection of a point A inside the visual of A is done. The preying behavior is executed if f(A) > f(A). Assume A as A and so, the fish at A will progress towards the point A. Otherwise, with its visual limit, it will move a step in a random fashion. In each iteration, the best solution obtained is termed as “board.” The search process can be terminated after the specified iterations, and the result present on the board is considered as the final solution. The position updating for the artificial preying fishes is formulated aswhere the next position of artificial fish is termed as Anext. The current position of the artificial fish is expressed as A and the position having a better objective function value is A. The random number is expressed as rand and it is in the range of [−1, 1]. Between the two position vectors, the distance is expressed as norm(A − A). The position updating for the artificial swarm fishes can be done as The position updating for the artificial chasing fishes is formulated as

4. Classification Procedures

The optimized values or the best gene values obtained after the second level optimization techniques are classified using NBC, Decision trees, SVM, and KNN algorithms. The Performance Analysis of Classifiers in terms of Classification accuracies with GO, MHO, BFO, KHO, and AFSO for different gene selection techniques using 50–200 selected genes is done here.

4.1. Naïve Bayesian Classifier

It is a famous probabilistic algorithm, where, given the class, the feature values based on Bayes rule are conditionally independent [47]. If a new sample observation is given, the assignment of the classifier to the class having the maximum conditional probability estimate is done.

4.2. Decision Tree

A famous rule-based classifier is DT, where leaf nodes represent classification outcomes and non-leaf nodes represent selected attributes [48]. A classification rule is reflected by the path from the root to a leaf node. The J4.8 algorithm is used here.

4.3. Support Vector Machine

The SVM analyzes the input data as two unique sets of vector in a p-dimensional space initially [49]. Then, in that space, a separate hyperplane is constructed, so that the margin is maximized between the two data sets. The SVM is utilized with SVM Polynomial kernels for training purposes.

4.4. KNN Algorithm

One of the famous instance-based classifiers is KNN [50]. The class label of a new testing sample is decided by the classifier. It is done by considering the majority of classes of its K closest neighbors dependent on their Euclidean distance. Here, the value of K is assigned to be 4.

5. Results and Discussion

It is classified with a 10-fold cross-validation method, and its performance is shown in the tables. The mathematical formulae for computing the Performance Index (PI), Sensitivity, Specificity, and Accuracy are mentioned in literature and using the same, the values are computed and exhibited. PC is Perfect Classification; MC is Missed Classification; and FA is False Alarm in the following expressions. The sensitivity is expressed as Specificity is expressed as Accuracy is expressed as Performance Index (PI) is expressed as Table 2 shows the performance analysis of classifiers for classification accuracy parameter with GO method for different gene selection techniques. As indicated in Table 2 that SVM classifier with 100 selected genes in Relief F test method and NBC with information gain method for 100 genes attained higher accuracy of 98.96%. The lower accuracy of 76% is thrown out by KNN classifier in all three statistical methods.
Table 2

Performance analysis of classifiers in terms of classification accuracies (%) with Grasshopper optimization for different gene selection techniques using 50–200 selected genes.

MethodNBCDTSVMKNN
Genes selected50100200501002005010020050100200
Information gain82.2998.967678.1283.5983.597683.5977.08767677.08
Relief–F test85.9391.6783.5983.5978.1289.678.1298.9695.8387.578.1277.08
Chi-square test97.9177.0883.5995.8393.7577.0882.2983.5991.677677.0883.59
T statistic test77.0882.2982.2997.9195.8378.1283.5993.7595.8377.087685.93
Table 3 indicates the performance analysis of classifiers for classification accuracy with MFO method for different gene selection techniques. As shown in Table 3, DT classifier with 50 selected genes in Relief F test method reached higher accuracy of 98.012%. The lower accuracy of 78.125% is depicted by SVM classifier with 100 genes selected in relief F test method. The lower accuracy of SVM is due to the presence of outlier in the gene samples.
Table 3

Performance analysis of classifiers in terms of classification accuracies (%) with Moth flame optimization for different gene selection techniques using 50–200 selected genes.

MethodNBCDTSVMKNN
Genes selected50100200501002005010020050100200
Information gain96.3489.693.7585.9383.5993.2691.6797.9189.689.683.5995.83
Relief–F test93.7582.2991.6798.0197.9186.1985.9378.1295.8382.2997.9189.6
Chi-square test93.7593.7590.7491.6785.9384.1097.9186.4085.9385.9397.9191.67
T statistic test85.9391.6795.8385.9393.7584.7596.1585.9391.6785.9385.9397.91
Table 4 demonstrates the performance analysis of classifiers for classification accuracy with BFO method for different gene selection techniques. From Table 4, it is identified that DT classifier with 50 selected genes in Chi-square test method reached higher accuracy of 98.56%. The lower accuracy of 82.24% is shown by SVM classifier with 100 genes selected in information gain method. Across the gene samples, all the classifiers performed well in this BFO method.
Table 4

Performance analysis of classifiers in terms of classification accuracies (%) with Bacterial foraging optimization for different gene selection techniques using 50–200 selected genes.

MethodNBCDTSVMKNN
Genes selected50100200501002005010020050100200
Information gain97.9196.8591.6495.8385.9385.9385.0282.2489.9285.9387.1293.75
Relief–F test97.9193.7589.689.691.6783.5997.9183.8187.3095.8397.9197.91
Chi-square test93.7589.691.6798.5690.5397.3397.7285.9397.9197.9195.8395.83
T statistic test91.6793.7585.9384.4195.8398.0486.7197.9184.9595.8393.7589.6
Table 5 reveals the performance analysis of classifiers for classification accuracy with KHO method for different gene selection techniques. Table 5 shows that SVM classifier with 50 selected genes in Relief F test method reached higher accuracy of 98.38%, as the number of selected genes increased gradually and given to SVM classifiers, which reported lower accuracy of 77.47% with 200 genes selected in Relief F test method.
Table 5

Performance analysis of classifiers in terms of classification accuracies (%) with Krill Herd optimization for different gene selection techniques using 50–200 selected genes.

MethodNBCDTSVMKNN
Genes selected50100200501002005010020050100200
Information gain97.9191.6781.5786.5889.696.7691.8993.7597.9192.0591.4096.92
Relief–F test89.697.9183.5982.2985.9387.3798.3879.2177.4783.4690.9784.25
Chi-square test93.7593.2781.7297.1282.5992.1288.5578.9998.6987.8878.6188.91
T statistic test95.8397.9191.6795.8396.9395.8394.7285.0696.6786.7388.0398.96
Table 6 reports the performance analysis of classifiers for classification accuracy with AFSO method for different gene selection techniques. As indicated in Table 6, DT classifier with 100 selected genes in Relief F test method reached the highest accuracy of 99.10%. The NBC classifier is settled at the lower accuracy of 77.08% with 200 selected genes for Relief F test method.
Table 6

Performance analysis of classifiers in terms of classification accuracies (%) with Artificial fish swarm optimization for different gene selection techniques using 50–200 selected genes.

MethodNBCDTSVMKNN
Genes selected50100200501002005010020050100200
Information gain90.2491.6789.0588.4186.8281.0394.9093.7593.7593.7589.693.75
Relief–F test95.9685.9377.0889.7999.1086.4295.8397.9197.9189.678.5980.56
Chi-square test82.5983.4878.1284.1095.3494.6195.8398.6393.7594.4997.9184.39
T statistic test93.7589.7587.586.7892.2989.3389.697.9197.9195.8394.7997.91
Table 7 signifies the performance analysis of classifiers for classification PI parameter with GO method for different gene selection techniques. As shown in Table 7, SVM classifier with 100 selected genes in Relief F test method and NBC with information gain for 100 genes attained higher PI of 97.87%. The lower PI of 7.69% is indicated by KNN classifier in all three statistical methods.
Table 7

Performance analysis of classifiers in terms of PI (%) with Grasshopper optimization for different gene selection techniques using 50–200 selected genes.

MethodNBCDTSVMKNN
Genes selected50100200501002005010020050100200
Information gain45.1697.877.6922.2251.1651.167.6951.1615.367.697.6915.36
Relief–F test60.8780.0151.16351.16322.2278.9322.2297.8791.5866.6622.2215.36
Chi-square test95.6515.3651.16391.5885.715.3645.1651.1680.017.6915.3651.16
T statistic test15.3645.1645.1695.6591.5822.2251.1685.791.5815.367.6960.87
Table 8 demonstrates the performance analysis of classifiers for classification PI with MFO method for different gene selection techniques. As shown in Table 8, DT classifier with 50 selected genes in Relief F test method reached higher PI of 95.85%. The lower PI of 22.22% is indicated by SVM classifier with 100 genes selected in Relief F test method. The lower PI of SVM is due to the presence of outlier genes in the samples.
Table 8

Performance analysis of classifiers in terms of PI (%) with Moth flame optimization for different gene selection techniques using 50–200 selected genes.

MethodNBCDTSVMKNN
Genes selected50100200501002005010020050100200
Information gain92.4578.9385.760.8751.1684.3180.0195.6578.9378.9351.1691.58
Relief-F test85.745.1680.0195.8595.6561.1460.8722.2291.5845.1695.6578.93
Chi-square test85.785.777.6280.0160.8753.3595.6562.5060.8760.8795.6580.01
T statistic test60.8780.0191.5860.8785.756.0892.1760.8780.0160.8760.8795.65
Table 9 represents the performance analysis of classifiers PI with BFO method for different gene selection techniques. From Table 9, it is known that DT classifier with 50 selected genes in Chi-square test method reached higher PI of 97.009%. The lower PI of 45.09% is indicated by SVM classifier with 100 genes selected in information gain method. Across the gene samples, all the classifiers performed well in this BFO.
Table 9

Performance analysis of classifiers in terms of PI (%) with Bacterial foraging optimization for different gene selection techniques using 50–200 selected genes.

MethodNBCDTSVMKNN
Genes selected50100200501002005010020050100200
Information gain95.6593.1779.9391.5860.8760.8757.2245.0978.3160.8764.0385.7
Relief-F test95.6585.778.9378.9380.0151.1695.6552.1163.8591.5895.6595.65
Chi-square test85.778.9380.0197.0077.1194.4594.9660.8795.6595.6591.5891.58
T statistic test80.0185.760.8752.9891.5895.9263.7695.6556.0091.5885.778.93
Table 10 exposes the performance analysis of classifiers for classification PI with KHO method for different gene selection techniques. Table 10 reported that SVM classifier with 50 selected genes in Relief F test method reached higher PI of 96.68%, as the number of selected genes increased gradually and given to SVM classifiers, which reported lower PI of 17.93% with 200 genes selected in Relief-F test method.
Table 10

Performance analysis of classifiers in terms of PI (%) with Krill Herd optimization for different gene selection techniques using 50–200 selected genes.

MethodNBCDTSVMKNN
Genes selected50100200501002005010020050100200
Information gain95.6580.0139.8762.4778.9393.4480.3685.795.6581.0779.2193.28
Relief-F test78.9395.6551.1645.1660.8764.6896.6828.9417.9350.0678.0453.72
Chi-square test85.784.4142.3193.6046.5681.0072.7927.5497.2868.9624.0174.91
T statistic test91.5895.6580.0191.5893.2291.5888.4557.4292.6963.0862.8397.87
Table 11 details the performance analysis of classifiers for classification PI with AFSO method for different gene selection techniques. As indicated in Table 11, DT classifier with 100 selected genes in Relief F test method reached the highest PI of 98.16%. The NBC classifier depicts lower PI of 15.36% with 200 selected genes for the same Relief F test method.
Table 11

Performance analysis of classifiers in terms of PI (%) with Artificial fish swarm optimization for different gene selection techniques using 50–200 selected genes.

MethodNBCDTSVMKNN
Genes selected50100200501002005010020050100200
Information gain77.6780.0175.7672.0264.1237.5988.0185.785.785.778.9385.7
Relief-F test91.2660.8715.3678.5598.1662.6791.5895.6595.6578.9325.1135.42
Chi-square test46.5650.6722.2253.3790.0888.0991.5897.1485.787.695.6554.61
T statistic test85.778.6766.6665.9182.6577.278.9395.6595.6591.5888.6495.65

6. Conclusion and Future Work

One of the most prominent lethal factors for human beings nowadays is cancer. The best chances of suitable treatment can sometimes be missed due to mistaken diagnosis. The accuracy of cancer diagnosis with machine learning along with clinical tests is very helpful in the treatment of cancer. Microarray expression data is highly redundant and with respect to most number of classes, the genes present are uninformative. Therefore, it is a critical necessity to select the best feature genes for the analysis of cancer. Out of a large dataset, the techniques should be capable of identifying a subset of most informative genes in a robust manner. In this work, a comprehensive analysis of lung cancer classification with the help of feature selection and optimization techniques is done. The best results are obtained when Relief-F test is computed with AFSO and classified with Decision Trees classifier for hundred genes, and a highest classification accuracy of 99.10% is obtained. Future works aim to work with other feature selection techniques and a variety of optimization techniques classified with deep learning techniques for effective classification of lung cancer.
  15 in total

1.  Coupled two-way clustering analysis of gene microarray data.

Authors:  G Getz; E Levine; E Domany
Journal:  Proc Natl Acad Sci U S A       Date:  2000-10-24       Impact factor: 11.205

2.  SVM-T-RFE: a novel gene selection algorithm for identifying metastasis-related genes in colorectal cancer using gene expression profiles.

Authors:  Xiaobo Li; Sihua Peng; Jian Chen; Bingjian Lü; Honghe Zhang; Maode Lai
Journal:  Biochem Biophys Res Commun       Date:  2012-01-28       Impact factor: 3.575

3.  Adaptive multinomial regression with overlapping groups for multi-class classification of lung cancer.

Authors:  Juntao Li; Yanyan Wang; Xuekun Song; Huimin Xiao
Journal:  Comput Biol Med       Date:  2018-06-19       Impact factor: 4.589

4.  3D laser scanning in conjunction with surface texturing to evaluate shift and reduction of the tibiofemoral contact area after meniscectomy.

Authors:  Roberto De Santis; Antonio Gloria; Stefano Viglione; Saverio Maietta; Francesco Nappi; Luigi Ambrosio; Dante Ronca
Journal:  J Mech Behav Biomed Mater       Date:  2018-08-09

5.  A hybrid gene selection approach for microarray data classification using cellular learning automata and ant colony optimization.

Authors:  Fatemeh Vafaee Sharbaf; Sara Mosafer; Mohammad Hossein Moattar
Journal:  Genomics       Date:  2016-05-03       Impact factor: 5.736

6.  Recursive partitioning for tumor classification with gene expression microarray data.

Authors:  H Zhang; C Y Yu; B Singer; M Xiong
Journal:  Proc Natl Acad Sci U S A       Date:  2001-05-29       Impact factor: 11.205

7.  Lung cancer prediction from microarray data by gene expression programming.

Authors:  Hasseeb Azzawi; Jingyu Hou; Yong Xiang; Russul Alanni
Journal:  IET Syst Biol       Date:  2016-10       Impact factor: 1.615

8.  Lung cancer gene expression database analysis incorporating prior knowledge with support vector machine-based classification method.

Authors:  Peng Guan; Desheng Huang; Miao He; Baosen Zhou
Journal:  J Exp Clin Cancer Res       Date:  2009-07-18

9.  Hybrid Binary Imperialist Competition Algorithm and Tabu Search Approach for Feature Selection Using Gene Expression Data.

Authors:  Shuaiqun Wang; Wei Kong; Weiming Zeng; Xiaomin Hong
Journal:  Biomed Res Int       Date:  2016-08-04       Impact factor: 3.411

10.  An Amalgamated Approach to Bilevel Feature Selection Techniques Utilizing Soft Computing Methods for Classifying Colon Cancer.

Authors:  Sunil Kumar Prabhakar; Harikumar Rajaguru; Sun-Hee Kim
Journal:  Biomed Res Int       Date:  2020-10-13       Impact factor: 3.411

View more
  1 in total

1.  Effect of Ropivacaine Combined with Small Doses of Triamcinolone and Continuous Nerve Block of Unilateral Paravertebral Canal Guided by Ultrasound on Metastasis after Radical Treatment of Lung Cancer.

Authors:  Dan Xu; Wei Luo
Journal:  J Oncol       Date:  2022-06-29       Impact factor: 4.501

  1 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.