Literature DB >> 34056623

Hybrid Binary Dragonfly Algorithm with Simulated Annealing for Feature Selection.

Hamouda Chantar1, Mohammad Tubishat2, Mansour Essgaer1, Seyedali Mirjalili3,4.   

Abstract

There are various fields are affected by the growth of data dimensionality. The major problems which are resulted from high dimensionality of data including high memory requirements, high computational cost, and low machine learning classifier performance. Therefore, proper selection of relevant features from the set of available features and the removal of irrelevant features will solve these problems. Therefore, to solve the feature selection problem, an improved version of Dragonfly Algorithm (DA) is proposed by combining it with Simulated Annealing (SA), where the improved algorithm named BDA-SA. To solve the local optima problem of DA and enhance its ability in selecting the best subset of features for classification problems, Simulated Annealing (SA) was applied to the best solution found by Binary Dragonfly algorithm in attempt to improve its accuracy. A set of frequently used data sets from UCI repository was utilized to evaluate the performance of the proposed FS approach. Results show that the proposed hybrid approach, named BDA-SA, has superior performance when compared to wrapper-based FS methods including a feature selection method based on the basic version of Binary Dragonfly Algorithm. SUPPLEMENTARY INFORMATION: The online version contains supplementary material available at 10.1007/s42979-021-00687-5.
© The Author(s), under exclusive licence to Springer Nature Singapore Pte Ltd 2021.

Entities:  

Keywords:  Dragonfly algorithm; Feature selection; Optimization; Simulated annealing algorithm

Year:  2021        PMID: 34056623      PMCID: PMC8147911          DOI: 10.1007/s42979-021-00687-5

Source DB:  PubMed          Journal:  SN Comput Sci        ISSN: 2661-8907


Introduction

Recently, data mining field has become an active research area due to presence of the huge amount of data in digital format that needs to be transformed into useful information. The main task of data mining is to build models for the discovery of useful hidden patterns in collections of huge data. It is considered as an essential step in the knowledge discovery process [16]. Preprocessing of data is a critical step in data mining. It has a direct impact on data mining techniques such as classification. It affects the quality of discovered patterns and the accuracy of the classification models [16, 25]. Feature selection is one of the main pre-processing steps in data mining that aims to discard noisy and irrelevant features while retaining the useful and informative ones. Selecting the ideal or near ideal subset of given features leads to accurate classification results and less computational cost [6, 25]. Feature selection approaches are classified based on estimation criteria of selected subset of features into two classes filter and wrapper approaches [6]. Wrapper techniques heavily rely on their search for optimal subset of features on the accuracy of machine learning classifiers such as KNN or SVM, whereas filter techniques use scoring matrices such as chi-square and information gain to assess the goodness of the selected subset of features. More precisely, in filter approaches, attributes are ranked using a filter approach, e.g., chi-square, and then the attributes with less than a predefined threshold are removed [1, 6, 14]. Generally speaking, finding an optimal subset of features is a challenging task. FS has gained the interest of many researchers in data mining and machine learning fields [8, 14]. The literature shows that meta-heuristic techniques have been very effective in tackling many optimization problems like machine learning, engineering design, data mining, production problems and feature selection [38]. For feature selection problem, Genetic algorithm (GA) and Particle swarm optimization algorithm (PSO) have been successfully utilized for solving many feature selection problems [6, 7, 13, 17]. Moreover, many biologically inspired approaches such as simulated annealing (SA) [29], Tabe search (TS) [45], Ant Colony Optimization (ACO) [22], Binary Bat Algorithm [32]. Moth-Flame Optimization Algorithm [44], Antlion Optimization Algorithm [43], Dragonfly Optimization Algorithm [26], and Whale Optimization Algorithm (WOA) [35] have been efficiently applied to discover the best subsets of features for many classification problems. As discussed in [37, 38], when designing and using meta-heuristic techniques, two main different criteria must be considered: diversification which refers to search space exploration, and intensification which means the exploitation of the optimal solution (e.,g., best subset of features found so far). Based on these criteria, meta-heuristic techniques can be distinguished into two branches. Population-based meta-heuristics (e.,g., PSO) and single-solution based techniques such as SA and TS which are biased towards exploitation. The performance of the search algorithm can be improved if an appropriate balance between the exploration and the exploitation is achieved. The desired balance between exploration and exploitation can be achieved via combining two techniques, e.g., population-based approach and single-solution based approach. Various hybrid meta-heuristic approaches have been proposed for solving wide range of optimization problem. Hybrid approaches have also been popular, because they benefit from advantages of two or more algorithms [37]. In [18], to control the search procedure, local search approaches were embedded in GA algorithm. Obtained results show better performance in comparison with classical versions of GA. In [42], simulated annealing algorithm in conjunction with Genetic Algorithms were used to optimize industrial production management problem. In addition, a hybrid approach based on Markov chain and simulated annealing algorithms was designed to tackle the travel salesman problem [28]. Furthermore, in [4], Particle Swarm Optimization algorithm was hybridized with Simulated Annealing algorithm to avoid PSO in getting trapped in local optima. The designed approach was applied to deal with complex multi-dimensional optimization problem. Finally, in [3], the ACO algorithm was used in conjunction with a GA as a hybrid approach for feature selection in text classification domain. The proposed approach recorded better results in comparison with filter approaches and the classic version of ACO. Despite the effectiveness that meta-heuristic algorithms have shown to solve many optimization problems, they have some shortcomings such as their sensitivity to parameter tuning, finding the optimal parameters for difference optimization problems is very important. Furthermore, in terms of performance, many of the present optimization algorithms have difficulties in dealing with high dimension optimization problems [5]. In addition, the computational cost of using meta-heuristics in general and the hybrid meta-heuristic approaches in particular is relatively high. In this work, SA is used to enhance the best solution found so far (subset of features) by the Binary Dragonfly Optimizer. The classic version of Binary Dragonfly Algorithm was used for feature selection [26], but to best of our knowledge, this is the first time that Binary Dragonfly hybridized with Simulated Annealing is utilized in feature selection domain. The remaining of this paper is organized as follows: related work is reviewed in section “Related Work”, whereas section “Algorithms” presents the BDA and SA Algorithms. Section “Proposed Feature Selection Method” explains the proposed FS approach. In section “Experiments”, the experimental results are presented and discussed. Conclusion and future work are shown in section “Conclusion”. In this paper, there are number of key contributions as follows: BDA-SA: an improved version of the standard binary version of BDA is proposed. The main improvement includes the combination of DA with SA to solve the local optima problem of standard BDA. BDA-SA wrapper feature selection model is developed in this paper. We evaluated and compared BDA-SA with a number of well-known algorithms including (BDA, PSO, BGWO and BALO). We compared the results with using all features, where we used 18 data sets from UCI repository which are frequently used by feature selection research. and ALO) using 18 benchmark data sets from UCI repository. From these results, it is clearly confirmed the superiority of BDA-SA in comparison to these baseline algorithms.

Related Work

Based on literature investigation, many optimization algorithms were improved by combining them with local search algorithms (LSA). For example, in a study by [11], Elgamal et al. improved Harris Hawks Optimization (HHO) Algorithm by SA and applied it for feature selection problem. In [2], the authors also improved water cycle optimization with SA and applied it for spam email detection. In work by [41], performance of Salp Algorithm (SSA) was enhanced by combining it with a new developed LSA and applied it for feature selection problem. Also, in [40], WOA was combined with new a new local search algorithm to overcome the problem of local optima, and applied it for rules selection problem. Furthermore, in [21], Jia et al. improved spotted hyena optimization using SA and applied it for feature selection problem. Also, Simulated Annealing was hybridized with GA in [27] to be utilized as feature selection method for the classification of power disturbance in the Power Quality problem. In [33], Genetic Algorithm GA was used with SA to extract features to tackle the examination timetabling problem. A local search strategy was embedded in Particle Swarm Optimization algorithm to guide the PSO during the search process for the best subset of feature in classification [31]. Mafarja and Mirjalili [25] proposed two hybrid wrapper feature selection approaches based on Whale Optimization and Simulated Annealing algorithms. The aim of using SA is to enhance the exploitation of WOA. Results showed that the proposed approaches improved the classification accuracy in comparison with other wrapper feature selection techniques. The literature also revealed that many successful efforts were done to improve the performance of Dragonfly optimization algorithm. For instance, in [19], Sayed et al. used several chaotic maps to adjust the movement parameters of dragonflies of DA algorithm through the iterations to accelerate its convergence rate. Hammouri et al. in [15] adopted several functions such as linear, quadratic, and sinusoidal for updating the main coefficients of Binary Dragonfly algorithm. Also, a hyper learning strategy was utilized by [39] to boost the Binary Dragonfly algorithm to avoid local optima, and enhance its search behaviour for an ideal subset of features. the proposed method was applied to a coronavirus disease (COVID-19) data set. Experimental results demonstrated the ability of hyper learning based BDA in improving the classification accuracy. Finally, in [34], Qasim et al. proposed a feature selection approach in which the binary dragonfly algorithm (BDA) was hybridized with statistical dependence (SD). The proposed hybrid approach confirmed its efficiency in increasing the classification accuracy. Thus, based on the achieved results and improvements conducted in these mentioned studies, it has motivated us to combine SA as a LSA with DA to improve its search ability for feature selection problem.

Algorithms

Dragonfly Algorithm

Dragonfly Algorithm is a recent biologically inspired optimization approach which was proposed by Seyedali Mirjalili in 2015 [30]. It was found that dragonfly swarming behavior depends on two sorts of swarming behavior: hunting and migration [30, 36]. Hunting swarm of dragonflies moves in small subgroups over a restricted area to find and hunt preys. This behavior was utilized to simulate the exploration part of optimization process. In the migration behavior, in contrast with hunting swarm, dragonflies move a long one direction in bigger subgroups. This behavior was exploited to simulate the exploitation part of the optimization [30, 36]. Generally, the aim of swarm members is to co-operate to discover food places, and to protect themselves form danger of enemies. Based on these two aims, a set of factors is mathematically modeled for adjusting the positions of members in the swarm. The mathematical models for implementing the swarming behavior of dragonfly insects are given as follows [12, 26, 30]: Separation indicates the way that flying dragonflies follow to avoid clashes between themselves. This can be mathematically written as in Eq. (1):where X refers to the current search agent, while denotes the th neighbor of X. N represents the number of neighbors. Alignment refers to the way of adjusting the velocity of an individual with respect to the velocity vector of other close dragonflies in the swarm. This can be mathematically written using Eq. (2):where refers to the j-th neighbor’s velocity vector. Cohesion is a factor for position update of search agents that represents the desire of search agents to travel towards the mass center. It is mathematically written as in Eq. (3):Attraction denotes the interest of search agents to travel in direction of food location. The tendency of th member in the swarm to move towards the food source is obtained using Eq. (4):where refers to the location of food source, and X refers to the current member. Distraction refers to mechanism that dragonflies follow to flee from enemy. The distraction of th dragonfly is defined as in Eq. (5):where presents current position of the enemy, and X is the position of the current member. To find the optimum solution for a given optimization problem, DA defines a position vector and a step vector for each search agent in the swarm. These vectors are utilized to update the positions of search agents in the search space of the given optimization problem. The step vector which refers to travelling direction of dragonflies is formulated as follow [26, 30]:where s, a, c, f, and e are known as weighting factors for separation (), alignment (), cohesion (), attraction () and distraction () of the i-th search agent, respectively. w refers to the inertia weight. The obtained step vector () is used to estimate the position vector of search agent X as follows:where t indicates the current iteration. The basic version of Dragonfly optimizer is proposed for the problems in continuous search space. The dragonflies can update their position by adding the step vector to the position vector. Feature selection is a binary optimization problem, so the update strategy as in Eq. (7) is not possible for binary search space. Mirjalili [30] utilized the following transfer function to convert the step vector values to a number restricted in [0,1].The above transfer function is used to find the probability of updating the position of dragonflies in the swarm, and then the following equation is employed to update the positions of dragonflies (search agents):where r is a number in the range of [0,1]. Algorithm 1 presents the pseudocode of Binary Dragonfly algorithm.

Simulated Annealing

SA is a single-solution based meta-heuristic optimization algorithm, introduced by Kirkpatrick et al. [23] in 1983. It has been widely used to tackle discrete and continuous optimization problems. SA is classified as a hill-climbing local search approach, in which a certain probability is used to decide weather to accept worse solution or not [23]. SA generates an initial solution (in our case, the best solution found so far by BDA is used as SA initial solution). A neighbour solution to the optimal one found so far is generated by SA based on fitness value of the neighbour and a specific neighbourhood structure. If the calculated fitness of neighbour is better (less or equal) than the fitness of optimal solution, then the neighbour solution is selected as the optimal one. Boltzmann probability, is applied as an acceptance condition of the neighbour solution. refers to the difference between the fitness of the optimal and neighbour solutions, and T is named temperature which is gradually reduced based on cooling schedule throughout the search procedure [20, 23, 25]. In this paper, as adopted in [25], the initial temperature equals , where N is the number of features in each data set, and , is applied to calculate the cooling schedule. Algorithm 2 presents the pseudocode of SA algorithm [25].

Proposed Feature Selection Method

Feature selection problem is distinguished as a binary optimization problem, so binary vectors are used to represent the solutions. In this way, if the value of a specific cell in the binary vector is set to 1, then the corresponding feature is retained. Otherwise, that feature is ignored. The size of the binary vector is equal to the number of features in the data set. Dragonfly optimization algorithm is a newly introduced optimization approach. The basic version of binary DA algorithm was used for feature selection in [26]. The main aim of this work is to improve the performance of BDA for feature solution problem. To achieve that purpose, the best solution obtained so for by BDA is passed to the SA algorithm to be used as an initial solution instead of the random generation of the initial solution. Therefore, SA will conduct a local search starting with the optimal solution found so far by BDA in attempt to find a better one. Figure 1 presents the flowchart of the proposed approach.
Fig. 1

Flowchart of BDA-SA algorithm

Flowchart of BDA-SA algorithm Feature selection is a multi-objective optimization problem, where the maximum accuracy of the classifier and least number of selected features are two related objectives need to be achieved. Eq. (10) is commonly used as fitness function for feature selection [6, 25, 26].where er denotes the error rate of KNN machine learning classifier using the selected subset of features. and are two parameters used to make a balance between the classification accuracy and the size of subset of features (selected by the search agent), is a number restricted in the range [0,1], and equals 1 - . N refers to the total number of features in the data set, and m indicates the cardinality of the subset of features selected by the search agent. In this work, since we are mostly interested in getting the highest classification accuracy, is set equal to 0.99 as in the previous work [26]. The value is set 0.99, because in this work we improved the binary version of DA algorithm which was developed in [26], and we used the same setting in our experiments to compare our results with the results were reported in BDA [26]. In general, diversification has larger importance than intensification in exploring potentially useful areas of the feature space especially at the beginning of the search process. In later phases, exploitation has larger importance, because the search for better solutions around the best one found by the exploration phase is required [24]. Hybrid approaches such as BDA-SA in our case can be used to achieve the desired balance between exploring and exploiting the search space. However, in comparison with classic wrapper approaches, where a heuristic technique and an evaluator are used, the computational cost of utilizing hybrid approaches is higher.

Experiments

Data Sets

In this work, 18 data sets from UCI repository were used to assess the performance of the proposed feature selection approach [10]. They are the same data sets used by many researchers to evaluate various feature selection approaches. Table 1 outlines the details of applied data sets for evaluating the proposed Binary Dragonfly algorithm based feature selection approach.
Table 1

Data sets used in the experiments

Data setFeaturesInstances
Breastcancer9699
BreastEW30569
CongressEW16435
Exactly131000
Exactly2131000
HeartEW13270
IonosphereEW34351
KrvskpEW363196
Lymphography18148
M-of-n131000
PenglungEW32573
SonarEW60208
SpectEW22267
Tic-tac-toe9958
Vote16300
WaveformEW405000
WineEW13178
Zoo16101
Data sets used in the experiments

Parameter Settings

As in [26], each data set is split into three equal sets: training set, validation set, and test set. In addition, K-fold-cross-validation procedure is used to evaluate KNN classifier (the parameter K of KNN classifier is set to five as adopted in [26]). Also, results of several metaheuristics-based wrapper feature selection algorithms including Binary Particle Swarm Optimization (BPSO), Binary Ant Lion Optimization (BALO), and Binary Gray Wolf Optimization (BGWO) were also used for comparison purpose. Furthermore, for Binary Dragonfly algorithm, the original paper of Dragonfly algorithm [30] comprehensively studied the appropriate values of swarming factors and the inertia weight, for that reason, the same best parameter settings reported in that paper were adopted in this work. Moreover, the same values of parameters adopted in [25] for SA algorithm were used. The parameters of BGWO, BPSO, and BALO algorithms were selected based on recommended setting reported in the original publications and related studies in feature selection domain. In all conducted experiments, common parameters were set as in Table 2. These values were set following a range of initial experiments. Comparison between approaches were made based on three criteria comprising classification accuracy, number of selected features, and best fitness. In addition, each approach was run 20 times with random initial solutions on a machine with Intel Core i5 processor 2.2 GHz and 4 GB of RAM.
Table 2

Parameter setting of algorithms

ParameterValue
K parameter for KNN classifier5
\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\propto$$\end{document} parameter for fitness function0.99
Number of search agents10
Maximum number of iterations100
a of BGWOfrom 2 to 0
\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\omega$$\end{document}ω of BPSOfrom 0.9 to 0.2
C1 and C2 parameters for velocity of BPSO2.0
Parameter setting of algorithms

Results and Discussion

This section presents all recorded results obtained from the proposed FS approach. The proposed hybrid approach BDA-SA was compared to the original version of BDA based approach. In terms of classification accuracy, it is clear from Table 3 that BDA-SA is able to classify most accurately on all data sets. It can be stated that SA succeeded to enhance the best solution found by BDA algorithm. In addition, in terms of best fitness, Table 3 reveals the averages of best fitness for the applied FS approaches on each data set. In most of the cases, BDA-SA obtained the lowest average of fitness value. BDA is slightly better than BDA-SA in only two cases (SonerEW and M-of-n dat asets). Although, we will see later that the differences are not statistically significant. It is clear from Table 3 that BDA-SA has less averages of selected features than BDA in some cases, while the averages recorded by BDA are less on other cases. We previously observed that BDA-SA is superior on all cases in terms of classification accuracy. It is evident that when the average of selected features by BDA-SA is greater than BDA, this means that BDA-SA approach managed to find informative and relevant features ignored by BDA, and when the average of selected features by BDA-SA is less than BDA, that means BDA-SA approach may removed some noisy or irrelevant features selected by BDA approach.
Table 3

Averages of classification accuracy, best fitness and selected features obtained from BDA and BDA-SA

Data setAccuracyBest FitnessSelected Features
BDABDA-SABDABDA-SABDABDA-SA
Breastcancer0.9680.9880.0380.0186.0006.000
BreastEW0.9600.9740.0430.03013.50012.450
CongressEW0.9670.9750.0350.0285.0004.150
Exactly0.9821.0000.0230.0056.2506.000
Exactly20.7440.7590.2440.2401.4501.450
HeartEW0.8420.8950.1590.1095.9506.900
IonosphereEW0.9190.9290.0790.07411.65010.400
KrvskpEW0.9580.9760.0310.02818.35014.400
Lymphography0.8720.9110.1310.0927.4008.500
M-of-n0.9950.9970.0060.0086.2006.150
PenglungEW0.9090.9300.0930.049123.350141.300
SonarEW0.9140.9170.0770.08627.55023.900
SpectEW0.8570.8660.1430.137 8.6508.850
Tic-tac-toe0.7840.8180.2070.1898.2007.800
BDA Vote0.9530.9640.0490.0376.5503.050
WaveformEW0.7550.8050.2360.19821.00021.100
WineEW0.9870.9990.0090.0086.2008.850
Zoo0.9590.9790.0420.0248.3505.650

Bold values represent the best results

Averages of classification accuracy, best fitness and selected features obtained from BDA and BDA-SA Bold values represent the best results Since the main aim of this work is to enhance the performance of BDA algorithm by hybridizing it with SA algorithm, we conducted further statistical analysis to demonstrate that the hybrid BDA-SA approach is better than using the basic BDA alone for feature selection problem. Table 4 presents the standard deviation values of the averages of classification accuracy, best fitness, and number of selected features for BDA and BDA-SA approaches on each data set. In terms of classification accuracy, as in Table 4, it can be observed that BDA-SA approach behaves more robust than BDA based approach on almost all data sets. In terms of best fitness and number of selected features, as shown in Table 4, it can be seen that BDA-SA approach is better than BDA approach on half of the cases.
Table 4

Standard deviation of the averages of classification accuracy, best fitness and selected features obtained from BDA and BDA-SA

Data setAccuracyBest FitnessSelected Features
BDABDA-SABDABDA-SABDABDA-SA
Breastcancer0.0012780.0006390.0000000.0006330.0000000.000000
BreastEW0.0051400.0040200.0043600.0041602.0131102.665000
CongressEW0.0041830.0034900.0036030.0033201.7167901.039900
Exactly0.0630000.0000000.0630700.0000000.9104600.000000
Exactly20.0358000.0015730.0003330.0004981.2343001.394500
HeartEW0.0142740.0030400.0080070.0031830.9986830.640720
IonosphereEW0.0128870.0108240.0092050.0110202.6808002.186000
KrvskpEW0.0410390.0034350.0051900.0036482.5188762.542270
Lymphography0.0113710.0081730.0114500.0080921.1424811.960129
M-of-n0.0169380.0081410.0031920.0082980.4103910.366348
PenglungEW0.0158690.0289960.0155860.0161017.7274298.676041
SonarEW0.0219430.0185610.0113500.0185944.0062453.567396
SpectEW0.0083520.0082550.0051230.0083041.2258192.433862
Tic-tac-toe0.0385100.0039620.0100510.0017971.6415651.880649
Vote0.0068400.0032620.0026800.0036122.4596751.234376
WaveformEW0.0282550.0060230.0009380.0062782.9912152.174009
WineEW0.0304130.0034580.0045650.0025951.5761381.814416
Zoo0.0167110.0043840.0131610.0042601.5652481.565200

Bold values represent the best results

Standard deviation of the averages of classification accuracy, best fitness and selected features obtained from BDA and BDA-SA Bold values represent the best results The average and standard deviation were used as measures to compare the overall results obtained form BDA and BDA-SA. To see whether the differences in the results are statistically significant or not, the non-parametric Wilcoxon test with significant level 0.05 was applied. This test is appropriate to compare the algorithms that have stochastic behaviour [9]. As in Table 5, the p values of the accuracy and the fitness show that BDA-SA recorded significantly better results than BDA on most of the data sets. In terms of selected features, the differences are statistically significant on eight cases, while on other data sets including BreastEW, CongressEW, WaveformEW, SpectEW, and IonosphereEW, p values show that the differences are not statistically significant. The superiority of BDA-SA particularly in terms of classification accuracy is expected, since it utilizes two powerful searching algorithms, DA which is efficient in exploration and SA that has a strong exploitation capability. The ability of DA is utilized in exploring the highly relevant regions in the feature space and avoid the trap of local optima, and then SA is used to intensify the nearby regions to the optimal solution (best subset of features) discovered by BDA algorithm. However, in terms of computational time, as revealed in Fig. 2, in all cases, the computation cost of BDA-SA is higher compared to BDA.
Table 5

P values of the Wilcoxon ranksum test over 20 runs for classification accuracy, Best Fitness and Selected Features of BDA and BDA-SA (P 0.05 have been underlined)

Data setAccuracyBest fitnessSelected features
Breastcancer0.0000800.000080N/A
BreastEW0.0000800.0000800.170680
CongressEW0.0005400.0003200.107400
ExactlyN/AN/AN/A
Exactly20.0001400.000080N/A
HeartEW0.0000800.0000800.006140
IonosphereEW0.0244400.0872600.123560
KrvskpEW0.0734600.0672400.000500
Lymphography0.000080.000080.047700
M-of-nN/AN/AN/A
PenglungEW0.0139000.0001000.000160
SonarEW0.6241400.2040800.020880
SpectEW0.0232000.0444400.968100
Tic-tac-toe0.0000800.000080N/A
Vote0.0002000.0000800.000420
WaveformEW0.0000800.0000800.703940
WineEW0.0587600.7794800.001000
Zoo0.0006400.0000800.001420

The underline values mean that there is no significant difference

Fig. 2

Averages of computational time for BDA and BDA-SA

P values of the Wilcoxon ranksum test over 20 runs for classification accuracy, Best Fitness and Selected Features of BDA and BDA-SA (P 0.05 have been underlined) The underline values mean that there is no significant difference Averages of computational time for BDA and BDA-SA The performance of BDA-SA was also compared with three meta-heuristic-based feature selection approaches including Binary Particle Swarm Optimization, Binary Ant Lion Optimization, and Binary Gray Wolf Optimization. In terms of accuracy rates, as revealed in Table 6, BDA-SA outperformed all its competitors. In addition, in terms of best fitness rates, as presented in Table 7, BDA-SA recorded lowest averages of best fitness on fifteen out of eighteen data sets. Furthermore, in terms of lowest number of selected features, Table 8 shows that BDA-SA outperformed other algorithms on more than 50% of tested cases. Also, the average of computational time for each approach was considered. Figure 3 shows the computational cost of BDA-SA, BPSO, BGWO, and BALO. It can be observed that BPSO is the best in terms of lowest computational time.
Table 6

Comparison between BDA-SA and other algorithms in terms of classification accuracy

Data setBPSOBALOBGWOBDA-SA
Breastcancer0.9490.9310.9570.988
BreastEW0.9240.9300.9350.974
CongressEW0.9120.9150.9230.975
Exactly0.6720.6260.6661.000
Exactly20.7250.7020.7170.759
HeartEW0.7890.7510.7510.895
IonosphereEW0.8450.8130.7870.929
KrvskpEW0.8500.7610.8570.976
Lymphography0.7300.6840.7860.911
M-of-n0.8140.7330.8000.997
PenglungEW0.7280.7890.7240.930
SonarEW0.7990.7780.7960.917
SpectEW0.8080.8310.7870.866
Tic-tac-toe0.7120.6940.7070.818
Vote0.9050.8820.9150.964
WaveformEW0.7430.7010.7430.805
WineEW0.9160.9160.8950.999
Zoo0.8460.8080.8290.979

Bold values represent the best results

Table 7

Comparison between BDA-SA and other algorithms in terms of best fitness

Data setBPSOBALOBGWOBDA-SA
Breastcancer0.0380.0380.0310.018
BreastEW0.0440.0360.0450.030
CongressEW0.0340.0450.0300.028
Exactly0.0370.1640.1630.005
Exactly20.2430.2570.2460.240
HeartEW0.1350.1650.1660.109
IonosphereEW0.1130.1410.1760.074
KrvskpEW0.0300.0450.0490.028
Lymphography0.1450.1560.1050.092
M-of-n0.0050.0460.0420.008
PenglungEW0.1650.1220.2160.049
SonarEW0.0930.1230.1260.086
SpectEW0.1340.0960.1270.137
Tic-tac-toe0.2010.2170.1910.189
Vote0.0320.0620.0460.037
WaveformEW0.2000.2190.2170.198
WineEW0.0180.0080.0380.008
Zoo0.0450.0330.0840.024
Table 8

Comparison between BDA-SA and other algorithms in terms of selected features

Data setBPSOBALOBGWOBDA-SA
Breastcancer6.0006.0004.0006.000
BreastEW12.80022.60020.60012.450
CongressEW3.2009.3009.5004.150
Exactly6.2007.7007.7506.000
Exactly22.0508.4508.2001.450
HeartEW3.5508.9008.5006.900
IonosphereEW12.30025.50020.55010.400
KrvskpEW16.85026.90024.30014.400
Lymphography6.05011.05011.8008.500
M-of-n6.0007.5007.6506.150
PenglungEW146.250253.650204.700141.300
SonarEW27.00047.45040.05023.900
SpectEW8.00015.35013.7508.850
Tic-tac-toe8.9009.0006.0007.800
Vote5.0008.5008.1003.050
WaveformEW21.15032.40030.60021.100
WineEW6.45010.2508.4008.850
Zoo8.15010.5509.3505.650
Fig. 3

Comparison between BDA-SA and other algorithms in terms of computational time

Comparison between BDA-SA and other algorithms in terms of classification accuracy Bold values represent the best results Comparison between BDA-SA and other algorithms in terms of best fitness Comparison between BDA-SA and other algorithms in terms of selected features Comparison between BDA-SA and other algorithms in terms of computational time

Conclusion

This work introduced BDA-SA as a hybrid feature selection approach. The main goal was to enhance the performance of Binary Dragonfly algorithm especially in terms for classification accuracy. The best solution found so far by BDA algorithm was used as initial solution by SA algorithm to conduct a local search to find better solution than the one obtained by BDA. The proposed approach was assessed on a set of frequently used data sets from UCI machine learning repository. The performance of BDA-SA was compared to the native BDA algorithm as well as various algorithms comprising BGWO, BPSO, and BALO. Experimental results show that BDA-SA outperformed BDA and the other algorithms. In the future, it is worth to evaluate the proposed hybrid approach on more complex data sets. Below is the link to the electronic supplementary material. Supplementary material 1 (PDF 454 KB) Supplementary material 1 (DOCX 17 KB)
  2 in total

1.  Hybrid genetic algorithms for feature selection.

Authors:  Il-Seok Oh; Jin-Seon Lee; Byung-Ro Moon
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  2004-11       Impact factor: 6.226

2.  Optimization by simulated annealing.

Authors:  S Kirkpatrick; C D Gelatt; M P Vecchi
Journal:  Science       Date:  1983-05-13       Impact factor: 47.728

  2 in total
  2 in total

1.  A hybrid binary dwarf mongoose optimization algorithm with simulated annealing for feature selection on high dimensional multi-class datasets.

Authors:  Olatunji A Akinola; Absalom E Ezugwu; Olaide N Oyelade; Jeffrey O Agushaka
Journal:  Sci Rep       Date:  2022-09-02       Impact factor: 4.996

2.  Binary dwarf mongoose optimizer for solving high-dimensional feature selection problems.

Authors:  Olatunji A Akinola; Jeffrey O Agushaka; Absalom E Ezugwu
Journal:  PLoS One       Date:  2022-10-06       Impact factor: 3.752

  2 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.