Literature DB >> 36056062

A hybrid binary dwarf mongoose optimization algorithm with simulated annealing for feature selection on high dimensional multi-class datasets.

Olatunji A Akinola1, Absalom E Ezugwu2, Olaide N Oyelade1, Jeffrey O Agushaka1.   

Abstract

The dwarf mongoose optimization (DMO) algorithm developed in 2022 was applied to solve continuous mechanical engineering design problems with a considerable balance of the exploration and exploitation phases as a metaheuristic approach. Still, the DMO is restricted in its exploitation phase, somewhat hindering the algorithm's optimal performance. In this paper, we proposed a new hybrid method called the BDMSAO, which combines the binary variants of the DMO (or BDMO) and simulated annealing (SA) algorithm. In the modelling and implementation of the hybrid BDMSAO algorithm, the BDMO is employed and used as the global search method and the simulated annealing (SA) as the local search component to enhance the limited exploitative mechanism of the BDMO. The new hybrid algorithm was evaluated using eighteen (18) UCI machine learning datasets of low and medium dimensions. The BDMSAO was also tested using three high-dimensional medical datasets to assess its robustness. The results showed the efficacy of the BDMSAO in solving challenging feature selection problems on varying datasets dimensions and its outperformance over ten other methods in the study. Specifically, the BDMSAO achieved an overall result of 61.11% in producing the highest classification accuracy possible and getting 100% accuracy on 9 of 18 datasets. It also yielded the maximum accuracy obtainable on the three high-dimensional datasets utilized while achieving competitive performance regarding the number of features selected.
© 2022. The Author(s).

Entities:  

Mesh:

Year:  2022        PMID: 36056062      PMCID: PMC9440036          DOI: 10.1038/s41598-022-18993-0

Source DB:  PubMed          Journal:  Sci Rep        ISSN: 2045-2322            Impact factor:   4.996


Introduction

Technological advancement in various fields of endeavor has resulted in a large amount of data being generated in the information industry. The massive data available today can only be meaningful if there are corresponding tools that can transform these data into information without stress. Data mining and machine learning are potent tools in this regard because there has been a tremendous growth in the use of these tools to transform massive data into meaningful information. However, this large amount of data comes with redundancies, noise, and many features which may hinder knowledge discovery activities like a classifier's performance. Knowledge discovery (KD) activities consist of repeatedly performing data cleaning, dimensional reduction, data integration and transformation, and many other activities. These activities form part of the pre-processing tasks, without which the performance of data mining and machine learning algorithms would be significantly affected. Data is so important nowadays that it is regarded as the 'new currency.' Careful handling of the 'new currency' is required, making data mining and machine learning a fast-growing field. The dimension of the vast data available is of great concern for data miners because it impacts their ability to transform the data into meaningful information. Many data mining and machine learning tools require considerable time to carry out their tasks. Therefore, noisy data with redundant features would increase these algorithms' time complexity. For this problem to be resolved, the pre-processing step of feature selection becomes crucial, which can impact the performance of learning algorithms. The feature selection then plays a notable role in many research areas[1,2]. Feature selection has become a prominent approach employed to remove irrelevant and unnecessary features, reducing the attributes that do not aid the purpose of classification but add more burden on computational cost and requirement for space. This process is often categorized into wrapper and filter approaches. The first employs one or more learning algorithms to extract a relevant subset of features. At the same time, the latter is independent of a learning algorithm and uses information gain, mutual information, Laplacian score, and many more to select useful features[3,4]. Generally, the wrapper methods yield better results in terms of classification accuracy[5] than the filter methods, which produce results faster than the wrapper approach. However, the wrapper methods are computationally expensive. This wrapper feature selection method is an optimization problem[6,7]. In tackling this optimization problem of feature selection, metaheuristic algorithms have played a prominent role in recent years. Among these methods are work done[6-11] which produced better solutions in comparison with the exact techniques like recursive feature elimination, mutual information, Laplacian score. The feature selection problem aims to locate the most appropriate or best subset of features, thereby choosing the minimum features possible and maximizing the classification accuracy. This has become a daunting task due to its contradicting, multi-objective nature. In recent years, metaheuristic algorithms have been utilized to solve this issue. As a result of the increasing time required to find the best feature subsets, especially in a high-dimension dataset, feature selection is considered an NP-hard problem[12]. Considering a dataset with an N feature, a sum of 2^N − 1 features combination is needed to investigate the location of the optima feature[13,14]. Therefore, a high-performing metaheuristic algorithm is required to reduce the processing time posed by this kind of problem. The random forest will likely solve this problem,however, the trees tend not to go deep and have a high bias[15]. Metaheuristic algorithms are deemed more appropriate for solving such problems due to their ability to cater to the worst-case scenario[16]. Many of these metaheuristic algorithms are found in literature, some of which include Particle Swarm Optimization (PSO)[17], Genetic Algorithm (GA)[18], Whale Optimization Algorithm (WOA)[19], DragonFly Algorithm (DA)[20], Cuckoo Search Optimization Algorithm (CSO)[21], Harmony Search algorithm (HS)[22], Grey Wolf Optimizer (GWO)[23], Krill Herd Algorithm (KHA)[24], Prairie Dog Optimization Algorithm (PDO)[25], and more. Each of these optimizers has been used to solve different problems with varying successes. Generally, metaheuristic algorithms' search process relies on the equilibrium between the diversification and intensification phases. Diversification is also an exploration where candidate solutions that are not neighboring are evaluated. Meanwhile, intensification is also regarded as exploitation where neighboring solutions are intensively searched for the best solutions. The importance of these two phases cannot be overemphasized as a factor in finding the best solution(s). Several researchers have employed metaheuristic algorithms to solve many feature selection problems[6,7,26]. Due to the great demand for these algorithms, researchers have invested a lot of time and effort in developing superior algorithms capable of generating high-quality solutions for the candidate problem. Since metaheuristics approaches can only give near-optimal results, developing new methods or improving existing algorithms to obtain better results or optimal solutions has become an ongoing endeavor. Furthermore, the No Free Lunch (NFL) theorem assumes that no single algorithm can produce optimal results for all optimization problems[27]. We can therefore conclude that there is no universal or one-size-fits-all algorithm for every optimization problem that can produce the best results. The reference to this theory has motivated research in this area. More researchers are developing efficient and novel metaheuristic algorithms to solve the FS problems. Though many metaheuristic algorithms have been used to solve the feature selection problems in the literature, most have inherent drawbacks or shortcomings[28]. Typically, these algorithms' performance largely depends on the innate traits of the datasets used to train those models. Usually, hybrid algorithms are often considered a better preference because the shortfall of one method is enhanced by the strength of the other[29]. Therefore, the success of the hybrid implementation of FS-based metaheuristic algorithms has motivated the current study, where a hybrid algorithm is proposed by hybridizing the binary variant of the standard DMO algorithm with the simulated annealing (SA). In this study, the binary DMO (or BDMO) was enhanced by improving its capacity to adequately exploit the intensification phase of the underlying problem landscape to produce better solutions. Therefore, the focus is on implementing a new hybrid solution that uses simulated annealing (SA)[30] to augment the identified limitation of the binary dwarf mongoose optimizer. The hybrid is subsequently employed to solve the problems of high-dimensional feature selection datasets. The new hybrid algorithm is called the binary dwarf mongoose simulated annealing optimization (BDMSAO) algorithm. In the new proposed feature selection method, the SA algorithm is used as a component in the BDMO or as a low-level team player in conjunction with the BDMO algorithm to increase the quality of the final feature selection results. Precisely, the SA is used to search the neighborhood of the best search agent to ensure that the local optima are enhanced. In addition, the SA is also employed after the BDMO is done exploring the solution search space to enhance the best-found solution, which afterwards is identified as the global best solution. It is noteworthy that several researchers have recently proposed similar hybrid methods, including the work presented in[1,31,32]. The results obtained by these hybrid methods proved their efficiency and outperformance over other related state-of-the-art techniques. Thus, in addition, this certainly also motivated our proposed hybrid version of the DMO. The major contribution of this study can be summarized as follows: A hybrid FS method called BDMSAO is introduced using BDMO and high-performing simulated annealing (SA). The proposed BDMSAO method is evaluated using 18 standard UCI datasets using the K-nearest Neighbors (KNN) classifier to prove its effectiveness. Also, the BDMSAO is applied to solve three high-dimensional datasets from the Arizona State University FS repository to prove its robustness further The performance of the proposed FS method is compared with many state-of-the-art metaheuristic-based FS methods. The remaining part of this paper is structured as follows: “Related work” section reviews related work. “Preliminaries” presents the two methods, BDMO and SA, that were hybridized in this study. Similarly, “The proposed hybrid method” presents the proposed method considering its solution presentation, fitness function, and computational complexity. In “Experimental results and discussion”, the experimental results of this study are discussed, and the statistical analysis test results are presented. “Testing on high-dimensional datasets” discusses the use of the proposed method on high-dimensional datasets to show its robustness. Finally, “Conclusion and future work” concludes this work by giving its limitation and future direction.

Related work

Recently, metaheuristic algorithms have gained ground in solving optimization problems, and these methods are regularly undergoing enormous improvements from researchers. Metaheuristic algorithms have become pivotal in finding optimal solutions through algorithms' learning using the process of iteration. In[33], metaheuristic algorithms were divided into population-based and single solution-based. Also[34], categorized these algorithms into non-nature inspired and nature-inspired metaheuristics. Many researchers have developed several hybrid forms of metaheuristic algorithms to solve feature selection problems. The hybridized methods have proven their superior performance in solving practical and real-world problems[29]. In the first hybrid method (metaheuristic), the Genetic Algorithm (GA) was combined with a local search algorithm to solve the optimization problem of feature selection[35]. From the inspirational viewpoint, metaheuristic algorithms can be generally grouped into swarm-based, physics-based, evolutionary-based, and human-based.

Swarm-based algorithms

Algorithms in this group are inspired by the social interaction or behaviour of birds, animals, insects, fish, schools, herds and so on. The main underlying idea of these algorithms is that everyone has a particular behaviour but coming together as a group or team and harnessing their joint effort enables them to solve very complex optimization problems. Several algorithms have been developed in this category in the last two decades, and researchers have also developed variants of some popular ones. Others have or/are still hybridizing them to solve various optimization problems. One of the prominent ones is the PSO[17] which has gained so much attention due to its rich mathematical basis for solving problems. Other algorithms in this group are Cuckoo Search (CS)[21], Grey Wolf Optimizer (GWO) by[23], Krill Herd Algorithm (KH)[24], Whale Optimization Algorithm (WOA)[36], Dwarf Mongoose Optimization (DMO)[37] algorithm, Gazelle Optimization Algorithm (GOA)[38] etc. As one of the most notable algorithms in the swarm-based category, the PSO has also been greatly hybridized to solve the feature selection problem. In[39], a local search algorithm was employed to assist the PSO in searching for the optimal solution and selecting the minimum reducts in relation to their correlation information. Talbi et al.[40] proposed a wrapper-based hybrid GA-SA method called GPSO using SVM as the classifier, and the work of[41] presented a multi-objective and hybrid mutation operator, which were both applied to the classification of microarray data. The study in[42] presented a novel hybridization of the GA with PSO in optimizing feature sets on the datasets of Digital Mammogram. The studies in[43,44] presented two different wrapper-based feature selection methods that hybridized the GA with Ant Colony Optimiser (ACO). Another study[45] combined the GA and Cuckoo search algorithms as a wrapper-based method to tackle the feature selection problem. Other hybrid methods include Harmony Search Algorithm and Stochastic Local Search (Nakkaa & Boughaci, 2016), Artificial Bee Colony (ABC) and Differential Evolution (DE) algorithms[46].

Evolutionary-based algorithms

Algorithms that fall under this category are inspired by nature or through biological process of evolution and begin their process by randomly generating their population solutions. The foremost algorithm in this category is the Genetic Algorithm (GA)[18] which generated its fittest individual using mutation and crossover in every generation. The GA has attracted a lot of attention with the creation of different variants, and improvements have been employed to solve many real-world problems. Other popular algorithms developed in this group include genetic programming[47], tabu search[48], evolution strategy, differential evolution, flower pollination algorithm[49], memetic algorithm[50], Biogeography-Based Optimization[51], and more. Apart from the presentation of these evolutional metaheuristic algorithms, the GA being the prominent algorithm in the evolutionary category has attracted significant attention where it was hybridized with other methods to solve different optimization problems[52-55] where those studies revealed the potency (in terms) of producing better output in comparison with either other local or global search models. It has also been widely hybridized in the domain of feature selection. The GA was also combined with the SA as a filter approach[56] to enhance the GA's local search capability to solve the feature selection problem. This method was evaluated using eight datasets from the UCI machine learning repository. It performed better in selecting the minimum number of a subset of features than other popular methods. Also in[57], the authors proposed a memetic feature selection algorithm where the study utilized the fuzzy logic in controlling the major parameters on two local search techniques, which was later combined with the GA. In application to the wrapper-based method, the crossover operator of the GA was combined into the metropolis acceptance criterium of the SA[58]. Furthermore, it was hybridized in[59] in classifying power disturbance in the problem of Power Quality (PQ) which also optimize SVM parameters. Moreso in[60], it was combined with Tabu Search, which employed the Fuzzy ARTMAP Neural Network to evaluate the wrapper feature selection method.

Physics-based algorithms

These algorithms draw their inspiration from the laws of physics in the world. Physics-based methods are inspired by physics principles, chemistry, music, complex dynamic systems, physics and metallurgy to mathematics[1]. Some prominent algorithms in this group include Gravitational Search Algorithms[61], Atom Search Optimizer[62], Ray Optimization[63], Galaxy-Base Search Algorithm[64], Equilibrium Optimizer (EO)[65], Sine Cosine Algorithm (CSA)[20] and so on. In[66], a hybrid metaheuristic feature selection method was proposed using a Golden Ratio Optimization (GRO) and Equilibrium Optimization (EO) algorithms called the Golden Ratio based Equilibrium Optimization (GREO) algorithm, which was applied in speech emotion recognition. Also, the study conducted in[67] presented a hybrid feature selection method that is based on the ReliefF filter technique and EO known as RBEO-LS, which have two phases: the first employed the ReliefR algorithm at the pre-processing stage for feature weights assignment, and the second utilized binary EO (BEO) as a wrapper search technique.

Human-based algorithms

Algorithms in this category are inspired by activities performed by humans or human behaviours. Human beings are involved in various activities that affect their performance, and researchers use these behaviours to develop algorithms. The most prevalent algorithms here are Teaching Learning-Based Optimization (TLBO) by[68] and League Championship algorithms (LCA)[69]. Others include Exchange Market Algorithm (EMA) by (Ghorbani & Babaei, [70]), Social-Based Algorithm (SBA) by[71], Seeker Optimization Algorithm (SOA) by[72], etc. It is observed from the literature that not many algorithms are human inspired. TLBO, a well-known method in this category, was hybridized in[73] with extreme learning machines (ELM), referred to as TLBO-ELM in solving data classification problems which feature selection falls under. It was tested on some UCI benchmark datasets. With the strength of metaheuristic algorithms comes to some issues, among which is premature convergence that results in locating limited optimal solutions. Frequently, researchers combine these algorithms with other methods like local search techniques. Generally, the local search algorithm tries to conduct an intensive search of each region of the solution, which can outperform existing metaheuristic solutions. Among the existing local search methods are Simulated Annealing (SA)[30], tabu search[48], and Hill climbing (HC). The HC have variations like the βHC, Adaptive βHC and Late Acceptance HC (LAHC). Several works exist in the literature that employed the combination of local search strategy with metaheuristic algorithms. Some of them includes[1,74-78].

Preliminaries

Dwarf Mongoose Optimization Algorithm

DMO[37] is a population-based stochastic metaheuristic algorithm inspired by the foraging and social behaviour of dwarf mongoose, also called Helogale. Each dwarf mongoose search for food individually since food search is not a collective exercise, but foraging is done collectively. Due to the seminomadic attribute of these animals, the building of a sleeping mound is close to an abundant source of food. The algorithm mathematically models the lifestyle of this animal to solve optimization problems. All population-based optimization algorithms commence with random initialization. After that, because of the intensification and diversification rules, every solution gathers around the global best optima. Similarly, the DMO starts its solution by initializing the mongoose's candidate population. This population is generated stochastically between a particular problem's lower and upper bounds.where X represents the set of the candidates' present population that are generated randomly using Eq. (2), indicates the position of the jth dimension of the ith population, n indicates the population size, and d is the dimension of the problem.where is a random number that is distributed uniformly, are lower bound and upper bound, respectively is the dimension of the problem. The best solution at each iteration is the best solution obtained so far. Like every metaheuristic algorithm, there are two phases in the DMO: exploitation (individual mongoose carries out a thorough search in each search space), also called intensification and exploration (a random search for a new abundant food source or new sleeping mound) or diversification. Three major social structures of the DMO carry out the activities of the two phases mentioned: the alpha group, scout group, and babysitters. The alpha female is the family unit controller and is selected using Eq. (3). matches the number of mongooses in the alpha group. The number of babysitters is denoted by and represents the sound of female alpha to the path of the other unit members. The sleeping mound is determined by abundant food which is expressed in Eq. (4).where is a random uniformly distributed number [− 1,1], after each iteration, there is an evaluation of the sleeping mound; Eq. (5) represents this.when a sleeping mound is found, an average value is derived using Eq. (6) Once the babysitter exchange criterium is attained, the next phase is the scouting, which evaluates the next sleeping mound determined by another food source. Since mongoose is known not to return to a prior sleeping mound, the scout group goes searching for the next sleeping mound to ensure exploration. The mongoose is known to forage and scout simultaneously in DMO with the justification that the farther the unit forage, the likelihood of finding the next sleeping mound. This is simulated using Eq. (7).where shows the parameter that directs the collective-volatile movement of the mongoose's group that linearly decreases during iterations. connotes the vector that motivates the movement of mongoose to a new sleeping mound. The babysitter's group remains with the juveniles when the scouting and foraging group searches for a sleeping mound and food source. The number of the members of this group is removed from the total number of candidate population as they do not go to forage or scout until the babysitter exchange parameter is met in Eq. (7). In the following section, the proposed binary variant of DMO is presented with the hybrid method of simulated annealing (SA).

The proposed hybrid method

In this section, the representation of the solution, the fitness function utilized and the proposed method's computation complexity are elaborated. The feature selection problem, an optimization problem represented in binary form, and its solution confined to 0 s and 1 s were taken care of by the BDMO. The agents update their position in each iteration using the BDMO optimization rules and afterwards pass the solution to the SA to locate the better neighborhood solution to improve and refine the results. As a multi-objective optimization problem where two opposing objectives of high classification accuracy and minimal features selected as possible need to be met. The achievement of these two objectives determines how best a solution is. In the proposed method, the BDMO utilizes the tournament selection mechanism to advance the algorithm's diversification capability, which affords a high chance of selecting weak solutions while searching for promising ones. The DMO is a recent metaheuristic algorithm which proves its efficacy in solving mechanical optimization problems. The algorithm employs a Tau operator, which signifies that if a new food source is not found mindless, the fitness value of the present solution and the one being operated upon the intensification should be performed (Eq. 7). This operator was replaced with the SA as a local search technique that takes an initial state of a solution, processes it, and replaces the improved solution in place of the original one. This technique represents the hybridization of the BDMO and SA as a local search method.

Solution representation

The feature selection problem being an optimization problem has its output . The zero indicates that the feature is redundant or irrelevant and is thereby rejected, while one signifies that the feature is useful and therefore selected. The possibility that the results might be out of range cannot be ruled out. Therefore, the binarization function is applied to every agent to ensure they remain within the specified range. This is performed using Eq. (8). To select a feature, the position index must be 0.5 and above, which rounds the value to 1, and for any feature to be rejected, its position index must be less than 0.5, which is rounded down to 0.where is the best solution in dimension . Thereby, a mongoose's position shows that a feature set is selected as the value of position increases for the dimensions[79].

Fitness function

Selecting a useful feature that assists the classifier in recognizing a class of a sample in a dataset is challenging. During the selection process of relevant features, there is a need to remove the redundant ones for the sake of classification automatically and to maximize the accuracy of the classification problem when the feature selected is to be used[80]. In this work, the BDMSAO is utilized to locate the best feature subset and employ the KNN classifier to calculate the classification accuracy. The classification accuracy of this model is gotten by a classifier, represent the feature subset dimension, and the total number of attributes contained in the dataset is signified by . Therefore, the classification error is and the subset of selected features from the dataset is denoted by . Hence, the fitness function is defined as:where [0,1] is the weight assigned to the error classification.

The BDMSAO algorithm

The binary version of DMO is proposed in this study to solve the problem of feature selection for many benchmark datasets. The aim is to investigate the performance of the new hybrid algorithm in solving the challenging problem of selecting minimal features from high-density datasets. The resulting BDMO algorithm is supported by the SA method to boost its operations in the local search tasks. The evaluation of the fitness function in the proposed BDMO method utilizes Eqs. (10–12).where the represents the number of items in a normalized value of , and is applied to compute the actual fitness value of which is assigned to . Algorithm 1 is a detailed listing of the pseudocode for the proposed hybrid algorithm BDMSAO. The search mechanism of the alpha groups in the DMO is now replaced by the SA for an improved local search operation. The algorithm accepts datasets prepared as trainX, testX, trainy, testy, and in addition the population size, the number of iterations, and the dimension of population. Using these inputs, the population is initialized and computation for the fitness value for each individual in the population is derived. During the iteration, three search processes are evaluated including the SA-based search, scout group-based search and the baby seater-based search. These three-level searches apply nature-specific operations which demonstrates the balance for the exploration and exploitation phases of the proposed BDMSAO algorithm. Specifically, the SA is adapted to improve the local search of BDMO for improved performance. Once the three stages of the search process are completed, the global best solution is identified so that classification accuracy for the solution is computed using the datasets. The algorithm returns the computed number of features selected, the accuracy of the classification leading to the selected number of features, and the best solution.

Complexity of computation

The computation complexity of every metaheuristic algorithm depends on the time each candidate takes to update its positions, the maximum iteration value, other operations such as sorting or comparison, and variable update time. The computational complexity of the BDMSAO is is the maximum iteration number, is the population size, denotes the search space dimension, represent the classifier's time required to calculate the fitness of a given solution. The SA is employed to locate the best solutions if they can be found in the neighborhood of the present solution. In terms of , the SA does not significantly affect the computation cost. The optimization steps of the developed hybrid BDMSAO algorithm for solving feature selection problems is presented in Fig. 1. In the given figure, the hybrid BDMSAO's first step is defining all the parameters (which includes both BDMO and SA algorithms parameters, respectively). Then the next step is to generate the population representing a set of solutions for the feature selection problem. Subsequently, the fitness function of the individual candidate solution is determined based on evaluating and selecting the best features, after which the current best solution is identified and retained. The next step for the BDMSAO algorithm is to update the current population by using either the BDMO or SA algorithms, again depending on the quality of the fitness function. The process is such that if the probability of fitness function for the current solution is greater than 0.5, then the BDMO is selected for update. Otherwise, the SA algorithm is used to update the current population. Note that the probability above is computed as a factor of the position index (P_index) being . Thereafter, the fitness function for each solution is computed using Eq. (9), and the best solution is determined after updating the population. The next step is for the BDMSAO to check if the stopping criteria have been met and if yes, then the algorithm returns the overall candidate's best solution. Otherwise, the algorithm will iteratively repeat the previous steps from checking whether P_index is until finally the stop condition is reached.
Figure 1

Flowchart illustrating the algorithmic structure of the BDMSAO algorithm.

Flowchart illustrating the algorithmic structure of the BDMSAO algorithm.

Experimental results and discussion

Dataset (low, medium and high-dimensional)

To evaluate the performance of the BDMSAO, eighteen University of California Irvine (UCI) low, medium, & high-dimensional datasets and two high-dimensional datasets from the Arizona State University feature selection repository. The details of the datasets used, including their feature number (N), instances, classes, and categories, are presented in Table 1. The high-dimensional datasets contain numerous features of at least two thousand (2000), and few of the datasets are multi-class in nature, ranging from 3 to 9 classes. These high-dimensional datasets usually represent real-world scenarios and are therefore more challenging. This allows us to ascertain the robustness of the proposed feature selection method. Although, there are studies in the literature where some feature selection methods were utilized to solve high-dimensional datasets problems, one of which is the work presented in[81]. However, the maximum number of features in those datasets was only limited to N = 4703 compared to the more high-dimensional feature sizes utilized in the current study. Not many metaheuristic algorithms perform reasonably on high-dimensional and multi-class datasets.
Table 1

Datasets and their properties.

NumberDatasets# features# instances# ClassesCategories
1Breastcancer96992Biological
2BreastEW305692Biological
3CongressEW164352Political
4Exactly1310002Biological
5Exactly21310002Biological
6HeartEW132702Biological
7IonosphereEW183514Electromagnetic
8KrvskpEW3631962Game
9Lymphography181484Biological
10M-of-n1310002Biological
11PengiunEW325732Biological
12SonarEW602052Biological
13SpectEW222672Biological
14Tic-tac-toe99582Game
15Vote163002Political
16WaveformEW4050003Physical
17WineEW131783Chemical
18Zoo161016Artificial
Datasets and their properties.

Experimental setup

The proposed BDMSAO was implemented using python. Most often, parameters play a key role in determining the outcome of multi-agent algorithms, particularly the agents' number and iteration's total number, which heavily influence the algorithm's performance. Therefore, the experiment was performed considering different population sizes to determine the suitable size of the population and number of iterations. To test the efficiency of this hybrid approach, we compared the proposed BDMSOA with the BDMO. The classification accuracy and number of features selected are shown in Tables 3 and 5 using various population sizes from 10 to 50. The convergence graphs for both methods are also depicted in Fig. 2 to show the solution's optimal position over the total iteration number of 50. For a fair comparison, each dataset was run 10 times, and the average values of the runs were taken. The computer configuration for this implementation is Core i7, 3.60 GHz CPU with 16 GB RAM. The finding of this experiment reveals that the population size of 10 produced better results which will serve as the basis for comparison in this study. Table 2 presents the parameter setting for the developed hybrid FS methods.
Table 3

The classification accuracy of BDMO and BDMSAO on different population sizes.

Agent_size1020304050
DatasetBDMOBDMSAOBDMOBDMSAOBDMOBDMSAOBDMOBDMSAOBDMOBDMSAO
Breastcancer93.5710069.2999.2992.8610065.7199.2990.7199.29
BreastEW89.4798.2592.1196.4991.2398.2592.1198.2586.8498.25
CongressEW94.2598.8591.9510087.3698.8585.0610097.70100
Exactly65.5010063100661006610067.5100
Exactly270797479.507379.5073806780.50
HeartEW79.6390.7473.3392.5964.8192.5962.9696.3068.5290.74
IonosphereEW88.5797.1482.8697.1484.2995.7182.8695.7182.8698.57
KrvskpEW66.6798.7577.1598.4410098.7560.7298.7572.3098.59
Lymphography73.3310076.6796.677093.3363.3310076.6796.67
M-of-n6810067100631006710067.5100
PengiunEW73.33100801008010073.331008093.33
SonarEW71.1910073.8197.6271.4210064.2997.6276.1997.62
SpectEW76.1997.6264.2997.6266.6710061.9095.2480.9597.62
Tic-tac-toe65.1084.3865.1082.8161.9883.8562.5084.9064.0686.46
Vote88.3310086.6710093.3310088.3310080100
WaveformEW66.3085.2062.808567.1084.606085.5069.2086
Wine88.8910083.3310086.1110094.4410086.11100
Zoo9010090100901008510085100
Table 5

Classification accuracy of BDMSAO in comparison with other well-known feature selection methods tested using the UCI dataset, and the highlighted results indicate the highest classification accuracy.

NoDatasetsBDMSAOASGWBSNDOHSGWBPSOBGARSGWSSDs + LAHCECWSA-4EHHM
1Breastcancer10098.510098.696.2997.4397.198.9395.21100
2BreastEW98.2510098.2598.197.1997.5498.298.2597.38100
3CongressEW98.8599.410097.596.3396.7996.110096.2398.85
4Exactly10099.910010010010099.710078.09100
5Exactly27977.780.581.576.87777.97978.979.1
6HeartEW90.7483.190.7492.383.787.4184.891.6785.6390.7
7IonosphereEW97.1497.295.7494.494.8994.8997.896.4386.7998.6
8KrvskpEW98.7597.198.4497.398.3198.597.297.8193.5397.81
9Lymphography10088.496.6793.489.1983.7889.396.6787.0296.9
10M-of-n10010010010010010010010092.47100
11PengiunEW10010010094.291.8991.8910010087.63100
12SonarEW10094.895.2496.494.2399.0497.997.6276.8492.85
13SpectEW97.628796.2286.288.8189.5581.595.1579.8490.74
14Tic-tac-toe84.3886.587.582.879.9679.9685.987.2478.7585
15Vote10098.410098.39697.3399.610095.0898.4
16WaveformEW85.2074.68774.875.678.3675.784.480.1886.8
17Wine10010010010097.7598.8810010098.02100
18Zoo10010010010096.0890.210010098.95100
Figure 2

Convergence behavior of all algorithms on the ionosphere, congressEW, Exactly, Exactly2, and Vote datasets.

Table 2

Experimental parameter setting.

ParameterValue
K-fold cross-validation number10
Agent number10–0
Number of runs10
Maximum iterations50
Dimension of problemsNumber of available features in the dataset N
Peep (vocalization) of female alpha mongoose in BDMOControlled
Babysitters in BDMOControlled
initial temperature2 ∗|N|
Cooling schedule (T)0.93
Convergence behavior of all algorithms on the ionosphere, congressEW, Exactly, Exactly2, and Vote datasets. Experimental parameter setting.

Result and discussion

This subsection discusses the results generated by BDMSAO and BDMO, which were evaluated using eighteen datasets from the UCI repository, with details in Table 1. Since the proposed method is a wrapper-based approach, the utilized classifier is K-Near Neighbor (KNN) since it is a well-known and most widely used classifier in wrapper-based feature selection[82] and it was used with in the experiment. The generated results show the outperformance of the BDMSAO over the binary DMO. The outcome of the experiment in Tables 3–4 indicates the efficacy of the proposed hybrid method over the BDMO in locating better solutions. We can conclude that the BDMSAO performed better on UCI datasets than the BDMO.
Table 4

Number of features selected by BDMO and BDMSAO.

Agent_No dataset1020304050
BDMOBDMSAOBDMOBDMSAOBDMOBDMSAOBDMOBDMSAOBDMOBDMSAO
Breastcancer2413244314
BreastEW147451412212410
CongressEW632.0345.8542.5246.714
Exactly66463.6666666
Exactly23628575829
HeartEW4.7444.9255.922636.304
IonosphereEW14196115111999
KrvskpEW17194.16180244.831813.4520
Lymphography6105.3464.6532.8103.545
M-of-n6626162656
PengiunEW69132152162177135137772108
SonarEW16267205268241632
SpectEW29227292211322425
Tic-tac-toe1425342616
Vote1322735323
WaveformEW224522122524122
Wine5.18433144262
Zoo6.2158584647.914
The classification accuracy of BDMO and BDMSAO on different population sizes. Number of features selected by BDMO and BDMSAO. A critical inspection of the results in Table 3 indicates that BDMSAO generates better results than BDMO on all datasets. The classification accuracy produced is greater than 90% on 16 of 18 datasets (88.88%) except on Exactly2 & Tic-tac-toe and yielded 100% on 9 of 18 datasets (50%). In the number of features selected, the BDMSAO selected fewer features on 6 datasets (BreastEW, CongressEW, HeartEW, SpectEW, Wine & Zoo), the same number of features on 2 datasets (Exactly & M-of-n), and BDMO selected fewer datasets on 11 datasets. The convergence behavior of both methods is shown in Fig. 2. It is observed that each algorithm converges steadily in all datasets. However, the BDMSAO achieved a better convergence to show its superiority over the BDMO. Furthermore, the fitness function's optimization pattern on the defined problem space was investigated, and the results obtained were graphed for comparative presentation. The graphing is grouped based on the dataset used for experimentation so that each graph presents a comparative outline of curves for some selected algorithms considered in this study. Values used for the graphing were the fitness value over all iterations in each case of the datasets on all optimization algorithms. Again, these plots illustrate the convergence pattern for each algorithm as experimented on different datasets. The BDMO, BDMSAO, GNDO-SA, ASGW, HSGW, AIEOU, WOA, RSGW, RTHS, and PSO algorithms were considered in the convergence plots. Figure 2 shows the convergence plots for ionosphere, congressEW, Exactly, Exactly2, and Vote datasets. For the ionosphere dataset, curves of GNDO-SA, ASGW, HSGW, AIEOU, WOA, RSGW, RTHS, and PSO run below those of BDMO and BDMSAO. A similar pattern is repeated for congressEW, Exactly, and Exactly2 datasets. Interestingly, for the Vote dataset, the PSO algorithms perform better than all algorithms with the BDMO algorithm. A competitive performance seen with both BDMO and BDMSAO overlap in the congressEW and Exactly2 datasets but shows the slight distance in the case of the ionosphere, Exactly, and Vote datasets. The highest fitness value of 1.0 obtained for all 50 iterations is reported by PSO in the Vote dataset. The relative highest values obtained for 0.9 and above were those seen in BDMO and BDMSAO on the ionosphere, congressEW, and Vote datasets. The performance of BDMO and BDMSAO algorithms demonstrates a superior performance when compared with all the similar methods in all the five datasets compared in the figure. This shows that the algorithms are suitable for finding the minimum number of features required for classifying class distributions in the datasets. In Fig. 3, the convergence curve for the ASGW algorithm for all datasets of the colon, HeartEW, BreastEW, BreastCancer, and Lymphography are seen to spike with a measure of instability from the first iteration to the last. Those for GNDO-SA, HSGW, RTHS, RSGW, WOA, and AEIOU are poorly fitted, considering the convergence curves for these algorithms lying far below those of PSO, BDMO, and BDMO-SA. Note that in all datasets listed in the figure, PSO, BDMO, and BDMO-SA are seen to competitively converge above all other methods confirming the superiority of the three methods when compared with others. However, in all the five datasets reported in the figure, BDMO and BDMO-SA performed better than PSO, which only converges based on its function evaluation values above the other two when HeartEW and BreastCancer datasets were experimented with. In all cases where BDMO and BDMO-SA performed well above others, we see the fitness values obtained lined through values above 0.8. This significant classification value confirms that the number of features selected for the two algorithms in those datasets represents a high-quality selection.
Figure 3

Convergence behavior of all algorithms on colon, HeartEW, BreastEW, BreastCancer, and Lymphography datasets.

Convergence behavior of all algorithms on colon, HeartEW, BreastEW, BreastCancer, and Lymphography datasets. The BDMO and BDMO-SA are reported to have performed well in three of the five datasets used for the plots in Fig. 4. Although the PSO algorithm showed a competitive performance with the two algorithms, we note that these are only limited to the SpaceEW and M-of-n datasets. Meanwhile, other algorithms have their curves running below those for PSO, BDMO and BDMO-SA in all the five datasets. We discovered that the nature of the five datasets observed in the figure is computationally demanding, given the unstable performance of the ASGW and RSGW.
Figure 4

Convergence behavior of all algorithms on waveformEW, sonar, SpaceEW, M-of-n, and Zoo datasets.

Convergence behavior of all algorithms on waveformEW, sonar, SpaceEW, M-of-n, and Zoo datasets. Figure 5 shows the convergence curves for all algorithms on PenglungEW, Tic-tac-toe, Wine, and KrVsKpEW datasets. Interestingly, the impact of hybridizing BDMO with the SA algorithm proved outstanding as the algorithm competes with PSO well in three cases. The computational difficulty experienced with KrVsKpEW for all algorithms still puts the BDMO-SA algorithm ahead of others to demonstrate that the proposed hybrid algorithm is suitable for selecting the optimal number of features required for solving the classification problem. These outstanding performances are not limited to the KrVsKpEW dataset alone but span across all datasets considered in this study.
Figure 5

Convergence behavior of all algorithms on PenglungEW, Tic-tac-toe, Wine, and KrVsKpEW datasets.

Convergence behavior of all algorithms on PenglungEW, Tic-tac-toe, Wine, and KrVsKpEW datasets. Considering the outstanding and competitive performance of the BDMSAO and BDMO algorithms, as reported in previous paragraphs, the plots for accuracy are concentrated only on the two algorithms for convenient comparative analysis. In Figs. 6, 7, 8 and 9. A plot for accuracy against iteration for all datasets is presented for BDMSAO and BDMO. In Fig. 6, the accuracy obtained for BDMSAO for all the five datasets indicates that it performs better than the base algorithm, which is the BDMO. Where we see the curve for BDMSAO rising, that of BDMO was dropping to rise in experimentation with some datasets. Meanwhile, all accuracies for the datasets considered rose above 0.9, with those of Exactly and Vote running on 1.0 accuracies for all iterations. Figure 7 shows the curve for accuracies obtained for HeartEW, BreastEW, BreastCancer, and Lymphography datasets over all iterations. Similar to the previous discussion, we see the hybrid method of BDMSAO doing well in all datasets when compared with BDMO.
Figure 6

Accuracy plot of the BDMSAO and BDMO on the ionosphere, congressEW, Exactly, Exactly2, and Vote datasets.

Figure 7

Accuracy plot of the BDMSAO and BDMO on HeartEW, BreastEW, BreastCancer, and Lymphography datasets.

Figure 8

Accuracy plot of the BDMSAO and BDMO on waveformEW, sonar, SpaceEW, M-of-n, and Zoo datasets.

Figure 9

Accuracy plot of the BDMSAO and BDMO on PenglungEW, Tic-tac-toe, Wine, and KrVsKpEW datasets.

Accuracy plot of the BDMSAO and BDMO on the ionosphere, congressEW, Exactly, Exactly2, and Vote datasets. Accuracy plot of the BDMSAO and BDMO on HeartEW, BreastEW, BreastCancer, and Lymphography datasets. Accuracy plot of the BDMSAO and BDMO on waveformEW, sonar, SpaceEW, M-of-n, and Zoo datasets. Accuracy plot of the BDMSAO and BDMO on PenglungEW, Tic-tac-toe, Wine, and KrVsKpEW datasets. The plot for the accuracy values obtained for the datasets on waveformEW, sonar, SpaceEW, M-of-n, and Zoo are shown in Fig. 8. There is a significant difference in the performance of BDMSAO when compared with its corresponding base algorithm. In most cases, the values for accuracy obtained were rising to 1.0, whereas those of BDMO were lying below 0.8. Again, this result demonstrates that the accuracy of the features selection result discussed in previous sections is consistent and represents an outstanding performance of the new hybrid method. Figure 9 shows a comparison of the curve for PenglungEW, Tic-tac-toe, Wine, and KrVsKpEW datasets. The BDMO performance in these datasets rose above what has been observed in other datasets, though it still lags behind that of the hybrid method, the BDMSAO proposed and implemented in this study.

Comparison

The discussion earlier clarifies that the BDMSAO outperforms the BDMO. In this subsection, the performance of the proposed hybrid method is compared with nine other state-of-the-art methods, out of which seven are hybrid methods. The algorithms that were compared with other methods like adaptive switching grey-whale optimizer (ASGW), social ski driver algorithm and late acceptance hill-climbing (SSDs + LAHC)[83], serial grey-whale optimizer (HSGW), embedded chaotic whale survival algorithm (ECWSA-4)[84], binary GA, random switching grey-whale optimizer (RSGW), electrical harmony-based metaheuristic (EHHM), BPSO, and binary simulated normal distribution optimizer (BSNDO)[1]. Given the results obtained in Table 5, we can conclude that the BDMSAO and BSNDO yield better results than other methods on 11 out of 18 datasets, which is 61.11%. Meanwhile, the BDMSAO produced a 100% accuracy on 9 out of 18 datasets (50%), while its competitor, the BSNDO, produced 100% accuracy on 8 out of 18 datasets (44.4%). On the Breastcancer dataset, the BDMSAO, BSNDO, and EHHM achieved 100% accuracy. The proposed BDMSAO produced the second-best result and BSNDO and SSDs + LAHC on the BreastEW dataset after ASGW and EHHM. On the CongressEW dataset, BDMSAO and EHHM came third after BSNDO & SSDs + LAHC and ASGW. Most of the methods in this study, including the BDMSAO, yield 100% accuracy on the Exactly dataset. On the Exactly2 dataset, the BDMSAO achieved the third best result after BSNDO and EHHM, which narrowly beat the BDMSAO by a 0.1% margin. The BDMSAO & BSNDO produced the third best result on HeartEW after HSGW & SSDs + LAHC. On IonosphereEW, BDMSAO also generated the third best result after EHHM & ASGW, which narrowly beat our method by 0.06%. The BDMSAO yielded the best result of 98.75% on the KrvskpEW dataset. On Lymphography and Sonar datasets, the BDMSAO yielded 100% and 97.62% accuracy on SpecEW to lead other methods. On M-of-n, PenguinEW, Wine, and Zoo datasets, the BDMSAO achieved 100% accuracy with other methods. Finally, the BDMSAO achieved the third best result on WaveformEW and produced the fifth best in Tic-tac-toe. Classification accuracy of BDMSAO in comparison with other well-known feature selection methods tested using the UCI dataset, and the highlighted results indicate the highest classification accuracy. Table 6 shows the performance of the BDMSAO and other feature selection methods used in this study regarding the number of features selected. This proposed approach selected the least number of features on the Tic-tac-toe dataset. It also selected the least features on the CongressEW dataset along with BPSO and the Vote dataset with BPSO & BSNDO. It performed second in selecting the least features on Breastcancer, BreastEW, Exactly, HeartEW, and M-of-n datasets, while it performed third on Exactly2 datasets. It attained the fourth best results on Lymphography, WaveformEW, Wine, & Zoo, and achieved fifth on IonosphereEW, KrvskpEW, PenguinEW, and SonarEW datasets.
Table 6

The number of features selected by BDMSAO compared with other well-known feature selection methods tested using the UCI dataset. The highlighted results indicate the least number of feature subsets selected.

NoDatasetsBDMSAOASGWBSNDOHSGWBPSORSGWSSDs + LAHCECWSA-4)EHHM
1Breastcancer44.8674545.9332.547
2BreastEW715.833416.667917.591315
3CongressEW38.83378.86739.75.575
4Exactly66.86766.767.1677
5Exactly267.93389.03319.2859
6HeartEW46.36748.76736.133589
7IonosphereEW1917.31618.167720.512710
8KrvskpEW1924.52224.81224.8201516
9Lymphography1011.2510.567510.5676.5610
10M-of-n66.86766.867.1675
11PengiunEW132170.3187135.33130181.21407493
12SonarEW2635.52734.32236.43323.52223
13SpectEW2210.167610.233613.39117
14Tic-tac-toe478767968
15Vote38.96737.56738.84.556
16WaveformEW2425.8333326.9331527.53322.52015
17Wine45.93334.53355.867317
18Zoo57.655.53355.34.517
The number of features selected by BDMSAO compared with other well-known feature selection methods tested using the UCI dataset. The highlighted results indicate the least number of feature subsets selected.

Statistical test

In Table 7, the Friedman mean ranking test results are shown in bold values as the best-ranked algorithm. In most cases, the BPSO ranked highest, as seen in the table, and our proposed method ranked second in all cases where the BPSO ranked first. However, the BDMSAO ranked better than the BPSO on Exactly and Exactly2 datasets and then tied with the BPSO on two datasets i.e. PenguinEW and Sonar. It also outranks other methods (that are hybrid methods as our proposed method) in the ranking. This statistically shows the performance significance of the BDMSAO over other algorithms.
Table 7

Friedman mean ranking test.

NoDatasetsBDMSAOASGWBSNDOHSGWBPSORSGW
1Breastcancer5.001.002.303.505.903.20
2BreastEW5.401.002.503.305.503.30
3CongressEW5.401.002.203.105.603.70
4Exactly5.401.002.003.855.303.45
5Exactly24.501.002.106.004.403.00
6HeartEW5.301.002.203.205.703.60
7IonosphereEW5.201.002.403.605.703.10
8KrvskpEW5.001.002.303.405.903.40
9Lymphography5.101.002.103.805.703.30
10M-of-n5.001.002.003.306.003.70
11PengiunEW5.501.003.053.005.502.95
12SonarEW5.401.002.353.355.403.50
13SpectEW5.201.002.303.405.803.30
14Tic-tac-toe5.101.002.303.705.603.30
15Vote5.101.002.803.105.903.10
16WaveformEW5.101.002.003.405.903.60
17Wine5.001.003.003.005.703.30
18Zoo5.101.002.153.655.903.20
Friedman mean ranking test. The statistical significance of the BDMSAO and other algorithms are the same on average on all measures on most of the datasets used, producing significant values less than 0.05. The 0.05 is the representation of the significant level of 5%, which is used in the acceptance of the null value. As our proposed method generated more values that are less than 0.05 than all other algorithms in all datasets except for the BPSO. Therefore, this validates the fact that the samples were produced from a continuous distribution having the same medians as opposed to the null hypothesis that does not. This gives convincing proof that the results obtained by the BDMSAO are statistically significant compared to other similar methods. Table 8 presents the Wilcoxon mean rank test.
Table 8

Wilcoxon mean rank test.

NoDatasetsASGW-BDMSAOBSNDO- BDMSAOHSGW- BDMSAOBPSO- BDMSAORSGW- BDMSAO
1Breastcancer0.0050.0050.0050.0740.005
2BreastEW0.0050.0050.0050.0580.005
3CongressEW0.0050.0050.0050.5080.005
4Exactly0.0050.0050.0050.1690.005
5Exactly20.0050.0050.0050.2020.005
6HeartEW0.0050.0050.0050.0470.005
7IonosphereEW0.0050.0050.0050.3860.005
8KrvskpEW0.0050.0070.0050.0070.005
9Lymphography0.0050.0050.0050.0740.005
10M-of-n0.0050.0050.0050.0050.005
11PengiunEW0.0050.0050.0050.6450.005
12SonarEW0.0050.0050.0050.5080.005
13SpectEW0.0050.0050.0050.3860.005
14Tic-tac-toe0.0050.0090.0050.0490.005
15Vote0.0050.0050.0050.0740.005
16WaveformEW0.0050.0050.0050.0090.005
17Wine0.0050.0070.0050.2850.007
18Zoo0.0050.0050.0050.0070.005
Wilcoxon mean rank test.

Testing on high-dimensional datasets

The results discussed above revealed the performance of the BDMSAO over other well-known algorithms used in this study. To evaluate the robustness of this proposed algorithm, we tested its application to three high-dimensional datasets known to be extremely challenging. The dataset's description is provided in Table 9. The efficacy of the BDMSAO is also proven in comparison with eight (8) well-known state-of-the-art feature selection methods mentioned in Table 10. The number of features selected by BDMSAO compared with some popular FS selection methods is shown in Figs. 10, 11, and 12. All algorithms, including the BDMSAO, yielded the highest classification accuracy on both high-dimensional datasets. The BDMSAO selected the least number of features on the colon and leukemia datasets, respectively, compared to those achieved by the AIEOU, SFO and BSNDO, thus confirming the proposed method's ability to select the least features as indicated previously. However, the BDMSAO selected the third best number of features after RTHS and AIEOU on the Prostate_GE dataset.
Table 9

High-dimensional datasets and their properties.

NumberDatasets# features# instances# ClassesCategories
1Colon2000622Biological
2Leukemia7070722Biological
3Prostate_GE59661022Biological
Table 10

Classification accuracy produced by BDMSAO and other FS methods on high-dimensional datasets.

DatasetsBDMSAOAIEOUASGWBSNDOBWOAHSGWRSGWRTHSSFO
Colon100100100100100100100100100
Leukemia100100100100100100100100100
Prostate_GE100100100100100100100100NA
Figure 10

The number of features selected by BDMSAO and other FS methods on the Colon dataset.

Figure 11

The number of features selected by BDMSAO and other FS methods on Leukemia.

Figure 12

The number of features selected by BDMSAO and other FS methods on the Prostate_GE dataset.

High-dimensional datasets and their properties. Classification accuracy produced by BDMSAO and other FS methods on high-dimensional datasets. The number of features selected by BDMSAO and other FS methods on the Colon dataset. The number of features selected by BDMSAO and other FS methods on Leukemia. The number of features selected by BDMSAO and other FS methods on the Prostate_GE dataset.

Conclusion and future work

This paper proposed a new hybridized feature selection problem-solving method called the BDMSAO. The hybridization concept emanates from the methodological enhancement of the standard BDMO and SA algorithms. The developed BDMSAO utilized the SA as a local search method to enhance the exploitation of the BDMO and aid a suitable balance between the exploitation and exploration of the hybrid method. Interestingly, the BDMSAO accomplished a substantial enhancement in solving feature selection problems regarding classification accuracy achieved against the BDMO and other well-known state-of-the-art algorithms used for comparison in this study. The performances of the proposed approaches were assessed and compared against nine other feature selection methods, including the ASGW, BSNDO, HSGW, BPSO, BGA, RSGW, SSDs + LAHC, ECWSA-4, and EHHM, respectively. The evaluation criteria reported for each approach include the classification accuracy, average feature selection size number, and the respective algorithms' convergence characteristics. Similarly, the BDMSAO was compared against the BDMO algorithm to ascertain the validity of the initial enhancement claim over the BDMO. This developed feature selection approach was mainly evaluated and validated on some UCI datasets confirmed to be challenging. The new method was also tested using three high-dimensional datasets to prove its robustness in finding reasonable solutions to real-world problems that are often considered complex and difficult to solve using conventional methods. The results obtained by the BDMSAO indicate that the proposed method is applicable in various publicly available datasets. A limitation of this study may be in the computation complexity due to the addition of a local search technique. In the future, it would be interesting to consider hybridizing the BDMO with other state-of-the-art metaheuristics such as the GA, PSO, CSO, GWO, PDO and KHA algorithms. Also, it will be worth considering employing the hybrid BDMSAO algorithm in other real-world problem areas like image processing, facial recognition, and text classification, where researchers utilize a huge dimensional vector of features without knowing the reputation of every dataset features.
  7 in total

1.  Hybrid genetic algorithms for feature selection.

Authors:  Il-Seok Oh; Jin-Seon Lee; Byung-Ro Moon
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  2004-11       Impact factor: 6.226

2.  Optimization by simulated annealing.

Authors:  S Kirkpatrick; C D Gelatt; M P Vecchi
Journal:  Science       Date:  1983-05-13       Impact factor: 47.728

Review 3.  A review of feature selection methods in medical applications.

Authors:  Beatriz Remeseiro; Veronica Bolon-Canedo
Journal:  Comput Biol Med       Date:  2019-07-31       Impact factor: 4.589

4.  Ant-cuckoo colony optimization for feature selection in digital mammogram.

Authors:  J B Jona; N Nagaveni
Journal:  Pak J Biol Sci       Date:  2014-01-15

5.  A hybrid genetic-simulated annealing algorithm for the location-inventory-routing problem considering returns under e-supply chain environment.

Authors:  Yanhui Li; Hao Guo; Lin Wang; Jing Fu
Journal:  ScientificWorldJournal       Date:  2013-12-29

6.  Binary Simulated Normal Distribution Optimizer for feature selection: Theory and application in COVID-19 datasets.

Authors:  Shameem Ahmed; Khalid Hassan Sheikh; Seyedali Mirjalili; Ram Sarkar
Journal:  Expert Syst Appl       Date:  2022-03-15       Impact factor: 8.665

7.  Hybrid Binary Dragonfly Algorithm with Simulated Annealing for Feature Selection.

Authors:  Hamouda Chantar; Mohammad Tubishat; Mansour Essgaer; Seyedali Mirjalili
Journal:  SN Comput Sci       Date:  2021-05-25
  7 in total
  1 in total

1.  Binary dwarf mongoose optimizer for solving high-dimensional feature selection problems.

Authors:  Olatunji A Akinola; Jeffrey O Agushaka; Absalom E Ezugwu
Journal:  PLoS One       Date:  2022-10-06       Impact factor: 3.752

  1 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.