Literature DB >> 36089930

A Novel Method Based on Nonlinear Binary Grasshopper Whale Optimization Algorithm for Feature Selection.

Lingling Fang1, Xiyue Liang1.   

Abstract

Feature Selection (FS) is considered as an important preprocessing step in data mining and is used to remove redundant or unrelated features from high-dimensional data. Most optimization algorithms for FS problems are not balanced in search. A hybrid algorithm called nonlinear binary grasshopper whale optimization algorithm (NL-BGWOA) is proposed to solve the problem in this paper. In the proposed method, a new position updating strategy combining the position changes of whales and grasshoppers population is expressed, which optimizes the diversity of searching in the target domain. Ten distinct high-dimensional UCI datasets, the multi-modal Parkinson's speech datasets, and the COVID-19 symptom dataset are used to validate the proposed method. It has been demonstrated that the proposed NL-BGWOA performs well across most of high-dimensional datasets, which shows a high accuracy rate of up to 0.9895. Furthermore, the experimental results on the medical datasets also demonstrate the advantages of the proposed method in actual FS problem, including accuracy, size of feature subsets, and fitness with best values of 0.913, 5.7, and 0.0873, respectively. The results reveal that the proposed NL-BGWOA has comprehensive superiority in solving the FS problem of high-dimensional data. © Jilin University 2022, Springer Nature or its licensor holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Entities:  

Keywords:  Biomimetic position updating strategy; Feature selection; High-dimensional UCI datasets; Hybrid bionic optimization algorithm; Multi-modal medical datasets; Nature-inspired algorithm

Year:  2022        PMID: 36089930      PMCID: PMC9449924          DOI: 10.1007/s42235-022-00253-6

Source DB:  PubMed          Journal:  J Bionic Eng        ISSN: 1672-6529            Impact factor:   2.995


Introduction

Nowadays, with the rapid accumulation of massive data in different fields, data are being produced at a never seen before scale in dimensions [1]. High-dimensional data usually contain an increasing amount of information, but many unrelated or weakly correlated features also exist, which affect the data processing [2]. Therefore, it is urgent to develop effective data mining technology to reduce the dimension of high-dimensional data in various fields, such as medicine, bioinformatics, text mining, and internet of drones [3, 4]. FS is an effective data dimensional reduction method, which has been widely used and plays an important role in machine learning and pattern recognition fields [5]. By reducing the dimension of the dataset, it improves the calculation speed of the model. As a preprocessing step, FS is essentially a combinatorial optimization problem [6]. The purpose of FS aims to remove irrelevant and redundant attributes of specific datasets [7]. Different FS methods can be selected for data processing due to different learning algorithms. When dealing with high-dimensional data, the search strategy in the FS method is particularly important [8]. Recently, meta-heuristics with various search strategies have been largely employed to solve FS problems [9]. With the global search ability, the Swarm Intelligence (SI)-based heuristic search methods can better alleviate the time complexity and computing costs in FS [10]. The SI optimization algorithms are inspired by natural animal communities’ behavior and simulate the foraging and predation law continuously [11]. At present, many popular SI algorithms have been employed to solve the FS problems, such as: Genetic Algorithm (GA) [12], Particle Swarm Optimization (PSO) [13], and Grey Wolf Optimization algorithm (GWO) [14]. Recently, some algorithms including Whale Optimization Algorithm (WOA) [15], Butterfly Optimization Algorithm (BOA) [16], and Grasshopper Optimization Algorithm (GOA) are proposed [17]. These optimization algorithms based on population generate a set of candidate solutions in each run. This can lead the better consequences in FS. Similarly, Abualigah et al. propose a novel nature-inspired meta-heuristic optimizer, called Reptile Search Algorithm (RSA), which is motivated by the hunting behavior of crocodiles [18]. The Dwarf Mongoose Optimization algorithm (DMO) algorithm is proposed to mimic the foraging behavior of the dwarf mongoose and displays the bionic behavior for searching [19]. Using a bionic disease transmission strategy, a meta-heuristic algorithm named Ebola Optimization Search Algorithm (EOSA) is proposed, which is based on the propagation mechanism of the Ebola virus disease [20]. Moreover, the distributed behavior of the major arithmetic operators can also establish a meta-heuristic approach to optimization problems, which is called Arithmetic Optimization Algorithm (AOA) [21]. Besides, a population-based optimization method is proposed, called Aquila Optimizer (AO), which is inspired by the natural behavior of aquila during prey capture [22]. Table 1 describes the bionic behavior differences of some nature-inspired meta-heuristic algorithms.
Table 1

Differences in bionic behavior between some nature-inspired meta-heuristic algorithms

AlgorithmInspirationReferences
Aquila Optimizer (AO)Aquila Bird[22]
Arithmetic Optimization Algorithm (AOA)Arithmetic Operators[21]
Butterfly Optimization Algorithm (BOA)Food Search and Mating Behavior of Butterflies[16]
Dwarf Mongoose Optimization algorithm (DMO)Behavior of The Dwarf Mongoose[19]
Ebola Optimization Search Algorithm (EOSA)Ebola Virus[20]
Genetic Algorithm (GA)Evolutionary Biology[12]
Grasshopper Optimization Algorithm (GOA)Foraging and Swarming Behavior of Grasshoppers[17]
Grey Wolf Optimization (GWO)Hunting Process of Grey Wolves[14]
Particle Swarm Optimization (PSO)Simplified Social Model[13]
Reptile Search Algorithm (RSA)Behavior of Crocodiles[18]
Whale Optimization Algorithm (WOA)Social Behavior of Humpback Whales[15]
Differences in bionic behavior between some nature-inspired meta-heuristic algorithms However, considering that many optimization problems are set in a binary space, the above SI optimization algorithms cannot meet this condition. Here, calculating the optimal position is the key to completing FS problems when employing SI optimization algorithms. Besides, appropriately balancing exploration and exploitation will lead to the improvement of the search algorithm’s performance. Therefore, how to explore the space of the search and exploit the optimal solutions are two contradictory principles to be considered [23]. To deal well with this issue, the majority of SI optimization algorithms have been further optimized. It is a common method to improve the location update in a binary way and redefine the space of position updating. A binary PSO (BPSO) in which a particle moves in a state space restricted to zero and one on each dimension is proposed [24]. To solve the diverse FS problems, Hussien et al. design two novel binary variants of the WOA called BWOA-S and BWOA-V [25]. By combining a mutation operator into the GOA, a novel version called binary GOA (BGOA) is proposed [26]. The BGOA can enhance the exploratory behavior of the original GOA. Although binary variant algorithms can solve the binary optimization problem in feature subset selection of high-dimensional datasets, it still makes errors because of careless search facing complex FS problems. New solutions are still urgently needed to solve the search problem in FS. The combination of the advantages of SI optimization algorithms can also balance the relationship between exploration and exploitation. Hybrid algorithms have attracted more and more attention in the field of optimization algorithms [27]. Table 2 summarizes the relevant studies on hybrid optimization algorithms in recent years. Marfarja et al. propose a hybrid WOA and simulated annealing (SA) algorithm in a wrapper feature selection method to enhance the exploitation of the WOA [28]. Al-Tashi et al. improve a binary version of the hybrid PSO and GWO as an FS method, which employs suitable operators to solve binary problems [29]. To surmount the inconveniences in dimensional space and non-informational peculiarities, Purushothaman et al. introduce a hybrid between the GWO and the GOA for text feature selection and clustering [30]. Besides the simple combination in the hybrid algorithms, some studies have improved the bionic variation by improving the coefficients in the function.
Table 2

Some hybrid optimization algorithms

ReferencesHybrid methodsAbbreviation
[28]WOA and SAWOASA(T)-1, 2
[29]GWO and PSOBGWOPSO
[30]GWO and GOAGWO–GOA
[31]MA and KHAMAKHA
[32]SOA and TEOSOA-TEO1, 2, 3
[33]MPMD and WOAMPMDIWOA
[34]GWO and CSAGWOCSA
[35]BA and PSOHBBEPSO
[36]CRO and SABCROSAT
[37]ACO and BCOAC-ABC
Some hybrid optimization algorithms Hence, a system for FS based on a hybrid Monkey Algorithm (MA) with Krill Herd Algorithm (KHA) is proposed [31]. The fitness function of the proposed algorithm incorporates both classification accuracy and feature reduction size. Jia H et al. propose three hybrid algorithms to solve FS problems based on Seagull Optimization Algorithm (SOA) and Thermal Exchange Optimization (TEO) [32]. A hybrid feature subset selection algorithm called the Maximum Pearson’s Maximum Distance Improved Whale Optimization Algorithm (MPMDIWOA) is proposed by designing two parameters to adjust the weights of the relevance and redundancy [33]. Hybrid bionic optimization algorithms perform well in feature subset selection, but there is no concrete practice in practical application. Recently, some practical feature selection problems have begun to focus on using hybrid algorithms. Considering the strengths of GWO and Crow Search Algorithm (CSA), a hybrid GWO with CSA, namely GWOCSA is proposed, which aims to generate promising candidate solutions and achieve global optima efficiently [34]. A new hybrid binary version of bat (BA) and enhanced PSO is proposed to solve FS problems, called Hybrid Binary Bat Enhanced Particle Swarm Optimization Algorithm (HBBEPSO) [35]. It combines the bat algorithm with its capacity for echolocation into the version of the particle swarm optimization. Moreover, Cy et al. introduce a hybrid model called BCROSAT for FS in high-dimensional biomedical datasets, in which the tournament selection mechanism and SA algorithm are combined with Binary Coral Reef Optimization (BCRO) [36]. Shunmugapriya and Kanmani propose a novel swarm-based hybrid algorithm named AC-ABC Hybrid, which combines the characteristics of Ant Colony Optimization (ACO) and Artificial Bee Colony (ABC) algorithms to optimize FS [37]. Although these hybrid bionic optimization algorithms solve the feature subset selection in some problems, they still fall into local optimum in high-dimensional datasets. In brief, the SI optimization algorithm has an excellent performance in FS, it is easy to fall into the local optimal condition in the process of feature selection and fail to select the most representative feature subset. Therefore, how to solve the defect and select the least feature subset is a crucial problem [38]. To improve this problem existing in most optimization algorithms, this paper proposes a nonlinear binary grasshopper whale optimization algorithm, namely NL-BGWOA, which combines the whale individual updating method in WOA with GOA and optimizes the position updating strategy. The proposed NL-BGWOA incorporates the adaptive weight and nonlinear adjustment coefficients. To verify the reliability of the proposed NL-BGWOA, it is compared with several other state-of-the-art algorithms in FS experiments, such as GOA, WOA, BGOA, and BWOA. Moreover, the proposed algorithm is also tested on high-dimensional UCI datasets, multi-modal Parkinson's speech datasets, and COVID-19 symptom dataset, whose experimental results are evaluated. The rest of the paper is organized as follows. Section 2 gives an outline of the standard GOA, WOA, and the corresponding binary version of them. Section 3 presents the proposed NL-BGWOA algorithm and its procedures. Section 4 describes the experimental design, results, and analysis of the basic FS problem and the actual problem in medical datasets. Finally, Sect. 5 draws some conclusions for this paper and presents the further work.

Background

Grasshopper Optimization Algorithm

The Original Algorithm

In nature, grasshoppers often gather in large-scale ways to prey. The characteristics of grasshoppers are related to their movement. The swarm in the nymph phase is characterized by slow movement with small steps by the grasshoppers. In contrast, the swarm in the adult phase is characterized by abrupt and long-distance movement, which corresponds to the exploration and exploitation phase of the nature-inspired algorithms. The grasshopper optimization algorithm gives a mathematical model that simulates the movement of grasshopper populations:where is the distance between the ith and the jth grasshopper: ; s is the strength of social forces: , where f indicates the intensity of attraction and l is the attractive length scale; and are the upper and lower bound, respectively; is the value of the best solution, and is a decreasing coefficient to shrink the comfort zone, repulsion zone, and attraction zone. The parameter is the core parameter of the GOA algorithm, and the parameter update strategy has an important influence on the convergence performance of the algorithm. The linearly decreasing parameter can realize the process of the algorithm from exploration to development and the ability to reduce the comfort zone between grasshoppers.where is the maximum value, is the minimum value, the value of is 1 or 2, indicates the current iteration, and is the maximum number of iterations. The outer is used to reduce the search coverage toward the target grasshopper as the iteration count increases, while the inner is used to reduce the effect of the attraction and repulsion forces between grasshoppers proportionally to the number of iterations.

Binary Grasshopper Optimization Algorithm

The grasshopper optimization algorithm has strong local development capabilities, so it has good performance in solving continuous optimization problems. However, according to the nature of the FS problem, the search space can be represented by binary values [0,1]. Since binary operators are expected to be much simpler than continuous counterparts, the GOA algorithm needs to be discretized. Therefore, Majdi Mafarja et al. propose a method by modifying the population update strategy of the GOA algorithm to obtain the Binary Grasshopper Optimization Algorithm (BGOA). Here, sigmoidal and hyperbolic tan functions are two transfer functions, which play key roles in BGOA. This paper takes the sigmoidal function as an example: The position of the current grasshopper will be updated based on the probability value :where represents the dth dimension of the grasshopper in the next iteration and is a random number in the range . A conceptual model of the interactions between grasshoppers and the comfort zone is illustrated in Fig. 1a.
Fig. 1

Behavior of individual grasshoppers and whale

Behavior of individual grasshoppers and whale

Whale Optimization Algorithm

The WOA is a SI optimization algorithm that imitates humpback whale predation, which includes the encircling prey phase, the exploitation phase, and the exploration phase. During the encircling prey phase, individual whales can identify the location of prey and enclose them. The corresponding computation is as follows:where is the number of iteration. is the optimal whale position and is the current whale position. is the corresponding step size. In the exploitation phase, the WOA designs two ways to simulate the bubble-net attacking method of whales. The first way is similar to the encircling prey phase when is established. The second way is called the spiral updating position, where the whale moves to the current optimal individual in a spiral movement. The corresponding process for position updating is defined as follows: where is the distance between the individual whale and the current optimal whale. is the logarithmic spiral shape constant, in general, is set to 1. is a random number in interval . In conclusion, the whale position can be updated by the above two cases: where is a random number in , which represents a probability value. When is less than 0.5, the whales use the first case to shrink and surround. Otherwise, the whales use the second case to spirally update their position. During the exploration phase, if there is no optimal prey in the shrinking range, the whale will jump out of the range and search randomly. When is true, the mathematical model is as follows: where is the position of a random whale in the current population.

Binary Whale Optimization Algorithm

The WOA has a simple structure and uses a few coefficients, so it is often used for various situations. However, the original WOA may exist local optimization. Therefore, the binary WOA for the feature selection redefined the space of position updating. In binary WOA, the positions of the individual whale in solution space are restricted to the binary space :where is the number in the interval . is the updated binary position at iteration. is a combination of the position updating and the sigmodal function, which is defined as follows: The general movement of the individual whale as it hunts and preys through the bubble-net attacking method is illustrated in Fig. 1b.

The Proposed Method

According to the fundamental purpose of the FS problem, the proposed NL-BGWOA finds the least optimal feature subset from the original datasets, improves the accuracy of the optimization algorithm, and simplifies the data processing. Besides, the proposed method improves the nonlinear decline coefficient and the adaptive weight that changes in the iteration process. The specific improvement contents are as follows.

The Related Work

(1) The nonlinear coefficient The swarm intelligence optimization algorithm is mainly divided into two stages: the exploration stage and the exploration stage. These two important stages determine the performance of the algorithm, so how to effectively balance them is a key consideration. However, there is a major problem in most swarm intelligence optimization algorithms: the core coefficient decreases linearly at a constant rate, which will slow down the convergence speed and easily make the algorithm fall into a local optimum. To tackle the above problem, this paper proposes a nonlinear change strategy, which adjusts the linear change of the coefficient by combining the change trend of the cosine function. The proposed nonlinear coefficient is described by the following equation: where and is the maximum and minimum, the value of is 1 or 2, respectively, indicates the current iteration, and is the maximum number of iterations. The proposed nonlinear coefficient speeds up the search speed in the exploitation stage, which enables the individual to obtain the target quickly. During the exploration phase, the speed of the coefficient slows down, so the population can carefully search the surrounding space. Therefore, the proposed method can effectively avoid falling into the local optimum and better balance the global and local search. The comparison between the nonlinear and linear coefficients is shown in Fig. 2.
Fig. 2

The comparison between the nonlinear and linear coefficients

The comparison between the nonlinear and linear coefficients 2. The adaptive weight The movement of individuals in the swarm intelligence optimization algorithm during position updating is also very important. During iterating, it should be considered that the position updating method in different iteration cycles and the degree of dependence between individuals. In this paper, the adaptive weight is integrated into the updating position, which is defined as follows:where is the number of current iteration and is the number of the maximum iterations. The position update after the adoption of adaptive weights will dynamically adjust the size of the weights according to the increase in the number of iterations, so that the current optimal position guides individuals differently. As the number of iterations increases, the individual with a larger weight will speed up the movement, which speeds up the convergence speed. As shown in Fig. 3, different has a great influence on the adaptive weight . The weight is large at the beginning and gradually become smaller, which is affected by the number of the maximum iteration . By comparison, it can be seen that when is 100, the condition can be met simultaneously. The optimal is highlighted in green.
Fig. 3

Comparison of adaptive weight with different

Comparison of adaptive weight with different

The Proposed NL-BGWOA

The purpose of the feature selection is to select relevant features and eliminate redundant ones. The BWOA remains the better performance in low-dimensional datasets, but it cannot select the effective features to handle the high-dimensional datasets. The BGOA has certain advantages in processing high-dimensional data, which can make up for the deficiencies of BWOA. Therefore, this paper proposes a novel algorithm named NL-BGWOA that combines the advantages of BGOA and BWOA. The proposed method can select the representative salient features of the adopted datasets and can find more optimal solutions in the search space. Concretely, features are selected from generated features that are potential solutions in the search space. The salient features are further selected by the NL-BGWOA to produce more representative and significant feature subsets. In the proposed NL-BGWOA, multiple influences are analyzed for optimal locations, with the aim of conducting more detailed searches in high-dimensional datasets. The proposed method can improve the efficiency of the search. Besides, the optimization between the current optimal position and the next one during each search can improve the accuracy of FS. Figure 4 shows the process of the proposed NL-BGWOA.
Fig. 4

The proposed NL-BGWOA

The proposed NL-BGWOA (1) Mathematical model of the proposed method In the proposed method, the selection of the significant feature subsets focuses on the updating position. Here, the new position update strategy is transformed by changing the calculation of the distance between individuals in the force function. In the position update strategy, the position of the individual in a potential search space depends on a variety of factors: Here, is the gravity, is the wind force, and is the interaction force between two individuals in populations: where is the number of individuals. is the distance between the individual to the individual: . is the corresponding unit vector, which is set by . Consequently, the individual position is determined by the position, distance, force, and the last optimal position between the populations. The position update model in the NL-BGWOA algorithm is defined as: In the proposed method, is the optimal position after selection, and are the two nonlinear coefficients used to balance the search, and are the upper and lower bound, respectively. or is initial position vector of the NL-BGWOA. or is the current position of the next optimal position in NL-BGWOA. is the adaptive weight in iteration, is the value of the best solution. In this paper, and are expressed by Eqs. 7 and 8, respectively. (2) The corresponding flowchart and pseudo-procedure The proposed method can expand its searching capability and locate the near global optimum solution. The architecture and the corresponding steps of the proposed NL-BGWOA are explained in Fig. 5. The specific steps of the proposed method are as follows:
Fig. 5

The overall structure of the proposed NL-BGWOA

The overall structure of the proposed NL-BGWOA Step 1: Initialize parameters , , , , and . Step 2: Initialize the positions of the search agents and the number of iterations. Step 3: Sort by the fitness of each individual and calculate the adaptive weight . Step 4: According to the probability number , update the position using the shrinking encircling mechanism or the spiral equation. Step 5: If the iteration gets to the maximum, then output the result. Otherwise, go to Step 3. Step 6: Randomly reinitialize the feature subset. Step 7: Calculate the fitness of initial features and keep the recent best fitness value and position in NL-BGWOA. Step 8: Enter the iterative process of NL-BGWOA, update the optimal solution, and return the best features. Step 9: Update the current optimal position and fitness value, then return the optimal features. Step 10: If the iteration gets to the maximum, then output the result. Otherwise, go to Step 6. The corresponding pseudo-code is illustrated as follows: The proposed NL-BGWOA fully integrates the strengths of BWOA and BGOA and tackles the FS of high-dimensional data. Besides, the proposed method expresses the datasets in the iteration and computes the fewer features. Therefore, it can guarantee the goodness of the feature subsets and increase the performance of the optimization algorithm.

Experimental Results and Evaluation

Data Description and Parameters’ Setting

To verify the performance of the proposed algorithm and extensively investigate the application, other state-of-the-art algorithms are selected for comparison experiments with the proposed algorithm. By analyzing some results in the feature selection experiment, such as the accuracy rate and the number of feature selections, the advantages between the proposed algorithm and other algorithms can be compared. The corresponding parameter settings are shown in Table 3. Each algorithm is conducted in 20 independent experiments with the following parameter settings.
Table 3

Parameter settings of the algorithms used for comparison in the current study

AlgorithmParameterValues
PSO(BPSO)Number of particles10
Maximum number of iterations100
Inertia weight \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$w$$\end{document}w in PSO(BPSO)1
Acceleration constants in PSO(BPSO)[2,2]
WOA(BWOA)Number of whales10
Maximum number of iterations100
Classification quality coefficient \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\alpha$$\end{document}α in WOA(BGOA)0.99
GOA(BGOA)Number of grasshoppers10
Maximum number of iterations100
Linear decrease coefficient \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$c_{{}}^{\max }$$\end{document}cmax and \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$c_{{}}^{\min }$$\end{document}cmin(1, 1.0E−04)
Classification quality coefficient \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\alpha$$\end{document}α in GOA(BGOA)0.99
NL-BGWOANumber of individuals10
Maximum number of iterations100
Classification quality coefficient \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\alpha$$\end{document}α in NL-BGWOA0.99
Linear decrease coefficient \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$c_{{}}^{\max }$$\end{document}cmax and \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$c_{{}}^{\min }$$\end{document}cmin(1, 1.0E−04)
Parameter settings of the algorithms used for comparison in the current study Here, several basic feature selection experiments have been performed over ten distinct high-dimensional UCI datasets [39]. The details of ten datasets are depicted in Table 4.
Table 4

Benchmark datasets

No. of datasetsNameNo. of featuresNo. of samples
D1Arrhythmia279452
D2BreastEW32596
D3Clean1165476
D4Clean21656598
D5Dermatology34366
D6Hill-Valley100606
D7LonosphereEW34351
D8SonarEW60208
D9Spambase574601
D10WaveformEW405000
Benchmark datasets Meanwhile, the multi-modal Parkinson's speech datasets from the UCI repository and COVID-19 symptom dataset are used to verify the specific practical problem of the proposed algorithm, which are described as follows: Here, the datasets P1 [40] and P2 [41] in Table 5 contain a group of linear and time–frequency-based features, measures of variation in fundamental frequency, amplitude, and so on. Especially for P1, besides the above fundamental features, the disease speech data set includes speech intensity, resonance frequencies, bandwidth-based features, and other specific features. The comprehensive multi-modal Parkinson's datasets can be referred to in Appendix A.
Table 5

Multi-modal Parkinson's disease speech datasets

No. of datasetsNo. of featuresNo. of samples
P1754756
P2261040
Multi-modal Parkinson's disease speech datasets Table 6 describes the information of the COVID-19 patient symptom dataset from Kaggle [42]. COVID-19 patient’s dataset is collected from a nearby hospital, where there are a total of 2575 positive and negative symptoms. The symptoms include age, fever, body pain, runny nose, difficult breath, and the infection probability of COVID-19 patients, of which the value is either 1 or 0. In difficult breathing, there are three types of values, which are no difficulty breathing problem, severe problem, and moderate problem.
Table 6

COVID-19 symptom dataset

No. of samplesLabelSymptoms
2575Boolean (COVID-19 positive or negative)Age, fever, body pain, runny nose, difficult breathing, and the infection probability of COVID-19 patients
COVID-19 symptom dataset

Evaluation Criteria

(a) Accuracy Accuracy is the criterion for comparing the classification accuracy when performing feature selection experiments, which is obtained from the 5-KNN classifier on the test data. Average accuracy is calculated in the total runs: where is the number of runs, is classification accuracy in the experiment. (b) The size of selected features subsets The size of selected feature subsets is another indicator to evaluate algorithms. It is an average value of the selected feature subsets when algorithms run times: where is the number of selected features in the experiment. (c) Fitness value The fitness value of the optimal solution is not only used to update the position of individuals, but also to evaluate the performance of the algorithms. It is the average of the fitness values acquired after running all the iterations of the optimization algorithms: where is the fitness value in the experiment.

Performance Evaluation

Feature Selection for Benchmark Datasets

In this paper, the 5-KNN classifier is used to measure the performance. All datasets are divided into two parts: 80% of the instances are devoted to training and 20% of the instances are used to test. Table 7 presents the average classification accuracy for feature selection after running 20 times. Note that the best results are highlighted in bold. From Table 7, it can be clearly seen that the proposed algorithm has advantages over other algorithms on the majority of high-dimensional datasets, especially D3 and D7.
Table 7

Mean classification accuracy

DatasetsAlgorithms
PSOWOAGOABPSOBWOABGOANL-BGWOA
D10.58410.59110.58330.58100.59770.59260.5994
D20.95880.95660.97370.95580.98230.98240.9823
D30.85490.91000.89310.90210.92050.91800.9258
D40.94650.97130.96450.96920.97230.96040.9725
D50.98630.98630.95340.98630.97860.98630.9895
D60.58160.55600.58180.55970.62370.60450.6320
D70.93070.94200.88930.92000.93710.91640.9443
D80.86670.89880.86670.86540.88050.90020.8857
D90.91050.88500.89340.87870.91600.91670.9239
D100.71200.80560.75820.76100.81200.81040.8125
Mean classification accuracy To describe the experimental results more intuitively, several representative datasets are selected in Fig. 6. Here, the red lines and the corresponding numbers represent the proposed NL-BGWOA. It can see the advantages of the proposed NL-BGWOA in classification accuracy compared with other algorithms.
Fig. 6

Mean classification accuracy for several representative datasets

Mean classification accuracy for several representative datasets Table 8 shows the average size of selected feature subsets on ten different high-dimensional datasets, where the best values are in bold. Meanwhile, several representative datasets are selected to represent the size of selected feature subsets, as shown in Fig. 7. As can be observed in Table 8 and Fig. 7, the proposed NL-BGWOA performs superior in the mean size of feature subsets on almost all the datasets.
Table 8

Mean size of selected feature subsets

DatasetsAlgorithms
PSOWOAGOABPSOBWOABGOANL-BGWOA
D1137.6143.0128.1140.0134.3173.9108.3
D214.713.512.814.810.210.95.6
D3104.986.878.782.164.986.635.8
D4109.47879.784.668.371.543.0
D519.722.117.318.320.515.811.9
D651.348.547.946.039.655.022.3
D714.713.615.415.05.113.13.2
D837.628.629.127.614.527.47.4
D929.931.22729.924.929.516.4
D1035.832.517.722.711.021.79.1
Fig. 7

Mean size of selected feature subsets for several representative datasets

Mean size of selected feature subsets Mean size of selected feature subsets for several representative datasets Especially on D3 and D9, the size of selected feature subsets obtained by the NL-BGWOA is far less than the results from other algorithms. This indicates that the proposed method can effectively reduce the number of features, select the most relevant optimal feature subsets, and implement the purpose of FS problems. The numbers in Fig. 7 specify the size of feature subsets selected by the proposed NL-BGWOA. Table 9 and Fig. 8 summarize the experimental results of the fitness for the proposed method. The best fitness values are in bold. Note that some representative datasets are also selected, as shown in Fig. 8.
Table 9

Mean fitness values

DatasetsAlgorithms
PSOWOAGOABPSOBWOABGOANL-BGWOA
D10.45690.44040.42070.41470.41130.41410.3934
D20.04120.04340.06590.04420.02050.01540.0205
D30.15080.17190.21700.17830.16360.21390.1215
D40.05960.03820.08000.03680.03710.07470.0336
D50.05490.02340.06970.06230.02150.05260.0138
D60.44680.44500.42430.44050.42560.41090.3614
D70.06930.05790.14490.08000.12680.17360.0563
D80.13820.11120.16840.13780.12070.23050.1143
D90.13050.12270.19110.19700.12410.12580.1061
D100.28710.19440.30630.22430.19650.22490.1876
Fig. 8

Mean fitness values comparison for several representative datasets

Mean fitness values Mean fitness values comparison for several representative datasets It can be observed from Table 9 and Fig. 8 that the majority of the best results highlighted in bold are obtained by the proposed NL-BGWOA. For instance, the NL-BGWOA provides average fitness values of 0.0336 and 0.0563 on D4 and D7, respectively. The numbers highlighted in Fig. 8 represent the values of fitness for the proposed algorithm. As can be seen in Table 9 and Fig. 8, the NL-BGWOA’s superior performance in mean fitness values demonstrates the ability to search and select the optimal objectives, which has a high potential to solve the promising problems. Overall, the average fitness values of NL-BGWOA prove the competency to efficiently find the optima in the search space.

Feature Selection Experiment for Medical Datasets

Table 10 and Fig. 9 show the results of three evaluation criteria obtained from the proposed method in comparison to other algorithms on the multi-modal Parkinson datasets, where the bold represents the best method. According to the results reported in Table 10 and Fig. 9, there is a significant contrast between the proposed NL-BGWOA and other algorithms, especially the number of selected features (Num-FS). The values of Num-FS are highlighted in Fig. 9, which are presented in the form of numbers around the corresponding algorithm.
Table 10

Evaluation of the current study on the multi-modal Parkinson datasets

AlgorithmsD1D2
AccuracyNum-FSFitnessAccuracyNum-FSFitness
PSO0.765378.60.23540.69413.60.3060
WOA0.782183.40.21790.66511.50.3351
GOA0.866365.50.23930.66511.10.3437
BPSO0.766371.10.23410.68813.10.3118
BWOA0.905180.70.09550.69510.50.3053
BGOA0.911485.70.09000.6929.20.3151
NL-BGWOA0.913111.70.08730.6985.70.3034
Fig. 9

The comparison of different algorithms on the multi-modal Parkinson datasets

Evaluation of the current study on the multi-modal Parkinson datasets The comparison of different algorithms on the multi-modal Parkinson datasets Similarly, the proposed NL-BGWOA is tested on the COVID-19 symptom dataset, with experimental results shown in Table 11 and Fig. 10. The results show that the proposed method can improve the accuracy, the number of feature subsets, and fitness. The best values are shown in bold in Table 11.  Especially the value of fitness and the number of feature subsets are improved significantly, which the optimal values are 1.0 and 0.4191, respectively. The numbers around the algorithms in Fig. 10 show a comparison using the number of selected features as an example.
Table 11

Evaluation of the COVID-19 symptom dataset

AlgorithmsAccuracyNum-FSFitness
PSO0.51842.00.4805
WOA0.50733.00.4629
GOA0.51882.00.4431
BPSO0.51673.00.4386
BWOA0.52043.00.4837
BGOA0.52012.00.4353
NL-BGWOA0.52171.00.4191
Fig. 10

The comparison of different algorithms on the COVID-19 symptom datasets

Evaluation of the COVID-19 symptom dataset The comparison of different algorithms on the COVID-19 symptom datasets The above experimental results indicate that the proposed NL-BGWOA has superior performance to the other algorithms in terms of three evaluation criteria during processing the medical datasets. The results mean that the proposed NL-BGWOA can tackle the foremost purpose of feature selection problems. Besides, it selects the features that can represent patients accurately. This application is also used the selected related features to determine whether the patient has a disease.

Statistical Test

Moreover, for further analysis of the proposed method, the statistical test is used to show the comparison statistically. It is a non-parametric test that provides a statistical value indicating whether the proposed method is significantly different from other methods. Table 12 summarizes the results of the Friedman test based on classification accuracy, with the optimal values in bold. From the above discussion and results, it can be observed that the best rank is given to NL-BGWOA corresponding to their mean rank, which is followed by BWOA, BGOA, BPSO, WOA, GOA, and PSO. It is concluded that the proposed NL-BGWOA has better performance on search quality.
Table 12

Results of Friedman test

AlgorithmsMeanRanking
PSO2.00007
WOA2.80775
GOA2.03856
BPSO3.20914
BWOA4.34622
BGOA4.07693
NL-BGWOA5.73081
Results of Friedman test

Conclusion

For the FS problem, BWOA and BGOA may select a large number of features and fall into optima. Therefore, they will add a certain degree of difficulty to subsequent data processing. In this paper, a novel FS method called NL-BGWOA is proposed, which integrates the strengths of BWOA and BGOA in the FS problem and tackles the FS of high-dimensional data. The proposed method expresses the datasets in the iteration by computing the fewer features, which can increase the performance of optimization algorithm and guarantee the goodness of the feature subsets. To verify the effectiveness of the proposed method, this paper conducts experiments on 10 benchmark datasets, multi-modal Parkinson datasets, and COVID-19 symptom datasets. The results demonstrated the superiority of the proposed method compared to most state-of-the-art algorithms in terms of the selected feature subsets. However, the classification accuracy and fitness still need to be improved facing datasets with fewer features, especially in various actual applications. As further work, the proposed method can be applied to more practices to solve the problems in the real life such as disease diagnosis, prediction, and engineering optimization problems.
No. of datasetsFeaturesDescription
P1Baseline features

Jitter variants

Shimmer variants

Fundamental frequency parameters

Harmonicity parameters

Recurrence period density entropy (RPDE)

Detrended fluctuation analysis (DFA)

Pitch period entropy (PPE)

Time frequency features

Intensity parameters

Formant frequencies

Bandwidth

Mel frequency cepstral coefficientsMFCCs
Wavelet transform based features

Wavelet transform (WT)

features related with F0

Vocal fold features

Glottis quotient (GQ)

Glottal to noise excitation (GNE)

Vocal fold excitation ratio (VFER)

Empirical mode decomposition (EMD)

P2Frequency features

Jitter (local)

Jitter (local, absolute)

Jitter (rap)

Jitter (ppq5)

Jitter (ddp)

Amplitude features

Shimmer (local)

Shimmer (local, dB)

Shimmer (apq3)

Shimmer (apq5)

Shimmer (apq11)

Shimmer (dda)

Harmonicity features

Autocorrelation

Noise-to-harmonic

Harmonic-to-noise

Pitch features

Median pitch

Mean pitch

Standard deviation

Minimum pitch

Maximum pitch

Pulse features

Number of pulses

Number of periods

Mean period

Standard deviation of period

Voicing features

Fraction of locally unvoiced frames

Number of voice breaks

Degree of voice breaks

  2 in total

1.  Collection and analysis of a Parkinson speech dataset with multiple types of sound recordings.

Authors:  Betul Erdogdu Sakar; M Erdem Isenkul; C Okan Sakar; Ahmet Sertbas; Fikret Gurgen; Sakir Delil; Hulya Apaydin; Olcay Kursun
Journal:  IEEE J Biomed Health Inform       Date:  2013-07       Impact factor: 5.772

2.  A novel unsupervised approach based on the hidden features of Deep Denoising Autoencoders for COVID-19 disease detection.

Authors:  Michele Scarpiniti; Sima Sarv Ahrabi; Enzo Baccarelli; Lorenzo Piazzo; Alireza Momenzadeh
Journal:  Expert Syst Appl       Date:  2021-12-16       Impact factor: 6.954

  2 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.