Literature DB >> 31137034

An improved adaptive memetic differential evolution optimization algorithms for data clustering problems.

Hossam M J Mustafa1, Masri Ayob1, Mohd Zakree Ahmad Nazri1, Graham Kendall2.   

Abstract

The performance of data clustering algorithms is mainly dependent on their ability to balance between the exploration and exploitation of the search process. Although some data clustering algorithms have achieved reasonable quality solutions for some datasets, their performance across real-life datasets could be improved. This paper proposes an adaptive memetic differential evolution optimisation algorithm (AMADE) for addressing data clustering problems. The memetic algorithm (MA) employs an adaptive differential evolution (DE) mutation strategy, which can offer superior mutation performance across many combinatorial and continuous problem domains. By hybridising an adaptive DE mutation operator with the MA, we propose that it can lead to faster convergence and better balance the exploration and exploitation of the search. We would also expect that the performance of AMADE to be better than MA and DE if executed separately. Our experimental results, based on several real-life benchmark datasets, shows that AMADE outperformed other compared clustering algorithms when compared using statistical analysis. We conclude that the hybridisation of MA and the adaptive DE is a suitable approach for addressing data clustering problems and can improve the balance between global exploration and local exploitation of the optimisation algorithm.

Entities:  

Mesh:

Year:  2019        PMID: 31137034      PMCID: PMC6538400          DOI: 10.1371/journal.pone.0216906

Source DB:  PubMed          Journal:  PLoS One        ISSN: 1932-6203            Impact factor:   3.240


Introduction

Data clustering is widely used in different applications to understand the structure of the data, to focus on a specific set of clusters for further analysis, and to detect the characteristics of each cluster. Data clustering has been developed and used as an essential tool for different disciplines, in areas such as Information Retrieval [1], the Internet of Things [2], Business [3], Medicine [4], and Image segmentation [5]. In recent times, clustering methods have been extensively studied [6-8]. The clustering methods can be classified based on the fitness function. The popular clustering methods are classified as partitioning clustering methods. The partitioning clustering methods attempt to divide the dataset into a set of disjoint clusters and try to optimise specific criterion function, which may emphasise the local structure of the data. The most popular partitioning clustering algorithms are k-means, k-medoids, expectation maximisation, clustering large applications and clustering large application based on randomised search [9]. The K-means algorithm is one of the popular for centre-based clustering [9], which is recognised as being simple and efficient. However, K-means can detect only well separated, compact or spherical clusters [10]. It is sensitive to noise due to the use of squared Euclidean distance, where any data object in the cluster can significantly influence the centre of clusters. The performance of K-means is highly sensitive to the selection of initial centres [11]. Improper initialisation may lead to empty clusters, weak convergence or a high possibility of getting trapped in a local optima [9]. Some researchers overcome these drawbacks by using meta-heuristics, such as Genetic algorithms [12], Particle Swarm Optimization [13], Ant Colony Optimization [14], Black Hole Algorithm [15], Gravitational Search Algorithm [16] and Krill Herd algorithm [17]. In the clustering problems, the balance between exploration and exploitation can affect the ability of the clustering algorithm to find good clusters among the datasets being used [18]. Some of the earlier proposed clustering algorithms, based on meta-heuristics, managed to find good clustering solutions for specific datasets. However, across all datasets, it was unable to find good results, or the results were not robust [7]. This might be due to the imbalance between exploration and exploitation of the meta-heuristic algorithm, which may lead to premature convergence or stagnation [19]. Some researchers have proposed a hybrid approach of a global search with a local search in order to achieve a better balance. The global search handles exploration, while exploitation is handled by the local search [20-23]. Memetic Algorithms (MAs) are one type of hybrid evolutionary algorithms that offer an efficient optimisation framework by combining perturbation mechanisms, local search strategies, population management [24] and learning strategies [25]. MAs can adopt the strength of other optimisation algorithms by combining them within the same framework, which can provide better performance and overcome the weakness of other algorithms. MAs comprise evolutionary phases that aid its success in complex optimisation problems [26-29]. More specifically, the mutation, the improvement and the restart phases are primarily responsible for the stability of a MAs performance [30,31]. The differential evolution (DE) algorithm can be hybridised with the MA in the mutation phase, where DE offers a superior mutation performance across many combinatorial and continuous domains’ problems [32,33]. However, the DE algorithm is subject to stagnation problems [34]. Many researchers tried to use the adaptation approach with the DE mutation operator, where two trends were mainly focusing on the control parameter adaptation strategy [35] and adaptive strategy control [36]. The importance of the mutation strategy can guide the search process to a global optimum [37]. Therefore, [38] proposed global and local neighbourhood-based mutation operators, where it can balance between the global and local search throughout the evolutionary processes. However, the mutation vectors require well-selected weights for the global and local strategy.

Related work

Many researchers have used the nature-inspired algorithms to overcome the shortcomings of the K-means algorithm, to avoid premature convergence. For example, [39] proposed a Gravitational Search Algorithm (GSA) for solving data clustering algorithm. The candidate solutions are created randomly and interact with one solution via Newton’s gravity law to find optimal solutions in the problem space. Later, [40] proposed a heuristic algorithm based on the Black Hole phenomenon, where it has a simple structure, easy, and free from parameter tuning implementation. A hybrid meta-heuristic algorithm is proposed by [41] by using Particle Swarm Optimisation and Magnetic Charge System Search algorithms for partitioning clustering problem. A dynamic shuffled differential evolution algorithm (DSDE) is proposed by [42]. The DSDE used the DE/best/1 mutation strategy and shuffled frog leaping algorithm to separate the population into two groups of the population during the evolving process. In recent researches, authors presented data clustering algorithms by integrating the K-means data clustering algorithm with the population-based meta-heuristics algorithms, for example; Abdeyazdan presented an enhanced data clustering approach for that adopts the combination of the K-harmonic means algorithm (KHM) and a modified version of the Imperialist Competitive Algorithm (ICA) algorithm [43]. Gong et al. presented an improved Artificial Bee Colony clustering algorithm by enhancing the initial clustering centres selection [44]. Mustafi et al. presented an improved Genetic Algorithm (GA) data clustering algorithm to overcome the K-means clustering algorithm drawbacks [12]. Niu et al. proposed an integrated Particle Swarm Optimisers (PSOs) with the K-means algorithm [13]. Pandey et al. proposed Improved Cuckoo Search data clustering that adopts the K-means [45]. In the research of [46], the authors proposed an improved Tabu Search strategy that is integrated with the K-means clustering algorithm. More recently, The research of [19] combined the K-Harmonic Means (KHM) algorithm with PSO and an improved Cuckoo Search (ICS). They used ICS and PSO to avoid the problem of falling into the local optima. Despite that the modified data clustering algorithms based on many evolutionary approached have better performance than other earlier algorithms, there still a problem with the weak convergence shortcoming in some evolutionary algorithms. More precisely, the exploitation and exploration balance of the evolutionary algorithms can be further improved. Therefore, in this work, we aim to improve the clustering algorithm based on an adaptive memetic differential evolution, named AMADE.

Contribution of this paper

The main objective of this paper is to address the issues discussed above by proposing an adaptive memetic differential evolution for solving data clustering problems. Specifically, the significance of our contribution is three-fold. We design an adaptive memetic differential evolution algorithm for the data clustering problem. The proposed algorithm data clustering algorithm used the approach of combining MA and DE in order to solve the data clustering problem. We develop an adaptive DE Mutation phase with an adaptive mutation strategy that can be used to narrow the search process through the evolutionary steps generations to the nearest possible centroids. We develop a local search algorithm utilising a neighbourhood selection heuristic that seeks better centroids based on the maximum and minimum values of each attribute centroid. More specifically, the algorithm proposed an adaptive DE mutation operator that was combined with memetic algorithm evolutionary steps. The mutation operator strengthens the search capabilities of DE through proposed DE mutation strategy. Thus, the algorithm also introduced an adaptive strategy to avoid any stagnation problem. The DE Mutation phase employs an adaptive DE/current-to-best/1 mutation strategy, to speed up the convergence speed of differential evolution algorithm under the guidance of both current and the best individuals. The memetic improvement phase included two steps: removing the duplicate solutions and local search using an improved neighbourhood search heuristic, the modification aimed to prevent the algorithm from falling into premature convergence. The hill-climbing local search algorithm is utilised as a local search algorithm to seek for better centroids by employing an improved neighbourhood selection heuristic with first improvement strategy. The neighbourhood selection heuristic seeks better centroids based on the maximum and minimum values of each attribute centroid. The restart phase was modified to replace the new partial population with good solution generate from the discrete differential algorithm, which can keep the diversity of the population as maximum as possible.

Organization of This Paper

This research is organised as the following: Section 2 introduces the theoretical background and concepts such as standard MA and DE. In section 3, briefly explains the improved adaptive memetic DE. Section 4 presents the experimental results of the proposed algorithm. Finally, Section 5 provides the conclusion and future works.

Background

This section discusses the fundamental aspects of clustering analysis problem, differential evolution (DE) and memetic algorithms, which have been used in the proposed data clustering algorithm. Thus, this section discusses the relevant population-based approaches in the data clustering.

Cluster analysis

Data Clustering is a process of partitioning a set of n objects into some clusters K, based on a specific similarity measure. The n objects are represented by the set X = {x, x, …, x}, the K clusters are denoted by C = {C, C, …, C}, such that data objects in the same clusters are similar, and other data objects are dissimilar. In the data clustering problem, clusters must maintain the following three hard constraints [47]: Each cluster should consist of at least one object: Different clusters should not have objects in common: Every object must be attached to a cluster: The Data clustering problem can be represented the Eq (4): The f(X, C) is the fitness function to measure the quality of the partitions generated by the clustering method. Thus, the fitness function can be maximised or minimised depending on the similarity/dissimilarity measure used. Moreover, the fitness function should be defined for adequate partitioning. The intra-cluster Distance similarity/dissimilarity measure is one of the most popular internal metrics that is utilised to measure the quality of the clustering solution [7], as in the Eq (5): The d(O, Z) represents the distance between the centre of cluster Z and data object Oi. The Euclidean distance, as in Eq (6), is one of the most famous distance functions [7]. It can measure the distance between two objects (Oi and Oj) inside the same cluster. Furthermore, the centres Z is can determine the mean value for all cluster objects as in Eq (7), where the number of data objects in cluster Z is denoted by n

Differential evolution algorithm (DE)

DE algorithm is an effective meta-heuristic optimisation algorithm for solving continuous and combinatorial optimisation problems [48]. The algorithm starts with initialising a population. The individuals are chosen as parents for the mutation and crossover operators to generate trial offspring individuals. The mutation operation perturbed a base individual by a scaled difference vector, where the vector can consist of many random individuals selected from the population in order to produce a mutant individual. The comparison between offspring individual with the parent in fitness value will result in a new individual for the next generation. The evolution process will be terminated when satisfying a termination condition. Finally, in the last generation, the best individual will be the solution to the problem. The DE algorithm starts the evolutionary process by initialising the population with individuals in the solution space. In each generation, the individuals are selected as parents for the mutation and crossover in order to generate the trial offspring individuals. In the mutation phase, the individual is perturbed by a scaled differential vector that contains several individuals that are randomly selected in order to produce the mutant individual. The offspring individual is then compared with the parent using the fitness value, and the superior one is chosen as the new individual for the next generation. The evolutionary processes are terminated when the termination condition is satisfied, and the solution to the problem will be the best individual in the last generation. The effectiveness of the DE in solving complicated optimisation problems depends mainly on choosing suitable mutation strategy and the related parameter values. Therefore, choosing suitable control parameter values for the DE algorithm is an essential task. Many researchers have been attracted to study the DE algorithm. For example, [35] proposed DE-PAS algorithm for selecting and incorporating a suitable adapting parameters scheme. [49] proposed a network intrusion detection based on efficient feature selection technique using decision tree algorithm and discretised differential evolution (DDE) from standard intrusion datasets. [50] presented a DE algorithm that can avoid premature convergence and improve the search quality. The population is grouped into many tribes and utilises an ensemble of different mutation and crossover strategies. They used an adaptive scheme to control the scaling factor and the crossover rate. [51] introduced a self-adaptive differential evolution algorithm (APDDE). The algorithm integrates the detecting values into two mutation strategies to produce the offspring population. [36] proposed a self-adaptive differential evolution algorithm with a hybrid mutation operator (SHDE) for parameters identification problem. In [52], researchers proposed a self-adaptive DE which can predict the control parameters based on the ensemble. [53] proposed a self-adaptive DE algorithm with discrete mutation control parameters (DMPSADE). Every individual contains its mutation control parameter, crossover control parameter and mutation strategy.

Memetic algorithms

The Memetic Algorithms (MAs) are a meta-heuristic approach that combines the problem-specific solvers with evolutionary algorithms. The problem solvers can be implemented using exact methods, approximation algorithms or local search heuristics. The hybridisation aims to accelerate the discovery of good solutions or to find the solutions that are unreachable by evolutionary algorithms or the local search methods alone. MAs have been proven successful performance for a broad range of problem domains, such as wireless sensor networks [54], Machine learning algorithms [55], scheduling problems [56], routing problems [57] and bioinformatics [58]. MAs received many names throughout the literature. Some of the alternative names are hybrid GA, Baldwinian EA, Lamarckian EA, genetic local search algorithms [59]. The MAs can combine techniques and approach from many search techniques, and most distinguished approaches from local search methods and population-based search techniques. The basic memetic algorithms template include procedures:

The initialisation procedure

The initialisation procedure is responsible for creating solutions to the initial set of the population. The MAs seeks to create high-quality solutions to be in the starting point. The initialisation procedure can be done either using a local search procedure or a constructive heuristic to improve the random initial solutions.

The cooperate and improve procedures

The cooperate and Improve procedures typically rely on the selection of the solutions from the population and recombine them. Both procedures utilise the approach of a local search in the population.

The compete procedure

The Compete procedure is used in the reconstruction of the current population using the old and the new population. A steady-state replacement strategy is one of the most popular strategies that could be used when the fitness function suffers from complexity and time-consumption and could lead to faster convergence [59].

The restart procedure

The restart procedure is invoked whenever the population falls into a degenerate state. Typically, one of the strategies that could be used is to keep a part of the current population and generate the remaining part by new solutions. Another approach is to apply a heavy mutation operator; this could generate a population different from the current state in the search space.

Improved adaptive memetic differential evolution

This section discusses the detailed steps of the propose AMADE algorithm along with the solution representation.

Solution representation

The optimal encoding aims to determine the data objects that belong to a particular cluster to perform optimal clustering analysis. A label-based one-dimensional array is used to represent the candidate solution in data clustering optimisation problem. Every solution representation is considered as a set of N data objects, where each cell represents a cluster number associated with that object. Fig 1 presents a candidate solution example for a problem with nine data objects and two clusters, In this example: objects O O O O and O is attached with clusters 1, and O O O and O is attached with clusters 2.
Fig 1

Example of a candidate solution represented by a label-based representation.

Additionally, a centroid-based representation that consists of a two-dimensional matrix is used to keep track of the positions of the cluster centroid and to be used by the local search. The matrix consists of K rows and D column, where K is the total number of the clusters and D is the total number of the attributes in the dataset. For example, in Fig 2, the dataset contains two clusters and two attributes; then the position of first cluster centroids is 4.5, 2.3, and the position of the second cluster centroids is 5.5, 7.4.
Fig 2

Cluster centroid-based representation of the solutions.

Constraint handling

The solution representation of the proposed AMADE guarantees that each data object is associated only with one cluster. An additional soft constraint is formed to prevent any duplicate solutions in the population in the improvement phase. Moreover, any possible duplicate solutions can lead through the evolutionary processes to premature stagnation. The duplicate solution is handled in the improvement phase by generating solutions randomly.

The AMADE proposed approach

In AMADE, the DE mutation operator with an adaptive strategy DE/current-to-best/1 has combined with the memetic algorithm evolutionary steps; this aims to have faster convergence speed faster by the best individual’s guidance. The new individuals are compared with the target vector, which can improve the guidance of the population evolution. However, AMADE may suffer from premature convergence. To this end, the restart phase can prevent falling into premature convergence by reconstructing the population diversity by generating new solutions in the population. The improvement phase plays a key role in finding better solutions, which can also improve the quality of the solution using a proposed improvement heuristic. The pseudo-code for the proposed AMADE algorithm is shown in Fig 3.
Fig 3

Pseudo-code of the proposed AMADE algorithm.

The population initialisation phase

For keeping better diversity of the population, a random constructive method is used. The initial solutions for the proposed AMADE are randomly generated. The data points of the dataset are randomly grouped into K clusters; all centroids of each cluster are calculated using Eq (7), where K is the total number of clusters. These two steps are repeated N times to generate N random solutions, where N is the population size parameter value of the AMADE algorithm.

The recombination phase

The recombination phase employs the mating pool approach [60] in evolutionary computation with CandPoolSize size. The tournament selection operators, with selection size TourSize [61] are applied to the entire population then placed into the mating pool. Thus, a two-point crossover operator [62] is then applied to the selected parents. At the end of the recombination phase, the mating pool is combined with the population by replacing some worst individuals with the new better individuals in the mating pool. The pseudo-code of the recombination phase is shown in Fig 4.
Fig 4

Pseudo-code of the recombination phase algorithm.

Fig 5 demonstrates an example of the two-point crossover with label-based solution representation chromosome with size N = 12 genes. The chosen cut points are randomly selected, and cut point 1 should be less than cut point 2.
Fig 5

Example of two-point crossover with label-based solution representation for a clustering solution using of twelve data.

The DE Mutation phase

The DE Mutation phase employs an adaptive DE/current-to-best/1 mutation strategy, to speed up the convergence speed of DE algorithm under the guidance of both current and the best individuals. The cluster centroids are modified by the mutation phase to achieve better cluster solution, as shown in Fig 6. This is performed by using Eq (8). Where C is the current individual centroid, C is the best individual centroid, C is a random individual centroid, CurrIteration is the current iteration in AMADE algorithm, and MaxIterations is the maximum number of iterations of AMADE.
Fig 6

Pseudo-code of creating the trial individual algorithm.

Such adaptive strategy will narrow the search process through the evolutionary steps generations to the nearest possible centroids. At the same phase, the data objects are rearranged to the closest clusters after modifying the centroids of the clusters. The new produced individual is immediately compared with the target vector in a current population, and the better individual could be retained. In order to demonstrate the effectiveness of the adaptive DE strategy, Fig 7 shows an example of a cluster centroid of value 6.5 that is adjusted throughout 1000 iterations of the AMADE algorithm. The adaptive DE strategy provides more exploration capabilities to cluster centroid, which in the first iteration is 6.5 and is adjusted to the new centroid that is 11.5. As the algorithm reaches the maximum number of iterations, the DE strategy produces more exploitation capability to the current centroids, which in the iteration 999 is 10.1 and the new centroid is 10.1004.
Fig 7

Example of the adaptive DE strategy performed on a cluster centroid of value 6.5 throughout 1000 iterations.

The improvement phase

The improvements phase consists of two solution quality improvement steps: the clear duplicate solutions step and the local search step. In the clear duplicate step, the algorithm ensures that the population retains better solution diversity in order to avoid any premature stagnation. At the second step, a hill-climbing local search [63] is employed on the centroid-based presentation by changing the current centroid with better cluster centroids. The hill-climbing local search algorithm, as shown in Fig 8, seeks better centroids by utilising the neighbourhood selection heuristic with a first improvement strategy [64]. The algorithm terminates when the current solution is improved.
Fig 8

Pseudo-code of the modified Hill climbing algorithm.

The neighbourhood selection heuristic, as shown as pseudo-code in Fig 9 and a flowchart in Fig 10, seeks better centroids based on the maximum and minimum values of each field’s centroid. The heuristic increases the centroid value with an increment step value until finding better centroid. Otherwise, the algorithm will change the search direction decreasingly to the minimum value of centroid.
Fig 9

Pseudo-code of the neighbourhood selection heuristic algorithm.

Fig 10

Flowchart of the neighbourhood selection heuristic.

The restart population phase

Once the population is having a state of degeneration, the restart procedure is employed immediately [59]. The restart strategy keeps part of the population and excludes the remaining individuals by generating new solutions. As shown Fig 11, AMADE keeps 75% of the population for the next evolutionary steps, while the remaining population is generated using a DE algorithm based on mutation strategy DE/rand/1 and the minimal number of generations, which can produce a new population with better diversity and good quality solutions.
Fig 11

Pseudo-code of the restart population algorithm.

The DE algorithm, shown in Fig 12, is applied to the solution representation with a discrete mutation operator. Each genome in the new chromosome is calculated using Eq (9). Where G, G, G is the gene in the chromosome of randomly selected individuals. The modulus is used to ensure that the result of the equation within the number of clusters in the dataset.
Fig 12

Pseudo-code of creating DE population algorithm.

Experimental setup and results

Experimental setup

The performance of the proposed AMADE clustering method are investigated based on six real data datasets from the UCI repository of the machine learning databases with a variety of complexity [65], which can be download at http://archive.ics.uci.edu/ml/index.php. The datasets that been used are Wisconsin Breast Cancer, Vowel, Wine, Iris, Contraceptive Method Choice (CMC) and Glass, as shown in Table 1. The datasets also include different complexity levels and classified from 1 to 10 levels based on the number of instances and attributes [66], where level 1 is the lowest complexity level, and level 10 represents the highest complexity level.
Table 1

The characteristics of the UCI repository datasets used in the experiments of AMADE algorithm.

DatasetNumber of clustersNumber of featuresNumber of data objectsDescriptionComplexity levels
Vowel63871Indian Telugu vowel4
Iris34150Fisher’s iris data3
Cancer29683Wisconsin breast cancer3
CMC391473Contraceptive method choice6
Glass69214Glass identification data3
Wine313178Wine data4
In order to evaluate the effectiveness of the proposed Memetic DE algorithm with proposed evolutionary phases, the AMADE performance is first compared with DE [48] with DE/best/1/bin strategy, Hybrid DE with DE/best/1/bin strategy, GA [67] and hybrid GA algorithms, where all algorithms are applied with the same experimental setup and local search heuristic. These algorithms have the same evolutionary phases of AMADA except restart phase. The selection of these algorithms is essential to show the strength of the combination of such algorithms in MA besides the proposed adaptive mutation operator and the modified restart phase. Moreover, for further testify the performance, the AMADE is compared with recent data clustering algorithms in the literature, including K-means [9], black hole [40], age-based particle swarm optimisation [68], dynamic shuffled differential evolution algorithm [42], the krill herd algorithm [69] and hybrid ICMPKHM [19]. The algorithm's performance is evaluated using the following criteria: The intra-cluster distances: is an internal quality measure that measures the distance between all objects in the cluster and its centre, as defined in Eq (5). The purpose of the data clustering algorithm is to minimise the sum of intra-cluster distances which can lead to high clustering quality. The intra-cluster distance value is given as best (minimum intra-cluster distance), the average value and worst (maximum intra-cluster distance) value of objective function value among entire runs. The F-measure: is an external measure that compares the ground truth with the obtained clusters to calculate the similarity between them. The high percentage of the F-measure value indicates a better clustering quality. The precision and recall of cluster Sj, and class Ri, i, j = 1, 2, …, k is shown in Eq (10) and Eq (11), Where |R| is the number of objects in class R and |S| is the number of data objects in cluster S, and L is the number of data objects of class R in cluster S. The F-measure of a class R is defined in Eq (12). The overall F-measure is computed as the weighted average of all classes is given in Eq (13). The accuracy: is an external measure indicates the proportionate number of data objects that correctly placed by the predictive model to match the class (ground truth) in the data, as shown in Eq (14): The parameter settings for the AMADE algorithm were independently tested on each of the six datasets for 31 times, the best, worst, average values, standard deviations and F-measure were computed. In AMADE, the maximum number of generations is set to 1000, and 100 to DE/rand/1 in population restart phase. Accordingly, F is set to 0.7 and Cr is set to 0.9. A Taguchi method [70,71] for the design of the experiment has been used to identify the best values of the parameters for AMADE algorithm. Five levels were considered for each factor as shown in Table 2. AMADE algorithm run for 31 times for each factor at each level was employed, and the mean of signal-to-noise (SN) ratio plot each level of the factors are shown in Fig 13. The level with the maximum SN ratio is the optimum parameter determined by Taguchi method.
Table 2

The AMADE algorithm parameter levels.

ParameterDefinitionLevel
12345
PopSizeThe population size2030405060
MGWImax generation without improvement1020304050
CandPoolSizerecombination mating pool size15101520
TourSizetournament selection pressure510152025
Fig 13

Graphical results of Taguchi method for AMADE algorithm.

According to Fig 13, the optimum value for population size is set to 20, and max generation without improvement is set to 50. The recombination mating pool size is set to 10, and the tournament selection pressure is set to 10. Last but not least, All algorithms are implemented in Oracle Java 1.8 and were run on CPU Intel Core i7 (2.4GHz) personal computer that contains 8 GB of RAM.

Experimental results and discussion

The objective function values comparison for the best solutions, average solutions and worst solutions, F-measure, standard deviation and execution time of solutions for 31 runs is shown in Table 3. Where Best, Mean and Worst are referred to the intra-cluster distances objective function values that were obtained out of 31 runs, where the smaller value is better, and the higher value of the F-measure is better. The results show that the proposed algorithm has a smaller best, average, worst and standard deviation compared with the other algorithms. For example, the Iris dataset results show that AMADE achieved 96.544 global optima whereas the best solutions of GA, DE, HyGA ad HyDE are 97.225, 97.101, 96.571 and 96.571. However, the worst, best, and average results of the solutions by HyGA and HyDE are close to AMADE on most of the datasets, but it did not perform well with the standard deviation and the results of worst solutions. Moreover, the results of F-measure of the proposed algorithm can be noticed as better than other algorithms in most datasets, except for the iris and cancer datasets which are similar to the global optimum.
Table 3

Comparison of intra-clusters distances among AMADE, HyDE, HyGA, DE and GA obtained from 31 runs.

Data SetCriteriaGADEHyGAHyDEAMADE
IrisBest97.22597.10196.57196.57196.544
Mean100.22100.23896.70496.68796.549
Worst106.63121.4297.08296.85196.56
Std.2.7998.2870.13320.09130.004
F-measure0.8880.9010.9010.9010.901
Time (s)0.0630.0623.3112.9082.442
WineBest16555.67916530.5316295.93216293.71616292.279
Mean17469.55416579.3016307.62616320.59116292.82
Worst21381.73218042.4616375.15116424.5516293.884
Std.857.221271.55014.75043.3440.395
F-measure0.6890.6960.6960.6960.708
Time (s)0.0640.10515.1706.2295.279
VowelBest213180.89228726.3149225.51149216.30148967.54
Mean338611.71246848.6150098.78150537.95149228.50
Worst414649.26264116.7151281.28157436.32150121.94
Std.58186.9438847.791388.2711415.782490.294
F-measure0.57310.5490.6450.5490.66209
Time (s)0.3520.82525.90115.33314.166
CMCBest8232.037414.525534.2095532.8555532.404
Mean9913.7258242.9655538.5355535.5665532.620
Worst10919.048724.3835591.4295538.7345534.836
Std.682.494306.2429.9281.7750.423
F-measure0.4860.48750.5170.4870.52107
Time (s)0.8702.09886.88772.47059.485
GlassBest223.509216.911214.382213.726210.17
Mean245.893221.685215.877216.187211.214
Worst352.186225.658220.359221.714213.686
Std.27.7203.0021.2661.8141.174
F-measure0.5890.6110.5970.6110.680
Time (s)0.0790.25337.19926.60628.459
CancerBest3022.2242984.0682965.9452964.7222964.393
Mean3342.2132984.6742973.2962966.0212964.522
Worst4316.2042985.6592994.2172976.2722964.73
Std.346.9720.6455.7432.7530.091
F-measure0.9570.9670.9640.9670.964
Time (s)0.2830.80223.60716.837014.736
Furthermore, the trade-off between the quality and the time-cost problem occurred, leading to the time-cost-quality trade-off problem. The hybrid metaheuristic approaches, such as AMADE, HyDE, and HyGA, can obtain optimal solutions in reasonable execution time. In contrast, the traditional metaheuristic algorithm, such as GA and DE, do not guarantee to find the optimal solution, but they usually obtain sub-optimal, good-quality solutions in less execution time. As shown in Table 3, The traditional DE and DE algorithm achieve best execution time for all dataset, but they were unable to obtain the optimal solution for the datasets. In contrast, AMADE algorithm produced the optimal results of the intra-clusters distances and the F-measure with reasonable execution time when compared to HyDE and HyGA. For example, AMADE obtained 5532.620 for the average intra-cluster distance on CMC dataset, and 0.52107 for the F-measure in 59.485 seconds, which were the optimal results with the best execution time compared to HyDE (72.470) and HyGA (86.887). Fig 14 shows the convergence curves of the first 200 iterations on six datasets. It demonstrates that AMADE has the best convergence rate results on the six datasets with faster converge in the early iterations of the search process; later, the convergence becomes slower. The HyDE achieved the second best convergence rate results, and the HyGA scored third best results. The GA and DE algorithm produced a slow convergence rate toward the optimum intra-cluster distance on all datasets. In general, the improved memetic phases by removing the duplicated solutions along with the local search and the adaptive strategy shown the effectiveness in preventing the algorithm from falling into premature convergence.
Fig 14

The convergence curves for the first 200 iterations on (a) cancer, (b) CMC; (c) glass; (d) iris; (e) vowel; (f) wine datasets.

The convergence curves for the first 200 iterations on (a) cancer, (b) CMC; (c) glass; (d) iris; (e) vowel; (f) wine datasets. Furthermore, Table 4 shows the result of the rankings of the mean values generated by Friedman’s test based on the average and best value of intra-clusters distances. Additionally, the Friedman’s test reveals the significance of the AMADE algorithm with a p-value of 0. 000189 for the test based on the average value of intra-clusters distances, and 0.0000128 for the test based on the best value of intra-clusters distances, which are both below the significance level (α = 0.05).
Table 4

Friedman tests based on the average and best intra-clusters distances obtained by AMADE.

#AlgorithmRanking(Based on Average)Ranking(Based on Best)
1AMADE1.001.00
2HyGA2.502.916
3HyDE2.502.083
4DE4.174.166
5GA4.834.833
The Holm’s procedure is employed as a post-hoc method to detect the statistical difference between the control case (ranked first) and the other remaining cases [72]. Table 5 shows the p-value obtained by the Holm’s procedure, where the rejection of the null hypothesis relies on the obtained p-value. Thus, the p-value must be less than the adjusted value of α (α/i), where i is the rank of the algorithm. Table 5 presents the adjusted p-value of Holm’s procedure, and the AMADE algorithm is used as the control algorithm. Holm’s procedure proves that AMADE is statistically better than DE, GA and HyGA, but the algorithm does not differ significantly from the HyDE algorithm. However, the results reported in Table 5 demonstrate that the proposed AMADE approach outperformed the HyDE in all of the tested datasets in all criteria. Based on the standard deviation criterion, AMADE is considered and more robust than HyDE as well as the other algorithms. Moreover, AMADE can found global optimal solutions for most of the cases.
Table 5

Holm’s procedure Adjusted p-value of the methods in the comparison.

iAlgorithmα/ip-value of Holms(based on average)p-value of Holms(based on best)Null Hypothesis
1HyGA0.05/1 = 0.05000.10030.03576Not rejected, rejected
2HyDE0.05/2 = 0.02500.10030.2353Not rejected, Not rejected
3DE0.05/3 = 0.01660.000520.000522Rejected, rejected
4GA0.05/4 = 0.01250.0002670.000026Rejected, rejected
Additionally, in order to show the superiority of the AMADE algorithm among the other algorithms, Fig 15 presents the box plots of all datasets from 31 runs. It reveals that AMADE did not produce any outlier on all datasets, and the median solutions obtained by AMADE distributions are centralised. The box plots for the AMADE was thick and near the minimum intra-clusters distance values. The thickness of the box plots indicates that results obtained have less deviation of the median value, which means that the algorithm performance was stable over the 31 runs. The HyDE algorithm achieved the second best performance on Cancer, CMC, and Iris datasets, while it almost obtained the same performance of the HyGA algorithm on the Glass, Vowel, and Wine datasets. The standard DE algorithm obtained a better result than the GA algorithm on all datasets, where both GA and DE performance are weak compared with other hybrid algorithms. In general, the improved memetic phases by the restart phase along with the DE mutation phase shown the effectiveness in keeping the diversity of the population as maximum as possible during the evolutionary process, which helped to avoid the instability of the obtained results.
Fig 15

Box plots of fitness of best solutions for AMADE, HyGA, HyDE, GA and DE algorithms on (a) cancer; (b) CMC; (c) glass; (d) iris; (e) vowel; (f) wine datasets.

Box plots of fitness of best solutions for AMADE, HyGA, HyDE, GA and DE algorithms on (a) cancer; (b) CMC; (c) glass; (d) iris; (e) vowel; (f) wine datasets. Furthermore, in order to validate the feasibility of the results, the centres of the clusters obtained by AMADE algorithm is shown in Tables 6–8, where all datasets with the same number of clusters are grouped in one table. The clusters centres can be used to validate the sum of intra-cluster distances given in Table 3. This could be manipulated by assigning the data objects within each dataset with the nearest clusters centres given accordingly in Tables 6–8, where the best intra-clusters distance values in Table 3 must be reached. For example, by allocating the 178 data objects in Wine dataset to the nearest centres with corresponding three cluster centres that are shown in Table 6, the best value of the sum of intra-cluster distances obtained by the AMADE algorithm on the Wine dataset, which is reported in Table 3, should be equals (16292.279). Otherwise, the best centres in Table 6 or the best values in Table 3 is invalid. This procedure can also be performed to validate other dataset’s cluster centres.
Table 6

The best clusters centres on the datasets Wine, Iris, and CMC obtained by the AMADE algorithm.

DatasetCentre 1Centre 2Centre 3
Iris6.7315.0195.93
3.0723.4232.797
5.6291.4694.416
2.1070.2381.417
CMC33.48724.40943.63
3.1373.043.01
3.5593.5123.458
3.6531.7984.595
0.8130.9330.797
0.7070.8170.764
2.112.3031.848
3.2932.9673.438
0.0730.0430.113
Wine12.52613.71812.832
2.3891.8612.582
2.3292.4242.393
21.39116.90519.509
92.435105.14698.835
2.092.762.088
1.8632.8461.51
0.4080.2930.429
1.4641.8951.433
4.3495.6015.801
0.9671.0720.898
2.5253.0152.258
463.8711137.495686.798
Table 8

The best clusters centres on the Vowel and Glass datasets obtained by the AMADE algorithm.

DatasetCentre 1Centre 2Centre 3Centre 4Centre 5Centre 6
Vowel506.853439.01623.906375.474407.997357.254
1839.705987.211309.8312149.2021018.0422291.144
2556.0512665.0012333.4762678.1362317.7632977.367
Glass1.521.5171.5131.5171.5221.521
12.84314.61213.02113.32413.81213.088
3.4560.0560.0183.5783.5690.259
1.3122.2043.031.4180.9351.426
73.02273.23970.58172.66971.85472.666
0.6020.0896.2210.5740.1650.342
8.5658.6516.948.2129.52311.978
0.0131.030.010.0060.0520.108
0.0790.0160.0010.050.0580.061

Comparison between AMADE and state of the art

In order to evaluate the performance of AMADE, the algorithm results are compared with well-known algorithms, such as the black hole (BH) [40], age-based particle swarm optimization (PSOAG) [68], A dynamic shuffled differential evolution algorithm (DSDE) [42], the krill herd algorithm (IKHCA) [69], hybrid of krill herd algorithm with harmony search algorithm (H-KHA) [17] and hybrid ICMPKHM [19]. The related comparison results are presented in Table 9. The results present the average of the intra-clusters distances for the AMADE and other Algorithms on Iris, Wine, CMC, Glass, and Cancer. The results indicate that AMADE has shown consistent performance and better result than IKHCA, ICMPKHM, PSOAG, H-KHA and BH on almost all the datasets. The AMADE achieved the second best results after the MSDE algorithm on Wine, CMC, Cancer datasets. Thus, The AMADE algorithm obtained the second best results on Iris and Glass datasets. The results shown in Table 9 reveal that the AMADE performance is consistent across all the datasets compared to the state of art algorithms concerning the average of the intra-clusters distances.
Table 9

Comparison between AMADE and other Algorithms based on the average of the intra-clusters distances.

DatasetPSOAGBHDSDEIKHCAICMPKHMH-KHAAMADE
Iris96.9796.6596.6596.6796.6196.5296.549
Wine16296.316294.316292.316589.16293.1816410.116292.8
CMC5559.985533.635532.185695.05695.135601.685532.62
Glass244.99211.49212.73223.03199.45215.66211.21
Cancer2984.242964.392964.382971.13024.792982.432964.52
To further analyse the results in Table 9, the rankings with the compared algorithms generated by Friedman’s test are shown in Table 10 based on the average function of the intra-clusters distances. Furthermore, the Friedman’s test has shown a significant difference of the AMADE among the other compared algorithms, with a p-value of 0.02465 based on the average function, which is below the significance level (α = 0.05). The AMADE algorithm shares the best ranked algorithm with the DSDE algorithm [42], which uses the DE algorithm with multiple population approaches to reach the best average function of the intra-clusters distances for the best solutions. The results show that AMADE achieved the best ranking among other clustering algorithm based on the average performance function of the intra-clusters distances. The BH algorithm achieved the third best rank, and the ICMPKHM algorithm achieved the fourth rank, then the H-KHA. Lastly, PSOAG and IKHCA achieved the worst rank compared to other algorithms. The rankings generated by Friedman’s test shown in Table 9 reveal that the AMADE performance is consistent compared to the state of art algorithms concerning the average of the intra-clusters distances.
Table 10

Friedman tests based on the average of the intra-clusters distances.

AlgorithmRanking
AMADE2.2
DSDE2.2
BH3.4
ICMPKHM4.2
H-KHA4.4
PSOAG5.8
IKHCA5.8
Furthermore, the performance of AMADE is compared based on the computed accuracy with four algorithms that reported accuracy performance measure in their research, such as PSOAG, K-means [73], PSOAG, DSDE and IKHCA as shown in Table 11. The accuracy obtained by AMADE is competitive with the other clustering algorithm, where it reaches the optimum accuracy on CMC and cancer datasets. The IKHCA algorithm achieved best results of the accuracy on Wine, CMC, and Glass datasets, while the PSOAG algorithm achieved the best result on the Iris dataset. However, the results of the accuracy reveal the consistent performance of the AMADE algorithm based on the accuracy on all datasets, where it obtained second best result of accuracy on Glass and Wine datasets and obtained the third best result of the accuracy on the Iris dataset.
Table 11

Comparison between AMADE and other population-based algorithms based on accuracy.

Data SetK-meansPSOAGDSDEIKHCAAMADE
Iris83.391.0390.0090.6790.0
Wine63.6270.9871.6573.0371.9
CMC41.839.8738.4945.6245.62
Glass60.851.2653.4865.8863.08
Cancer93.3796.3196.48695.1696.486
At last but not least, the performance of AMADE is compared based on the computed F-measure with three algorithms that have reported the F-measure external performance measure in their research, such as K-means [74], KSC-LCA [74], ICMPKHM [19] as shown in Table 12. The F-measure obtained by AMADE outperformed other clustering algorithms, where it reached the optimum F-measure value on the Iris, CMC, Cancer and Vowel datasets, while it obtained the second best results of the F-measure on Wine and Glass. The KSC-LCA algorithm achieved the best result of the F-measure on Wine dataset, and ICMPKHM algorithm achieved the best result on Glass and Cancer datasets. The results shown in Table 12 reveals the consistent performance of AMADE across all dataset based on the F-measure.
Table 12

Comparison between AMADE and other population-based algorithms based on the F-measure.

Data SetK-meansKSC-LCAICMPKHMAMADE
Iris80.457289.877589.23290.1
Wine66.978173.022168.8170.8
CMC36.927342.898147.5152.10
Glass45.944049.673369.5268.0
Cancer95.686396.173096.496.4
Vowel48.685951.936065.866.20

Conclusions and future work

In this work, an adaptive memetic differential evolution (AMADE) was proposed for efficient data clustering. The combination between MA and DE algorithms aimed to balance between the exploration and exploitation. The algorithm proposed an adaptive DE mutation operator and a neighbourhood selection heuristic that are combined with memetic algorithm evolutionary steps. The enhancements helped to avoid the instability of the obtained results by keeping the diversity of the population as maximum as possible during the evolutionary process. Experiments conducted on six real-life datasets with different level of complexity have demonstrated that the AMADE showed consistent performance compared to the state of art algorithms concerning the average of the intra-clusters distances, accuracy, and F-measure validity measures. AMADE algorithm achieved the optimum result of the accuracy on CMC (45.62%) and Cancer (96.486%) datasets, and also reached the optimum result of the F-measure on Iris (90.1%), CMC (52.10%), Cancer (96.4%), and Vowel (66.20%) datasets. Moreover, future work will focus on using other data clustering objective functions to solve a variety of categorical and mixed data datasets. Additionally, future work will focus on how to associate validity measures with each other when combined in multi-objective approaches.
Table 7

The best clustering centres on the Cancer data set obtained by the AMADE algorithm.

DatasetCentre 1Centre 2
Cancer2.8867.113
1.1276.639
1.2016.624
1.1655.613
1.9935.227
1.1228.097
2.0086.078
1.0996.021
1.0332.32
2.8867.113
1.1276.639
1.2016.624
1.1655.613
  5 in total

1.  An Adaptive Multipopulation Differential Evolution With Dynamic Population Reduction.

Authors:  Mostafa Z Ali; Noor H Awad; Ponnuthurai Nagaratnam Suganthan; Robert G Reynolds
Journal:  IEEE Trans Cybern       Date:  2016-10-25       Impact factor: 11.448

2.  A memetic optimization algorithm for multi-constrained multicast routing in ad hoc networks.

Authors:  Rahab M Ramadan; Safa M Gasser; Mohamed S El-Mahallawy; Karim Hammad; Ahmed M El Bakly
Journal:  PLoS One       Date:  2018-03-06       Impact factor: 3.240

3.  An optimized digital watermarking algorithm in wavelet domain based on differential evolution for color image.

Authors:  Xinchun Cui; Yuying Niu; Xiangwei Zheng; Yingshuai Han
Journal:  PLoS One       Date:  2018-05-21       Impact factor: 3.240

4.  A differential privacy protecting K-means clustering algorithm based on contour coefficients.

Authors:  Yaling Zhang; Na Liu; Shangping Wang
Journal:  PLoS One       Date:  2018-11-21       Impact factor: 3.240

5.  Clustering algorithms: A comparative approach.

Authors:  Mayra Z Rodriguez; Cesar H Comin; Dalcimar Casanova; Odemir M Bruno; Diego R Amancio; Luciano da F Costa; Francisco A Rodrigues
Journal:  PLoS One       Date:  2019-01-15       Impact factor: 3.240

  5 in total
  3 in total

1.  Solving text clustering problem using a memetic differential evolution algorithm.

Authors:  Hossam M J Mustafa; Masri Ayob; Dheeb Albashish; Sawsan Abu-Taleb
Journal:  PLoS One       Date:  2020-06-11       Impact factor: 3.240

2.  Velocity prediction of nanofluid in a heated porous pipe: DEFIS learning of CFD results.

Authors:  Meisam Babanezhad; Iman Behroyan; Azam Marjani; Saeed Shirazian
Journal:  Sci Rep       Date:  2021-01-13       Impact factor: 4.379

3.  A novel hybrid soft computing optimization framework for dynamic economic dispatch problem of complex non-convex contiguous constrained machines.

Authors:  Ijaz Ahmed; Um-E-Habiba Alvi; Abdul Basit; Tayyaba Khursheed; Alwena Alvi; Keum-Shik Hong; Muhammad Rehan
Journal:  PLoS One       Date:  2022-01-26       Impact factor: 3.240

  3 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.