Literature DB >> 27656201

A Modified MinMax k-Means Algorithm Based on PSO.

Xiaoyan Wang1, Yanping Bai2.   

Abstract

The MinMax k-means algorithm is widely used to tackle the effect of bad initialization by minimizing the maximum intraclustering errors. Two parameters, including the exponent parameter and memory parameter, are involved in the executive process. Since different parameters have different clustering errors, it is crucial to choose appropriate parameters. In the original algorithm, a practical framework is given. Such framework extends the MinMax k-means to automatically adapt the exponent parameter to the data set. It has been believed that if the maximum exponent parameter has been set, then the programme can reach the lowest intraclustering errors. However, our experiments show that this is not always correct. In this paper, we modified the MinMax k-means algorithm by PSO to determine the proper values of parameters which can subject the algorithm to attain the lowest clustering errors. The proposed clustering method is tested on some favorite data sets in several different initial situations and is compared to the k-means algorithm and the original MinMax k-means algorithm. The experimental results indicate that our proposed algorithm can reach the lowest clustering errors automatically.

Entities:  

Year:  2016        PMID: 27656201      PMCID: PMC5021908          DOI: 10.1155/2016/4606384

Source DB:  PubMed          Journal:  Comput Intell Neurosci


1. Introduction

Clustering has broad applications in pattern recognition, image processing, machine learning, and statistics [1, 2]. The aim is to partition a collection of patterns into disjoint clusters, such that patterns in the same cluster are similar, and, on the other hand, patterns of two different clusters are distinct. One of the most popular clustering methods is k-means algorithm, where clusters are identified by minimizing the clustering error. The k-means algorithm is widely accepted in the literature. However, the k-means algorithm is sensitive to the choice of initial starting conditions [3, 4]. To deal with this problem, several methods have been proposed. For example, a method has been proposed to eliminate the dependence on random initial conditions. Global k-means algorithm [5] is an incremental approach that starts from one cluster and at each step a new cluster is deterministically added to the solution according to an appropriate criterion. Based on the algorithm, Bagirov et al. proposed some modifications [6, 7]. Tzortzis and Likas extended the algorithm to kernel space [8, 9]. Zang et al. developed a fuzzy c-means clustering algorithm and applied such algorithm to the investigation of speech signal [10]. An alternative approach to eliminate the influence of initial starting conditions is to use the multi-restarting k-means algorithm [11-14]. A new version of this method is the MinMax k-means clustering algorithm [15], which starts from a randomly picked set of cluster centers and tries to minimize the maximum intraclustering error. Its application [16] shows that the algorithm is efficient in intrusion detection. Particle Swarm Optimization (PSO), a population-based stochastic search process, was firstly proposed by Eberhart and Kennedy in 1995 [17]. Such process was introduced to simulate the social behaviors of bird flocking or fish schooling when the group of birds or fish searches for food. PSO is very fast, simple, and easy to understand and implement. PSO does not require the adjustment of many parameters and the memory space required by the algorithm is little. PSO has been widely used to improve other algorithms' performances such as ANN [18-20], scheduling problems [21, 22], traveling salesman problems [23, 24], and anomaly detection problems [25]. PSO also has been successfully applied in clustering problem [26-30]. Recently, PSO and k-means algorithm are combined to develop novel clustering algorithms [31, 32]. In this paper, a new version of modified MinMax k-means algorithm is proposed. Recent investigation indicates that if the parameter of p max has been set, the programme can reach the lowest E max at p ∈ [p init, p max] [15]. However, experiments imply that the above conclusion is not always correct. Different parameters of p max result in different values of E max, and as such the value does not always comply with the rule that the larger the value of p max, the lower the value of E max. In MinMax k-means algorithm, parameter β needs priory set as well, and different values of β also result in different values of clustering errors. The value of clustering errors does not have any regulation. Therefore we should decide the values of parameters to minimize cluster errors. In this paper, we calculate the clustering errors for each parameter p, respectively, without using the automatically adapted exponent p as in [15]. By utilizing PSO, we choose the parametric value and obtain the minimum clustering errors. Thus, we can obtain the minimum clustering errors without choosing parameters manually. We carry out many experiments on different data sets, including synthetic data sets and real world data sets in five different initial situations. Balanced type, unbalanced type, and almost balanced type data sets are considered in the study. Our investigation indicates that the proposed algorithm can search proper parameters to get the minimum clustering errors. The rest of the paper is organized as follows. We briefly describe the k-means, MinMax k-means, and PSO algorithms in Section 2. In Section 3 we propose our algorithms. Experimental evaluation is presented in Section 4. Lastly, Section 5 concludes our work.

2. Preliminaries

2.1. k-Means Algorithm

Given a data set X = {x 1, x 2,…, x }, x ∈ R   (n = 1,2,…, N), we aim to partition this data set into M disjoint clusters C 1, C 2,…, C , such that a clustering criterion is optimized. Usually, the clustering criterion is the sum of the squared Euclidean distances between each data point x and the corresponding cluster center m . This kind of criterion is called clustering error and depends on the cluster centers m 1, m 2,…, m :where Generally, we call ∑ I(x ∈ C )‖x − m ‖2 intraclustering error (variance). Obviously, clustering error is the sum of intraclustering error. Therefore, we use E sum to denote E(m 1, m 2,…, m ); that is, E sum = E(m 1, m 2,…, m ). The k-means algorithm finds locally optimal solutions with respect to the clustering error. The main disadvantage of the method is its sensitivity to initial position of the cluster center.

2.2. The MinMax k-Means Algorithm

As is known, in the k-means algorithm, we minimize the clustering error. The MinMax k-means algorithm minimizes the maximum intraclustering error:where m , I(x) are defined as (1). Since directly minimizing the maximum intracluster variance E max is difficult, a relaxed maximum variance objective was proposed [15]. The authors constructed a weighted formulation E of the sum of the intracluster variances:where the p exponent is a constant. The greater (smaller) the p value is, the less (more) similar the weight values become, as relative differences of the variances among the clusters are enhanced (suppressed). In [15], the authors give a practical framework that extends the MinMax k-means to automatically adapt the exponent p to the data set. It begins with a small p (p init) that is increased by p step after each iteration, until a maximum value p (p max) is attained. For the method, we should first decide the values of parameters p init, p max, and p step. Now, all clusters contribute to the objective, according to different degrees regulated by the w values. It is clear that the more a cluster contributes (higher weight), the more intensely its variance will be minimized. So w are calculated by To enhance the stability of the MinMax k-means algorithm, a memory effect could be added to the weights:

2.3. PSO

PSO is a population-based metaheuristic algorithm. It is launched with a population (called a swarm) of individuals (called particles), where each particle represents a potential solution in the problem space. All of the particles move around the problem space to find the optimization solution (the best position) according to some speed (velocity) iteratively. In the n-dimension problem space, the position and the velocity of the ith particle are, respectively, denoted by the following vectors: The solution is evaluated by the fitness value for each particle at every iteration. Afterwards, a record of the best position of each particle based on fitness value is saved. The best previously visited position of the particle i at iteration t is denoted by vector P (t) = (p (t), p (t),…, p (t)) as the personal best. The position of all the particles which give the best fitness value at iteration t is also recorded as the global best position denoted by G(t) = (g 1(t), g 2(t),…, g (t)). At each iteration, the velocity and the position of each particle are updated according to the following equations:where ω is an inertia weight that introduces a preference for the particle to continue moving in the same direction. Here, c 1 and c 2 are two positive constant parameters called coefficients, and r 1 and r 2 denote two random numbers between (0,1). In order to prevent the particle's blind search, each component of V and each component of X are kept within the ranges [−V max, V max] and [−X max, X max], respectively. Inertia weight is not included in the original version [17]. The inclusion of an inertia weight in the PSO was first proposed by Shi and Eberhart [33]; then they subsequently investigated the effect of the inertial weight and maximum velocity on the performance of the particle swarm optimizer, and they provided guidelines for selecting these two parameters [34]. There are several strategies of inertial weight ω described in [20, 35–37], and there are some other ways to insure the convergence of the PSO, that is, a constriction factor [38] and a conditional random [39].

3. The Proposed Algorithm

The effectiveness and robustness of the MinMax k-means algorithm depend on initializations of parameters [15]. Reference [15] introduces a practical framework that extends the MinMax k-means to automatically adapt the exponent p to the data set. They concluded that if p max has been set, the programme can reach the lowest E max at p ∈ [p init, p max]. However, our experiments show that it is not always correct. We do experiments using the well-known data set Pendigits to support our claim. The description of this data set is given in Section 4. In our calculations the results of the MinMax k-means algorithm are the average over 100 runs of E max and E sum defined by (3) and (1), respectively. The results are reported in Table 1. One can see from Table 1 that E max and E sum have different values for different p and β. To address this problem we propose a new algorithm to find optimal values of parameters p and β which provide the minimum values of E max and E sum. We call it the PSO MinMax k-means algorithm.
Table 1

Comparative results in different parameter value on the Pendigit data set.

Value of parameter E max E sum
p max = 0.5, β = 0.3851062094
p max = 0.4, β = 0.3760961184
p max = 0.3, β = 0.3794860993
p max = 0.5, β = 0.11682671546
p max = 0.4, β = 0.11039463285
p max = 0.3, β = 0.1791860993
p max = 0.5, β = 0774461116
p max = 0.4, β = 0774061100
p max = 0.3, β = 0792460994
The proposed algorithm includes the PSO process and the MinMax k-means process. We utilize PSO to optimize the two parameters. That is to say, we find optimal parameters and put them into MinMax k-means process to obtain the minimum clustering errors. The specific method is illustrated as follows.

Algorithm 1 (PSO MinMax k-means algorithm).

See Figure 1.
Figure 1

Flowchart of the PSO MinMax k-means algorithm.

Step 1. Set up parameters of PSO, including iteration, population size, maximum velocity (V max), inertial weight (ω), and two learning factors (c 1, c 2); give the number of clusters k and initial weight w = 1/k. Step 2. Initialize each particle position (X ) and velocity (V ) randomly; randomly choose the center of each cluster. Note that each particle is a vector of the two parameters (p, β) for MinMax k-means algorithm. Therefore, X can be represented as X = (p , β ). Step 3. Calculate cluster assignments I(x ∈ C ), i = 1,2,…, N, by the Euclidean distance for each particle. Step 4. Calculate the weighted sum of the intracluster variances E by (4) for each particle. Step 5. Update cluster center C , j = 1,2,…, k, by the following equation: m = (1/|C |)∑ x for each particle. Step 6. Update weight value by (6). Step 7. If stopping conditions of MinMax k-means algorithm are not satisfied, go back to Step 3; otherwise, go to Step 8. Step 8. Calculate fitness value for each particle using formula (1) or (3). That is to say, clustering errors are the fitness functions. Step 9. Update the personal best P and the global best G. Step 10. According to the best positions P and G, update the velocity and position for each particle using formula (8). It should be noted that V cannot be larger than V max or smaller than −V max, and X cannot be larger than X max or smaller than −X max. Thus, Step 11. Record the final best p and β if the specified number of iterations is reached; otherwise, go back to Step 8. After performing all the steps above, we find the optimal parameters of MinMax k-means algorithm and then get the final clustering results by plugging the optimal parameter to MinMax k-means algorithm. Actually Steps 3 to 7 are the process of MinMax k-means algorithm.

4. Computational Results

In the following subsections, we report experimental results for five different states (State 1 to State 5) using both synthetic and real world data sets. We also compare the k-means, the MinMax k-means, and the PSO MinMax k-means algorithms using numerical results. In each state, the results of the MinMax k-means algorithm are tested in different parameters (p max, β), and we set p init = 0, p step = 0.01 as in [15]; we set the population size 20 and the generation number 100 in PSO MinMax k-means algorithm. The experiments in each state of PSO MinMax k-means algorithm just have two sets of results. The value of parameters in each set is one of the optimization values which can result in one minimum of the clustering errors.

4.1. Synthetic Data Sets

In this subsection, two synthetic data sets S 1 and S 2 from [40] are used to test algorithms. Typically, they are generated from a mixture of four or three bivariate Gaussian distributions on the plane coordinate system. Thus, a cluster takes the form of a Gaussian distribution. Particularly, all the Gaussian distributions have the covariance matrices of the form σ 2 I, where σ is the standard deviation. The first data set S 1 with four Gaussian distributions and 300 sample points is located at (−1,0), (1,0), (0,1), and (0, −1), respectively. Actually, σ takes the values of 0.4. As for data set S 2, we give three Gaussian distributions located at (1,0), (0,1), and (0, −1), with 400, 300, and 200 sample points, respectively. Therefore, S 2 represents the asymmetric situation where the clusters do not take the same shape with different number of sample points. The data sets are shown in Figure 2, respectively.
Figure 2

The sketch of two typical synthetic data sets: (a) S 1 and (b) S 2.

4.2. Real World Data Sets

Coil-20 is a data set [41], which contains 72 images taken from different angles for each of the 20 included objects. We use three subsets, Coil15, Coil18, and Coil19, with images from 15, 18, and 19 objects, respectively, as the data set in [15]. The data set includes 216 instances and each of the data instances has 1000 features. Yeast (UCI) [42] includes 1484 instances about the cellular localization sites of proteins and eight attributes. Proteins belong to ten categories. Five of the classes are extremely underrepresented and are not considered in our evaluation. The data set is unbalanced. Pendigits (UCI) [42] includes 10992 instances of handwritten digits (0–9) from the UCI repository [19] and 16 attributes. The data set is almost balanced. In the experiment, the sample data of Pendigits data set will be firstly normalized and then the algorithm will be implemented on the normalized data. Ecoli (UCI) [42] is composed of 336 protein localization sites for the E. coli bacterium and seven attributes. Proteins belong to eight different categories. Four of the classes are extremely underrepresented and are not considered in our evaluation. The data set is unbalanced. A summary of the data sets is provided in Table 4.
Table 4

The brief description of the real data sets.

Data set InstancesAttributesClassesBalanced
Coil2 21610003Yes
Yeast 135085No
Pendigits 109921610Almost
Ecoli 30774No

4.3. Performance Analysis

The comparison of the algorithms across the various data sets in five different states is shown in Tables 2, 3, and 5 –8. The values of parameters for PSO MinMax algorithm in Tables 2, 3, and 5 –8 are shown in Table 9 correspondingly. Based on the analysis shown in the tables, first, we find that our proposed algorithm can attain the lowest E max and E sum values in optimal parameters except in Table 5 (States 2–5). In States 2–5 of Table 5, our proposed algorithm did not attain the lowest E max. It lies in the drawbacks of PSO algorithm itself which just gets the local optimal solution. Sometimes our proposed algorithm has better E max than k-means algorithm and the original MinMax k-means algorithm (see Tables 3 and 7) and sometimes we have both better E max and E sum than other algorithms. We also find that E max and E sum cannot reach the lowest value simultaneously.
Table 2

Comparative results on S 1 data set.

MethodInitial State 1Initial State 2Initial State 3Initial State 4Initial State 5Mean error
E max E sum E max E sum E max E sum E max E sum E max E sum E max E sum
k-means 90.8431 329.4133 90.8431329.425890.8431329.425890.8431329.425890.8431 329.4133 90.8431329.4208
MinMax (p max = 0.5, β = 0.3) 87.1170 329.6677 87.1170 329.6677 87.1170 329.6677 87.1170 329.6677 87.1170 329.6677 87.1170 329.6677
MinMax (p max = 0.5, β = 0.1) 87.1170 329.6677 87.1170 329.6677 87.1170 329.6677 87.1170 329.6677 87.1170 329.6677 87.1170 329.6677
MinMax (p max = 0.5, β = 0) 87.1170 329.6352 87.1170 329.6352 87.1170 329.6352 87.1170 329.6352 87.1170 329.6352 87.1170 329.6352
MinMax (p max = 0.3, β = 0.3)89.3927329.4572 87.1170 329.5055 87.1170 329.505589.3927329.457289.3927329.457288.4824329.4765
MinMax (p max = 0.3, β = 0.1)89.3927329.4766 87.1170 329.5055 87.1170 329.505589.3927329.457289.3927329.457288.4824329.4804
MinMax (p max = 0.3, β = 0)89.3927329.4572 87.1170 329.5055 87.1170 329.505589.3927329.457289.3927329.457288.4824329.4765
MinMax (p max = 0.1, β = 0.3)90.0813329.418888.5673329.461688.5673329.461690.8431 329.4133 90.0813329.418889.6281329.4348
MinMax (p max = 0.1, β = 0.1)90.0813329.418888.5673329.461688.5673329.461690.8431 329.4133 90.0813329.418889.6281329.4348
MinMax (p max = 0.1, β = 0)90.0813329.418888.5673329.461688.5673329.461690.8431 329.4133 90.0813329.418889.6281329.4348
MinMax (p max = 0, β = 0)90.8431 329.4133 90.8431329.425890.8431329.425890.8431329.425890.8431 329.4133 90.8431329.4208
PSO MinMax 90.8431 329.4133 90.8431 329.4133 90.8431 329.4133 90.0813 329.4133 90.8431 329.4133 90.8431 329.4133
PSO MinMax 87.1170 329.6207 87.1170 329.5857 87.1170 329.5857 87.1170 329.4833 87.1170 329.5857 87.1170 329.5722
Table 3

Comparative results on S 2 data set.

Method Initial State 1Initial State 2 Initial State 3Initial State 4Initial State 5Mean error
E max E sum E max E sum E max E sum E max E sum E max E sum E max E sum
k-means 62.5878 105.5999 62.5878 105.5999 62.5878 105.5999 62.5878 105.5999 62.5878 105.5999 62.5878 105.5999
MinMax (p max = 0.5, β = 0.3)54.0427109.092754.0427109.092754.0427109.092754.0427109.092754.0427109.092754.0427109.0927
MinMax (p max = 0.5, β = 0.1)54.0464109.122654.0464109.122654.0464109.122654.0464109.122654.0464109.122654.0464109.1226
MinMax (p max = 0.5, β = 0)54.0464109.122654.0464109.122654.0464109.122654.0464109.122654.0464109.122654.0464109.1226
MinMax (p max = 0.3, β = 0.3)57.3660106.693757.3660106.693757.3660106.693757.3660106.693757.3660106.693757.3660106.6937
MinMax (p max = 0.3, β = 0.1)57.3660106.693757.3660106.693757.3660106.693757.3660106.693757.3660106.693757.3660106.6937
MinMax (p max = 0.3, β = 0)57.3660106.693757.3660106.693757.3660106.693757.3660106.693757.3660106.693757.3660106.6937
MinMax (p max = 0.1, β = 0.3)61.0903105.649061.0903105.649061.0903105.649061.0903105.649061.0903105.649061.0903105.6490
MinMax (p max = 0.1, β = 0.1)61.0903105.649061.0903105.649061.0903105.649061.0903105.649061.0903105.649061.0903105.6490
MinMax (p max = 0.1, β = 0)61.0903105.649061.0903105.649061.0903105.649061.0903105.649061.0903105.649061.0903105.6490
MinMax (p max = 0, β = 0)62.5878 105.5999 62.5878 105.5999 62.5878 105.5999 62.5878 105.5999 62.5878 105.5999 62.5878 105.5999
PSO MinMax 62.5878 105.5999 62.5878 105.5999 62.5878 105.5999 62.5878 105.5999 62.5878 105.5999 62.5878 105.5999
PSO MinMax 52.1071 110.8688 52.1071 110.8688 52.1071 110.8688 52.1071 110.8688 52.1071 110.868852.1071 110.8688
Table 5

Comparative results on Coil2 data set.

Method Initial State 1Initial State 2Initial State 3Initial State 4Initial State 5Mean error
E max E sum E max E sum E max E sum E max E sum E max E sum E max E sum
k-means 58.8338 154.0274 58.8338 154.0274 78.2208156.605768.2271156.9333122.1251158.163077.2501155.9514
MinMax (p max = 0.5, β = 0.3) 58.8645154.104958.8645154.104958.8645154.104958.8645154.104959.0939156.514158.9104154.5867
MinMax (p max = 0.5, β = 0.1) 57.0210154.7626 57.0210 154.7626 57.0210 154.7626 57.0210 154.7626 57.0210 154.7626 57.0210 154.7626
MinMax (p max = 0.5, β = 0) 58.8645154.104958.8645154.104958.8645154.104958.8645154.104958.8645154.104958.8645154.1049
MinMax (p max = 0.3, β = 0.3) 58.8338 154.0274 58.8338 154.0274 58.8338 154.0274 58.8338154.027461.6945156.395459.4061154.5010
MinMax (p max = 0.3, β = 0.1) 58.8338 154.0274 58.8338 154.0274 58.8338 154.0274 58.8338154.027458.8338 154.0274 58.8338154.0274
MinMax (p max = 0.3, β = 0) 58.8338 154.0274 58.8338 154.0274 58.8338 154.0274 58.8338154.027458.8338 154.0274 58.8338154.0274
MinMax (p max = 0.1, β = 0.3) 58.8338 154.0274 58.8338 154.0274 58.8338 154.0274 77.6769153.8029122.1251158.337575.2607154.8445
MinMax (p max = 0.1, β = 0.1) 58.8338 154.0274 58.8338 154.0274 58.8338 154.0274 77.6769153.8029122.1251158.607475.2607154.8985
MinMax (p max = 0.1, β = 0) 58.8338 154.0274 58.8338 154.0274 78.2208156.605777.6769153.8029122.1251158.607479.1381155.4142
MinMax (p max = 0, β = 0) 58.8338 154.0274 58.8338 154.0274 78.2208156.605768.2271156.9333122.1251158.163077.2501155.9514
PSO MinMax 58.8338 154.0274 58.8338 154.0274 58.8338 154.0274 78.5526 153.7484 58.8338 154.0274 62.7816 153.9716
PSO MinMax 56.9911 154.684658.8338 154.0274 58.8338 154.0274 58.8338154.027458.8338 154.0274 58.4653154.1588
Table 6

Comparative results on Yeast data set.

Method Initial State 1Initial State 2Initial State 3Initial State 4Initial State 5Mean error
E max E sum E max E sum E max E sum E max E sum E max E sum E max E sum
k-means 11.8980 51.1611 13.618850.992019.448553.528313.618850.992013.583750.990714.433651.5328
MinMax (p max = 0.5, β = 0.3) 16.062453.810311.167251.384716.062453.810311.167251.384711.070151.379513.105952.3539
MinMax (p max = 0.5, β = 0.1) 10.644051.386721.160264.824110.644051.386721.160264.824111.070151.379514.935756.7602
MinMax (p max = 0.5, β = 0) 10.790551.374512.042651.248110.965051.328712.042651.248111.698851.249711.507951.2898
MinMax (p max = 0.3, β = 0.3) 10.971951.291311.691751.268011.022351.295811.691751.272011.681051.226811.411751.2708
MinMax (p max = 0.3, β = 0.1) 11.006751.291212.010551.248211.056551.289912.010551.248211.681051.226811.553051.2609
MinMax (p max = 0.3, β = 0) 10.971951.291311.991551.248211.022351.295811.991551.248211.681051.226811.531655.2621
MinMax (p max = 0.1, β = 0.3) 11.490651.202412.641351.124011.586351.187912.746451.093412.756451.020612.244251.1260
MinMax (p max = 0.1, β = 0.1) 11.490651.202412.686951.123611.595851.187912.750551.096212.756451.020612.256051.1261
MinMax (p max = 0.1, β = 0) 11.490651.202412.686951.123611.595851.191312.750551.096212.756451.020612.256051.1261
MinMax (p max = 0, β = 0) 11.8980 51.1611 13.618850.992019.448553.528313.618850.992013.583750.990714.433651.5328
PSO MinMax 11.8980 51.1611 13.0750 50.9871 13.3736 50.9891 13.0750 50.9871 13.1598 50.9869 12.9163 51.0233
PSO MinMax 10.6149 51.5056 10.6360 51.3925 10.5965 51.4087 10.6205 51.4043 10.9874 51.4089 10.6911 51.4240
Table 7

Comparative results on Pendigit data set.

Method Initial State 1Initial State 2Initial State 3Initial State 4Initial State 5Mean error
E max E sum E max E sum E max E sum E max E sum E max E sum E max E sum
k-means 2044367183151716147814322 59899 78396173312262 59628 1400761984
MinMax (p max = 0.5, β = 0.3) 8969636349160622338955617677311619111095163265909662562
MinMax (p max = 0.5, β = 0.1) 7355618749389620971565473892232967724726142752721636770076
MinMax (p max = 0.5, β = 0) 7480617929779619031008560836753161839689160234835361320
MinMax (p max = 0.3, β = 0.3) 108146155410842614741046360617760061809699460181934261127
MinMax (p max = 0.3, β = 0.1) 75996173710842614751047460611760061809699460179870161162
MinMax (p max = 0.3, β = 0) 75996173710842614751047460611760061809699460179870161162
MinMax (p max = 0.1, β = 0.3) 13231609861323560924125215999077366174011434597131163160670
MinMax (p max = 0.1, β = 0.1) 1104161520132356092412542 59989 77366174011434597131119760777
MinMax (p max = 0.1, β = 0) 1323160986132356092412542 59989 77366174011434597131163660670
MinMax (p max = 0, β = 0) 2044367183151716147814322 59899 78396173312262 59628 1400761984
PSO MinMax 7319 60047 13394 60907 14322 59899 7296 60055 12262 59628 10919 60107
PSO MinMax 6678 60393 6887 61424 8491 63718 6672 60396 6669 60393 7079 61265
Table 8

Comparative results on Ecoli data set.

Method Initial State 1Initial State 2Initial State 3Initial State 4Initial State 5Mean error
E max E sum E max E sum E max E sum E max E sum E max E sum E max E sum
k-means 6.7989 15.3683 7.181415.36726.058915.76075.684115.78286.798915.36836.503915.5294
MinMax (p max = 0.5, β = 0.3) 4.7952 15.7294 4.7952 15.7294 4.7952 15.7294 4.7952 15.7294 4.7952 15.7294 4.7952 15.7294
MinMax (p max = 0.5, β = 0.1) 5.294415.70085.294415.7008 4.7952 15.7281 5.038415.86465.294415.70085.143415.7390
MinMax (p max = 0.5, β = 0) 5.434716.19185.434716.19185.173815.71095.173815.71095.434716.19185.330315.9994
MinMax (p max = 0.3, β = 0.3) 5.536215.48525.536215.48525.684115.79245.465215.83005.536215.48525.551615.6156
MinMax (p max = 0.3, β = 0.1) 5.536215.48525.536215.48525.684115.79245.465215.83005.536215.48525.551615.6156
MinMax (p max = 0.3, β = 0) 5.536215.48525.536215.48525.684115.79245.465215.83005.536215.48525.551615.6156
MinMax (p max = 0.1, β = 0.3) 6.294115.39436.798915.36836.058915.76075.684115.78226.7989 15.3683 6.327015.5348
MinMax (p max = 0.1, β = 0.1) 5.867215.43326.798915.36836.058915.76075.684115.78226.7989 15.3683 6.241615.5425
MinMax (p max = 0.1, β = 0) 5.867215.43326.798915.36836.058915.76075.684115.78226.7989 15.3683 6.241615.5425
MinMax (p max = 0, β = 0) 6.7989 15.3683 7.181415.36726.058915.76075.684115.78286.7989 15.3683 6.503915.5294
PSO MinMax 6.7989 15.3683 6.9020 15.3664 4.7952 15.7281 5.6920 15.4645 6.7989 15.3683 6.1974 15.4591
PSO MinMax 4.7952 15.7281 4.7952 15.7294 4.7952 15.7294 4.7952 15.7294 4.7952 15.7281 4.7952 15.7289
Table 9

Value of parameters in PSO MinMax algorithm.

Data set State 1State 2State 3State 4State 5
p β p β p β p β p β
S 1 00.39860.03420.13520.03500.47100.03310.28760.03310.2876
0.40520.18710.38890.24070.37120.11940.28820.10000.36240.3775

S 2 0.00910.00760.00880.415100.403400.40690.01160.0604
0.50000.23510.50000.11760.50000.08440.50000.23520.50000.2933

Coil20.43820.25970.32830.44050.05700.00980.08640.32740.40670.4891
0.50000.33700.32830.44050.05700.00980.40790.30100.40670.4891

Yeast0.05700.00980.02840.46020.00810.16610.03070.30090.02040.3449
0.49060.48700.49940.50000.49060.40300.50000.50000.50000.5000

Pendigit0.186200.08520.362200.41250.19000.044600.3126
0.50000.39960.27540.03150.50000.26090.50000.26090.50000.4477

Ecoli00.07930.05360.31770.42620.17830.18980.05060.08880.2523
0.47730.28230.48480.44480.47560.30740.42130.16100.48900.4521
Second, it follows from the tables that the values of clustering errors in k-means and in MinMax k-means are equal when setting p max = 0, β = 0, implying that the k-means algorithm can be considered as a special case of the MinMax k-means algorithm. Third, the proper parameter in our algorithm is not a single value. Here, we just give one of the values. Hence, MinMax k-means algorithm and our proposed algorithm can both reach the lowest clustering errors on different parameter value. Fourth, for the operation time of the algorithm, it is easy to observe that k-means algorithm consumes the least time. The operation time of our proposed algorithm depends on the population size (s), the number of generation (n), and the speed of convergence. For convenience, denote the running time of single MinMax k-means algorithm as t; then the operation time of our proposed algorithm is snt. Comparing the running time of MinMax k-means to that of our proposed algorithm, it is hard to identify which method consumes less time. For example, when we perform Coil2 data set (State 1) with p max = 0.5, β = 0.3, the running time of MinMax k-means is 0.7403 s, and the running time of our proposed algorithm (s = 3, n = 2) is 0.6653 s. However, when we perform experiment on Ecoli data set (State 1), the running time is 0.3637 s for MinMax k-means (p max = 0.5, β = 0.3) and the running time for our proposed algorithm (s = 10, n = 5) is 0.7848 s. Finally, Tzortzis and Likas [15] stated that high p value forces clusters with large variance to lose most or even all of their instances as their enormous weights excessively distance the instances from their centers. Hence, their solution has the following properties: whenever an empty or singleton cluster emerges, no matter if p max has been reached or not, decreasing p by step reverts back to the cluster assignments corresponding to the previous p value and clustering must be resumed from there. Our algorithm can resolve the problem automatically. We do experiment on all data sets mentioned in this paper. The result is as follows. On Ecoli data set (State 5) when p max = 0.5, β = 0, MinMax k-means converges for p = 0.44 with E max = 5.4347, E sum = 16.1918. On the other hand, if p ≥ 0.45, β = 0, we have E max = 49.9010, E sum = 49.9010. It is clear that when p is bigger than convergence value, the corresponding values of E max and E sum are much bigger than that of convergence value. The situations on States 1–4 are similar to State 5. On Pendigits data set, when β = 0, MinMax k-means algorithm converges for p < 0.5. If p is bigger than the convergence value, we get E max = E sum, which is much bigger than that of proper p. Similar situation occurs on Yeast data set. In summary, we find that when p is bigger than the convergence value, we get E max = E sum, which are much bigger than that of proper p; meanwhile the data sets are clustered just for one class. Based on the above analysis, our proposed algorithm cannot choose the corresponding parameter in PSO process. Therefore our proposed algorithm cannot be chosen singleton or empty for one of the clusters.

5. Conclusions

We modified the MinMax k-means algorithm to attain the lowest clustering errors automatically. Firstly, we use PSO to search the optimal parameters which can result in the minimum errors. Then we plug the parameters obtained by PSO into MinMax k-means algorithm. Experiments are tested on different data sets in different initial states, and the results show that our proposed algorithm is efficient in most situations. As for future work, we plan to accelerate the proposed algorithm. A possible direction is data sets processing. For example, we can use the method of PCA. We also plan to achieve time efficiency of the PSO process, since many iterations in the algorithm may have repetitive calculations.
  3 in total

1.  Frequency-sensitive competitive learning for scalable balanced clustering on high-dimensional hyperspheres.

Authors:  Arindam Banerjee; Joydeep Ghosh
Journal:  IEEE Trans Neural Netw       Date:  2004-05

Review 2.  Survey of clustering algorithms.

Authors:  Rui Xu; Donald Wunsch
Journal:  IEEE Trans Neural Netw       Date:  2005-05

3.  The global kernel k-means algorithm for clustering in feature space.

Authors:  Grigorios F Tzortzis; Aristidis C Likas
Journal:  IEEE Trans Neural Netw       Date:  2009-05-29
  3 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.