Literature DB >> 28042291

A Self-Adaptive Fuzzy c-Means Algorithm for Determining the Optimal Number of Clusters.

Min Ren1, Peiyu Liu2, Zhihao Wang2, Jing Yi2.   

Abstract

For the shortcoming of fuzzy c-means algorithm (FCM) needing to know the number of clusters in advance, this paper proposed a new self-adaptive method to determine the optimal number of clusters. Firstly, a density-based algorithm was put forward. The algorithm, according to the characteristics of the dataset, automatically determined the possible maximum number of clusters instead of using the empirical rule [Formula: see text] and obtained the optimal initial cluster centroids, improving the limitation of FCM that randomly selected cluster centroids lead the convergence result to the local minimum. Secondly, this paper, by introducing a penalty function, proposed a new fuzzy clustering validity index based on fuzzy compactness and separation, which ensured that when the number of clusters verged on that of objects in the dataset, the value of clustering validity index did not monotonically decrease and was close to zero, so that the optimal number of clusters lost robustness and decision function. Then, based on these studies, a self-adaptive FCM algorithm was put forward to estimate the optimal number of clusters by the iterative trial-and-error process. At last, experiments were done on the UCI, KDD Cup 1999, and synthetic datasets, which showed that the method not only effectively determined the optimal number of clusters, but also reduced the iteration of FCM with the stable clustering result.

Entities:  

Mesh:

Year:  2016        PMID: 28042291      PMCID: PMC5153549          DOI: 10.1155/2016/2647389

Source DB:  PubMed          Journal:  Comput Intell Neurosci


1. Introduction

Cluster analysis has a long research history. Due to its advantage of learning without a priori knowledge, it is widely applied in the fields of pattern recognition, image processing, web mining, spatiotemporal database application, business intelligence, and so forth. Clustering is often as unsupervised learning and aims to partition objects in a dataset into several natural groupings, namely, the so-called clusters, such that objects within a cluster tend to be similar while objects belonging to different clusters are dissimilar. Generally, the datasets from different application fields vary in feature, and the purposes of clustering are multifarious. Therefore, the best method of cluster analysis depends on datasets and purposes of use. There is no universal clustering technology that can be widely applicable to the diverse structures presented by various datasets [1]. According to the accumulation rules of objects in clusters and the methods of applying these rules, clustering algorithms are divided into many types. However, for most clustering algorithms including partitional clustering and hierarchical clustering, the number of clusters is a parameter needing to be preset, to which the quality of clustering result is closely related. In practical application, it usually relies on users' experience or background knowledge in related fields. But in most cases, the number of clusters is unknown to users. If the number is assigned too large, it may result in more complicated clustering results which are difficult to be explained. On the contrary, if it is too small, a lot of valuable information in clustering result may be lost [2]. Thus, it is still a fundamental problem in the research of cluster analysis to determine the optimal number of clusters in a dataset [3]. Contributions of this paper are as follows. (1) A density-based algorithm is proposed, which can quickly and succinctly generate high-quality initial cluster centroids instead of choosing randomly, so as to stabilize the clustering result and also quicken the convergence of the clustering algorithm. Besides, this algorithm can automatically estimate the maximum number of clusters based on the features of a dataset, hereby determining the search range for estimating the optimal number of clusters and effectively reducing the iterations of the clustering algorithm. (2) Based on the features of compactness within a cluster and separation between clusters, a new fuzzy clustering validity index (CVI) is defined in this paper avoiding its value close to zero along with the number of clusters tending to the number of objects and obtaining the optimal clustering result. (3) A self-adaptive method that iteratively uses improved FCM algorithm to estimate the optimal number of clusters was put forward.

2. Related Work

The easiest method of determining the number of clusters is data visualization. For the dataset that can be effectively mapped to a 2-dimensional Euclidean space, the number of clusters can be intuitively acquired through the distribution graph of data points. However, for high-dimensional and complicated data, this method is unserviceable. Rodriguez and Laio proposed a clustering algorithm based on density peaks, declaring that it was able to detect nonspherical clusters and to automatically find the true number of clusters [4]. But, in fact, the number of cluster centroids still needs to be selected artificially according to the decision graph. Next, relevant technologies to determine the optimal number of clusters are summarized below.

2.1. Clustering Validity Index Based Method

Clustering validity index is used to evaluate the quality of partitions on a dataset generated by the clustering algorithm. It is an effective method to construct an appropriate clustering validity index to determine the number of clusters. The idea is to assign different values of the number of clusters c within a certain range, then run fuzzy clustering algorithm on the dataset, and finally to evaluate the results by clustering validity index. When the value of clustering validity index is of the maximum or the minimum or an obvious inflection point appears, the corresponding value of c is the optimal number of clusters c opt. So far, researchers have put forward lots of fuzzy clustering validity indices, divided into the following two types. (1) Clustering Validity Index Based on Fuzzy Partition. These indices are according to such a point of view that for a well-separated dataset, the smaller the fuzziness of fuzzy partition is, the more reliable the clustering result is. Based on it, Zadeh, the founder of fuzzy sets, put forward the first clustering validity index degree of separation in 1965 [5]. But its discriminating effect is not ideal. In 1974, Bezdek put forward the concept of partition coefficient (PC) [6] which was the first practical generic function for measuring the validity of the fuzzy clustering and subsequently proposed another clustering validity function partition entropy (PE) [7] closely related to partition coefficient. Later, Windham defined proportion exponent by use of the maximum value of a fuzzy membership function [8]. Lee put forward a fuzzy clustering validity index using the distinguishableness of clusters measured by the object proximities [9]. Based on the Shannon entropy and fuzzy variation theory, Zhang and Jiang put forward a new fuzzy clustering validity index taking account of the geometry structure of the dataset [10]. Saha et al. put forward an algorithm based on differential evolution for automatic cluster detection, which well evaluated the validity of the clustering result [11]. Yue et al. partitioned the original data space into a grid-based structure and proposed a cluster separation measure based on grid distances [12]. The clustering validity index based on fuzzy partition is only related to the fuzzy degree of membership and has the advantages of simpleness and small calculating amount. But it lacks direct relation with some structural features of the dataset. (2) Clustering Validity Index Based on the Geometry Structure of the Dataset. These indices are based on such a point of view that for a well-separated dataset, every cluster should be compact and separated from each other as far as possible. The ratio of compactness and separation is used as the standard of clustering validity. This type of representative clustering validity indices include Xie-Beni index [13], Bensaid index [14], and Kwon index [15]. Sun et al. proposed a new validity index based on a linear combination of compactness and separation and inspired by Rezaee's validity [16]. Li and Yu defined new compactness and separation and put forward a new fuzzy clustering validity function [17]. Based on fuzzy granulation-degranulation, Saha and Bandyopadhyay put forward a fuzzy clustering validity function [18]. Zhang et al. adopted Pearson correlation to measure the distance and put forward a validity function [19]. Kim et al. proposed a clustering validity index for GK algorithm based on the average value of the relative degrees of sharing of all possible pairs of fuzzy clusters [20]. Rezaee proposed a new validity index for GK algorithm to overcome the shortcomings of Kim's index [21]. Zhang et al. proposed a novel WGLI to detect the optimal cluster number, using global optimum membership as the global property and modularity of bipartite network as the local independent property [22]. The clustering validity index based on the geometric structure of the dataset considers both the fuzzy degree of membership and the geometric structure, but its membership function is quite complicated with large calculating amount. Based on clustering validity index, the optimal number of clusters is determined through exhaustive search. In order to increase the efficiency of estimating the optimal number of clusters c opt, the search range of c opt must be set; that is, c max⁡ is the maximum number of clusters, assigned to meet the condition c opt ≤ c max⁡. Most researchers used the empirical rule , where n is the number of data in the dataset. For this problem, theoretical analysis and example verification were conducted in [23], which indicated that it was reasonable in a sense. However, apparently, this method has the following disadvantages. (1) Each c must be tried in turn, which will cause a huge calculation. (2) For each c, it cannot be guaranteed that the clustering result is the globally optimal solution. (3) When noise and outliers are existing, the reliability of clustering validity index is weak. (4) For some datasets like FaceImage [24], if , the empirical rule will be invalid. Researches showed that due to the diversified data types and structures, no universal fuzzy clustering validity index can be applicable to all datasets. The research is and will still be carried on urgently.

2.2. Heuristic Method

Some new clustering algorithms have been proposed in succession recently. The main idea is to use some criteria to guide the clustering process, with the number of clusters being adjusted. In this way, while the clustering is completed, the appropriate number of clusters can be obtained as well. For example, X-means algorithm [25] based on the split hierarchical clustering is representative. Contrary to the aforesaid process, RCA [26] determined the actual number of clusters by a process of competitive agglomeration. Combining the single-point iterative technology with hierarchical clustering, a similarity-based clustering method (SCM) [27] was proposed. A mercer kernel-based clustering [28] estimated the number of clusters by the eigenvectors of a kernel matrix. A clustering algorithm based on maximal θ-distant subtrees [29] detected any number of well-separated clusters with any shapes. In addition, Frey and Dueck put forward an affinity propagation clustering algorithm (AP) [24] which generated high-quality cluster centroids via the message passing between objects to determine the optimal number of clusters of large-scale data. Shihong et al. proposed a Gerschgorin disk estimation-based criterion to estimate the true number of clusters [30]. José-García and Gómez-Flores presented an up-to-date review of all major nature-inspired metaheuristic algorithms used thus far for automatic clustering, determining the best estimate of the number of clusters [31]. All these have widened the thoughts for relevant researches.

3. Traditional FCM for Determining the Optimal Number of Clusters

Since Ruspini first introduced the theory of fuzzy sets into cluster analysis in 1973 [32], different fuzzy clustering algorithms have been widely discussed, developed, and applied in various areas. FCM [33] is one of the most commonly used algorithms. FCM was first presented by Dunn in 1974. Subsequently, Bezdek introduced weighted index to the fuzzy degree of membership [34], further developing FCM. FCM divides the dataset X into c fuzzy clusters. This algorithm holds that each object belongs to a certain cluster with a different degree of membership, that is, a cluster is considered as a fuzzy subset on the dataset. Assume that X = {x 1, x 2,…, x } is a finite dataset, where x = (x , x ,…, x ) is an l-dimensional object and x is the jth property of the ith object. C = {C 1, C 2,…, C } denotes c cluster(s). V = {v 1, v 2,…, v } represents c l-dimensional cluster centroid(s), where v = (v , v ,…, v ). U = (u )( is a fuzzy partition matrix, and u is the degree of membership of the ith object in the kth cluster, where ∑ u = 1, ∀i = 1,…, n. The objective function is the quadratic sum of weighed distances from the samples to the cluster centroid in each cluster; that is,where d = ‖x − v ‖ shows the Euclidean distance between the ith object and the kth cluster centroid. m (m ∈ [1, ∞)) is a fuzziness index, controlling the fuzziness of the memberships. As the value of m becomes progressively higher, the resulting memberships become fuzzier [35]. Pal and Bezdek advised that m should be between 1.5 and 2.5, and usually let m = 2 if without any special requirement [36]. According to the clustering criterion, appropriate fuzzy partition matrix U and cluster centroid V are obtained to minimize the objective function J . Based on the Lagrange multiplier method, U and V are, respectively, calculated by the formulas below: FCM algorithm is carried out through an iterative process of minimizing the objective function J , with the update of U and V. The specific steps are as follows.

Step 1 .

Assign the initial value of the number of clusters c, fuzziness index m, maximum iterations I max⁡, and threshold ξ.

Step 2 .

Initialize the fuzzy partition U (0) randomly according to the constraints of the degree of membership.

Step 3 .

At the t-step, calculate c cluster centroids V ( according to (3).

Step 4 .

Calculate the objective function J ( according to (1). If |J ( − J (| < ξ or t > I max⁡, then stop; otherwise continue to Step  5.

Step 5 .

Calculate U ( according to (2) and return to Step  3. At last, each object can be arranged into one cluster in accordance with the principle of the maximum degree of membership. The advantages of FCM may be summarized as simple algorithm, quick convergence, and easiness to be extended. Its disadvantages lie in the selection of the initial cluster centroids, the sensitivity to noise and outliers, and the setting of the number of clusters, which have a great impact on the clustering result. As the random selection of the initial cluster centroids cannot ensure the fact that FCM converges to an optimal solution, different initial cluster centroids are used for multiple running of FCM; otherwise they are determined by using of other fast algorithms. The traditional method to determine the optimal number of clusters of FCM is to set the search range of the number of clusters, run FCM to generate clustering results of different number of clusters, select an appropriate clustering validity index to evaluate clustering results, and finally obtain the optimal number of clusters according to the evaluation result. The method is composed of the following steps. Input the search range [c min⁡, c max⁡]; generally, c min⁡ = 2 and . For each integer kn ∈ [c min⁡, c max⁡]. Step  2.1. Run FCM. Step  2.2. Calculate clustering validity index. Compare all values of clustering validity index. kn corresponding to the maximum or minimum value is the optimal number of clusters c opt. Output c opt, the optimal value of clustering validity index, and clustering result.

4. The Proposed Methods

4.1. Density-Based Algorithm

Considering the large influence of the randomly selected initial cluster centroid on the clustering result, a density-based algorithm is proposed to select the initial cluster centroids, and the maximum number of clusters can be estimated at the same time. Some related terms will be defined at first.

Definition 1 (local density).

The local density ρ of object x is defined aswhere d is the distance between two objects x and x and dc is a cutoff distance. The recommended approach is to sort the distance between any two objects in descending order and then assign dc as the value corresponding to the first p% of the sorted distance (appropriately p ∈ [2,5]). It shows that the more objects with a distance from x less than dc there are, the bigger the value of ρ is. The cutoff distance finally can decide the number of initial cluster centroids, namely, the maximum number of clusters. The density-based algorithm is robust with respect to the choice of dc.

Definition 2 (core point).

Assume that there is an object x ∈ X, if x ∉ C and ρ = max⁡(ρ ), k = 1,…, c, j = 1,…, n, then x is a core point.

Definition 3 (directly density-reachable).

Assume that there are two objects x , x ∈ X, if their distance d < dc, then x is directly density-reachable to x and vice versa.

Definition 4 (density-reachable).

Assume that X is a dataset and objects x , x ,…, x , x belong to it. If, for each i ∈ [s, t − 1], x is directly density-reachable to x , then x is density-reachable to x .

Definition 5 (neighbor).

Neighbors of an object x are those who are directly density-reachable or density-reachable to it, denoted as Neighbor (x ).

Definition 6 (border point).

An object is called a border point if it has no neighbors. The example is shown in Figure 1.
Figure 1

An example. Point A is of the highest local density. If A does not belong to any cluster, then A is a core point. Points B, C, D, and E are directly density-reachable to point A. Point F is density-reachable to point A. Point H is a border point.

The selection principle of initial cluster centroids of density-based algorithm is that, usually, a cluster centroid is an object with higher local density, surrounded by neighbors with lower local density than it, and has a relatively large distance from other cluster centroids [4]. Density-based algorithm can automatically select the initial cluster centroids and determine the maximum number of clusters c max⁡ according to local density. The pseudocode is shown in Algorithm 1. These cluster centroids obtained are sorted in descending order according to local density. Figure 2 demonstrates the process of density-based algorithm.
Algorithm 1

Density-based algorithm.

Figure 2

Demonstration of the process of density-based algorithm. (a) is the initial data distribution of a synthetic dataset. The dataset consists of two pieces of 2-dimensional Gaussian distribution data with centroids, respectively, as (2, 3) and (7, 8). Each class has 100 samples. In (b), the blue circle represents the highest density core point as the centroid of the first cluster, and the red plus sign represents the object belonging to the first cluster. In (c), the red circle represents the core point as the centroid of the second cluster, and the blue asterisk represents the object belonging to the second cluster. In (d), the purple circle represents the core point as the centroid of the third cluster, the green times sign represents the object belonging to the third cluster, and the black dot represents the final border point which does not belong to any cluster. According to a certain cutoff distance, the maximum number of clusters is 3. If calculated in accordance with the empirical rule, the maximum number of clusters should be 14. Therefore, the algorithm can effectively reduce the iteration of FCM algorithm operation.

4.2. Fuzzy Clustering Validity Index

Firstly based on the geometric structure of the dataset, Xie and Beni put forward the Xie-Beni fuzzy clustering validity index [13], which tried to find a balance point between the fuzzy compactness and separation so as to acquire the optimal cluster result. The index is defined as In the formula, the numerator is the average distance from various objects to centroids, used to measure the compactness, and the denominator is the minimum distance between any two centroids, measuring the separation. However, Bensaid et al. found that the size of each cluster had a large influence on Xie-Beni index and put forward a new index [14], which was insensitive to the number of objects in each cluster. Bensaid index is defined aswhere n is the fuzzy cardinality of the kth cluster and defined as ∑ u ‖x − v ‖2 shows the variation of the kth fuzzy cluster. Then, the compactness is computed as ∑ ‖v − v ‖2 denotes the separation of the kth fuzzy cluster, defined as the sum of the distances from its cluster centroid to the centroids of other c − 1 clusters. Becausethis index is the same as the Xie-Beni index. When c → n, the index value will be monotonically decreased, close to 0, and will lose robustness and judgment function for determining the optimal number of clusters. Thus, this paper improves Bensaid index and proposes a new index V : The numerator represents the compactness of the kth cluster, where n is its fuzzy cardinality. Its second item, an introduced punishing function, denotes the distance from the cluster centroid of the kth cluster to the average of all cluster centroids, which can eliminate the monotonically decreasing tendency as the number of clusters increases to n. The denominator represents the mean distance from the kth cluster centroid to other cluster centroids, which is used for measuring the separation. The ratio of the numerator and the denominator thereof represents the clustering effect of the kth cluster. The clustering validity index is defined as the sum of the clustering effect (the ratio) of all clusters. Obviously, the smaller the value is, the better the clustering effect of the dataset is, and the corresponding c to the minimum value is the optimal number of clusters.

4.3. Self-Adaptive FCM

In this paper, the iterative trial-and-error process [16] is still used to determine the optimal number of clusters by self-adaptive FCM (SAFCM). The pseudocode of SAFCM algorithm is described in Algorithm 2.
Algorithm 2

SAFCM.

5. Experiment and Analysis

This paper selects 8 experimental datasets, among which 3 datasets come from UCI datasets, respectively, Iris, Wine, and Seeds, a dataset (SubKDD) is randomly selected from KDD Cup 1999 shown in Table 1, and the remaining 4 datasets are synthetic datasets shown in Figure 3. SubKDD includes normal, 2 kinds of Probe attack (ipsweep and portsweep) and 3 kinds of DoS attack (neptune, smurf, and back). The first synthetic dataset (SD1) consists of 20 2-dimensional Gaussian distribution data with 10 samples. Their covariance matrixes are second-order unit matrix I 2. The structural feature of the dataset is that the distance between any two clusters is large, and the number of classes is greater than . The second synthetic dataset (SD2) consists of 4 2-dimensional Gaussian distribution data. The cluster centroids are, respectively, (5,5), (10,10), (15,15), and (20,20), each containing 500 samples and the covariance matrix of each being 2I 2. The structural feature of the dataset is of short intercluster distance with a small overlap. The third synthetic dataset (SD3) has a complicated structure. The fourth synthetic dataset (SD4) is a nonconvex one.
Table 1

The data type and distribution of SubKDD.

Attack behavior Number of samples
normal200
ipsweep50
portsweep50
neptune200
smurf300
back50
Figure 3

Four synthetic datasets.

In this paper, the numeric value x is normalized aswhere and s denote the mean of the jth attribute value and its standard deviation, respectively; then For the categorical attribute of SubKDD, the simple matching is used for the dissimilarity measure, that is, 0 for identical values and 1 for different values.

5.1. Experiment of c max⁡ Simulation

In [37], it was proposed that the number of clusters generated by AP algorithm could be selected as the maximum number of clusters. So this paper estimates the value of c max⁡ by the empirical rule, AP algorithm, and density-based algorithm. The specific experimental results are shown in Table 2. Let p = p in AP algorithm. dc is, respectively, selected as the distance of the first 1%, 2%, 3%, and 5%.
Table 2

c max⁡ estimated by several methods. n is the number of objects in the dataset, c is the actual number of clusters, c ER is the number of clusters estimated by the empirical rule, that is, , c AP is the number of clusters obtained by AP algorithm, and c DBA is the number of clusters obtained by density-based algorithm.

c DBA
Dataset n c c ER c AP p = 1 p = 2 p = 3 p = 5
Iris1503129201496
Wine1783131514763
Seeds21031413181252
SubKDD1050632242117107
SD1200201419382220
SD220004442516342
SD388532927241953
SD494733031231384
Experiment results show that, for the dataset with the true number of clusters greater than , it is obviously incorrect to let . In other words, the empirical rule is invalid. The number of clusters finally obtained by AP algorithm is close to the actual number of clusters. For the dataset with the true number of clusters smaller than , the number of clusters generated by AP algorithm can be used as an appropriate c max⁡, but, sometimes, the value is greater than the value estimated by the empirical rule, which enlarges the search range of the optimal number of clusters. If c max⁡ is estimated by the proposed density-based algorithm, the results in most cases are appropriate. The method is invalid only on SD1 when the cutoff distance is the first 5% distance. When the cutoff distance is selected as the first 3% distance, c max⁡ generated by the proposed algorithm is much smaller than . It is closer to the true number of clusters and greatly narrows the search range of the optimal number of clusters. Therefore, the cutoff distance is selected as the distance of the first 3% in the later experiments.

5.2. Experiment of Influence of the Initial Cluster Centroids on the Convergence of FCM

To show that the initial cluster centroids obtained by density-based algorithm can quicken the convergence of clustering algorithm, the traditional FCM is adopted for verification. The number of clusters is assigned as the true value, with the convergence threshold being 10−5. Because the random selection of initial cluster centroids has a large influence on FCM, the experiment of each dataset is done repeatedly for 50 times, and the round numbers of the mean of algorithmic iterations are compared. The specific experimental result is shown in Table 3.
Table 3

Comparison of iterations of FCM algorithm. Method 1 uses the random initial cluster centroids, and Method 2 uses the cluster centroids obtained by density-based algorithm.

DatasetMethod 1Method 2
Iris2116
Wine2718
Seeds1916
SubKDD3123
SD13814
SD21812
SD33022
SD42621
As shown in Table 3, the initial cluster centroids obtained by density-based algorithm can effectively reduce the iteration of FCM. Particularly, on SD1, the iteration of FCM is far smaller than that with randomly selected initial cluster centroids whose minimum iteration is 24 and maximum iteration reaches 63, with unstable clustering results. Therefore, the proposed method can not only effectively quicken the convergence of FCM, but also obtain a stable clustering result.

5.3. Experiment of Clustering Accuracy Based on Clustering Validity Index

For 3 UCI datasets and SubKDD, clustering accuracy r is adopted to measure the clustering effect, defined as Here b is the number of objects which co-occur in the ith cluster and the ith real cluster, and n is the number of objects in the dataset. According to this measurement, the higher the clustering accuracy is, the better the clustering result of FCM is. When r = 1, the clustering result of FCM is totally accurate. In the experiments, the true number of clusters is assigned to each dataset and the initial cluster centroids are obtained by density-based algorithm. Then, experimental results of clustering accuracy are shown in Table 4.
Table 4

Clustering accuracy.

Dataset IrisWineSeedsSubKDD
Clustering accuracy84.00%96.63%91.90%94.35%
Apparently, clustering accuracies of the datasets Wine, Seeds, and SubKDD are high, while that of the dataset Iris is relatively low for the reason that two clusters are nonlinearly separable. The clustering results of the proposed clustering validity index on 4 synthetic datasets are shown in Figure 4, which depict the good partitions on the datasets SD1 and SD2 and the slightly poor partition on SD3 and SD4 because of the complexity of their structure or nonconvexity. When the number of clusters is set as 4, SD3 is divided into 4 groups that means the right and the left each have 2 groups.
Figure 4

Clustering results of two synthetic datasets.

5.4. Experiment of the Optimal Number of Clusters

At last, Xie-Beni index, Bensaid index, Kwon index, and the proposed index are, respectively, adopted for running of SAFCM so as to determine the optimal number of clusters. The results are shown in Table 5. V XB, V B, V K, and V R represent Xie-Beni index, Bensaid index, Kwon index, and the proposed index, respectively.
Table 5

Optimal number of clusters estimated by several clustering validity indices.

Dataset V XB V B V K V R
Iris2922
Wine3633
Seeds2522
SubKDD101094
SD120202020
SD24444
SD35354
SD42823
It shows that, for synthetic datasets SD1 and SD2 with simple structure, these indices can all obtain the optimal number of clusters. For 3 UCI datasets, SubKDD and SD4, Bensaid index cannot obtain an accurate number of clusters except SD3. For the dataset Wine, both Xie-Beni index and Kwon index can obtain the accurate number of clusters, while for the datasets Iris, Seeds, SubKDD, SD3, and SD4, they only obtain the result approximate to the true number of cluster. The proposed index obtains the right result on the datasets Wine and SD4 and a better result compared to Xie-Beni index and Kwon index on the datasets SubKDD and SD3, while it obtains the same result of the two indices on the datasets Iris and Seeds. There are two reasons. First, the importance degree of each property of the dataset is not considered in the clustering process but assumed to be the same, thereby affecting the experimental result. Second, SAFCM has a weak capacity to deal with overlapping clusters or nonconvexity dataset. These are the authors' subsequent research contents. Tables 6 –13 list the detailed results of 4 indices on the experimental datasets in each iteration of SAFCM. For the dataset Wine, when the number of clusters is 4, 5, or 6, the result based on Xie-Beni index or Kwon index is of no convergence. However, the proposed index can achieve the desired results.
Table 6

The value of clustering validity index on Iris.

c V XB V B V K V R
2 0.114234 0.223604 17.384973 0.473604
30.2239730.12459834.5721460.551565
40.3167420.09910349.2794880.615436
50.5601090.08910887.5409680.676350
60.5744750.07220190.5633790.691340
70.4000710.06700563.3286790.735311
80.2756820.03628345.7369720.614055
90.250971 0.027735 42.8684490.584244
Table 7

The value of clustering validity index on Wine.

c V XB V B V K V R
20.6634061.328291118.339021.578291
3 0.468902 0.513071 83.939058 1.346421
40.4732541.735791
50.3736681.846686
6 0.256698 1.683222
Table 8

The value of clustering validity index on Seeds.

c V XB V B V K V R
2 0.147247 0.293609 31.170349 0.543609
30.2121270.15089945.3262160.599001
40.2434830.12772052.2153340.697943
50.348842 0.085477 75.4936540.701153
Table 9

The value of clustering validity index on SubKDD.

c V XB V B V K V R
20.6469891.324434550.1666761.574431
30.2607550.378775222.0208381.090289
40.1338430.062126119.544560 0.511238
50.2344020.052499202.6412040.537852
60.1807280.054938156.8122710.583800
70.1346360.047514119.0292650.619720
80.1045110.03284991.98527400.690873
90.1297210.027639 74.0128470 0.562636
10 0.06822 0.027025 91.35285600.528528
Table 10

The value of clustering validity index on SD1.

c V XB V B V K V R
20.2216930.44339044.5929680.693390
30.2060350.19885340.2642510.726245
40.1277310.09365326.2202000.655550
50.1307810.06984827.1548670.651465
60.1448940.05006722.9221210.639325
70.1365620.04027529.1261520.636258
80.1124800.03287424.3236250.627442
90.1150900.02683324.2425800.624936
100.1414150.02261128.5745790.616701
110.1266800.01925628.8217070.611524
120.1031780.01663423.9318650.605990
130.1103550.01325326.5170650.588246
140.0955130.01108323.6350220.576808
150.0759280.00981719.3020950.562289
160.0660250.00882417.2361380.557990
170.0543140.00724814.9952840.544341
180.0453980.00609013.2088100.534882
190.0394920.00536511.9774370.527131
20 0.034598 0.004917 10.835952 0.517481
Table 11

The value of clustering validity index on SD2.

c V XB V B V K V R
20.0662860.132572132.815030.382572
30.0682420.063751137.525350.394200
4 0.060916 0.028149 123.09967 0.378337
Table 12

The value of clustering validity index on SD3.

c V XB V B V K V R
20.1483790.300899131.5572690.570876
30.195663 0.142149 173.9005510.599680
40.1275120.142150113.748947 0.557198
5 0.126561 0.738070 113.262975 0.589535
Table 13

The value of clustering validity index on SD4.

c V XB V B V K V R
2 0.104434 0.208832 99.1482710 0.473748
30.1703260.142561162.044450 0.458832
40.2218840.081007211.6926990.583529
50.1562530.053094157.6839210.603211
60.1231910.041799118.2791160.575396
70.1654650.032411107.2100820.592625
80.145164 0.028698 139.3109690.606049

6. Conclusions

FCM is widely used in lots of fields. But it needs to preset the number of clusters and is greatly influenced by the initial cluster centroids. This paper studies a self-adaptive method for determining the number of clusters by using of FCM algorithm. In this method, a density-based algorithm is put forward at first, which can estimate the maximum number of clusters to reduce the search range of the optimal number, especially being fit for the dataset on which the empirical rule is inoperative. Besides, it can generate the high-quality initial cluster centroids so that the the clustering result is stable and the convergence of FCM is quick. Then, a new fuzzy clustering validity index was put forward based on fuzzy compactness and separation so that the clustering result is closer to global optimum. The index is robust and interpretable when the number of clusters tends to that of objects in the dataset. Finally, a self-adaptive FCM algorithm is proposed to determine the optimal number of clusters run in the iterative trial-and-error process. The contributions are validated by experimental results. However, in most cases, each property plays a different role in the clustering process. In other words, the weight of properties is not the same. This issue will be focused on in the authors' future work.
  5 in total

1.  A similarity-based robust clustering method.

Authors:  Miin-Shen Yang; Kuo-Lung Wu
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  2004-04       Impact factor: 6.226

2.  Clustering by passing messages between data points.

Authors:  Brendan J Frey; Delbert Dueck
Journal:  Science       Date:  2007-01-11       Impact factor: 47.728

3.  Mercer kernel-based clustering in feature space.

Authors:  M Girolami
Journal:  IEEE Trans Neural Netw       Date:  2002

4.  Machine learning. Clustering by fast search and find of density peaks.

Authors:  Alex Rodriguez; Alessandro Laio
Journal:  Science       Date:  2014-06-27       Impact factor: 47.728

5.  A new validity measure for a correlation-based fuzzy c-means clustering algorithm.

Authors:  Mingrui Zhang; Wei Zhang; Hugues Sicotte; Ping Yang
Journal:  Conf Proc IEEE Eng Med Biol Soc       Date:  2009
  5 in total
  3 in total

1.  Automated detection of photoreceptor disruption in mild diabetic retinopathy on volumetric optical coherence tomography.

Authors:  Zhuo Wang; Acner Camino; Miao Zhang; Jie Wang; Thomas S Hwang; David J Wilson; David Huang; Dengwang Li; Yali Jia
Journal:  Biomed Opt Express       Date:  2017-11-07       Impact factor: 3.732

2.  Magnetic Resonance Features of Acquired Immune Deficiency Syndrome Involving Central Nervous System Diseases by Intelligent Fuzzy C-Means Clustering (FCM) Algorithm.

Authors:  Gang Huang; Jiaqi Chen; Yuli Ge; Xiaomei Zhu; Meixiao Ding; Xugao Chen; Chunsheng Qu
Journal:  Comput Math Methods Med       Date:  2022-07-05       Impact factor: 2.809

3.  Clustering of fMRI data: the elusive optimal number of clusters.

Authors:  Mohamed L Seghier
Journal:  PeerJ       Date:  2018-10-03       Impact factor: 2.984

  3 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.