Literature DB >> 33265364

KL Divergence-Based Fuzzy Cluster Ensemble for Image Segmentation.

Huiqin Wei1, Long Chen1, Li Guo1.   

Abstract

Ensemble clustering combines different basic partitions of a dataset into a more stable and robust one. Thus, cluster ensemble plays a significant role in applications like image segmentation. However, existing ensemble methods have a few demerits, including the lack of diversity of basic partitions and the low accuracy caused by data noise. In this paper, to get over these difficulties, we propose an efficient fuzzy cluster ensemble method based on Kullback-Leibler divergence or simply, the KL divergence. The data are first classified with distinct fuzzy clustering methods. Then, the soft clustering results are aggregated by a fuzzy KL divergence-based objective function. Moreover, for image segmentation problems, we utilize the local spatial information in the cluster ensemble algorithm to suppress the effect of noise. Experiment results reveal that the proposed methods outperform many other methods in synthetic and real image-segmentation problems.

Entities:  

Keywords:  KL divergence; ensemble learning; fuzzy clustering; image segmentation; spatial information

Year:  2018        PMID: 33265364      PMCID: PMC7512790          DOI: 10.3390/e20040273

Source DB:  PubMed          Journal:  Entropy (Basel)        ISSN: 1099-4300            Impact factor:   2.524


1. Introduction

Image segmentation has become increasingly important in a wide variety of applications like biomedical image analysis [1,2,3,4] and intelligent robotics [5]. Unfortunately, due to the variations and noise of images, finding a good image partition is still a great challenge, especially when we want to stress the semantic meanings of different regions. There is a growing number of methods available for image-segmentation problems over recent years [1,2,3,4,5,6,7,8]. Fuzzy approaches show considerable advantages among these methods by carefully handling the ubiquitous uncertainty and unknown noise in images. In contrast to hard segmentation methods, the fuzzy ones could retain much more information from the original data [9,10,11]. The fuzzy c-means (FCM) clustering algorithm is the best known one in fuzzy segmentation methods [12]. FCM derives the segmentation by iteratively minimizing a cost function that is dependent on the distances of image pixels to cluster centers in the feature domain. However, the standard FCM does not consider any spatial information in the image context, and hence suffers from high sensitivity to noise. Many extensions of the standard FCM have been proposed to suppress the effects of noise in images [9,10,13,14]. For example, the spatial neighborhood information is incorporated into the membership function of FCM for clustering in [9,10]. With so many available algorithms, one may obtain very different clustering results for a given dataset. Without ground truth, it is difficult to select the most suitable method for a given problem. In addition, most of existing algorithms require the specification of some parameters to obtain a decent grouping of the data. Instead of choosing a decent setting of parameters, a particular algorithm, a good clustering configuration, or a special similarity measure that best suits a given problem, ensemble clustering can integrate results from multiple weak partition algorithms into a single robust and stable solution. The inputs of ensemble clustering are a set of data partitions. Previously, ensemble approaches for clustering problems have been studied extensively [15,16,17,18,19,20,21,22,23]. These ensemble methods generate many partitions by using the same method with various parameter settings, distinct methods, different inputs, or divergent feature sets. The final merged partition is obtained by approaches like majority voting or evidence accumulation [20]. For the fuzzy c-means and its extensions, their outputs are soft partitions of data. A soft partition assigns a degree of association of each instance to every cluster. So instead of a label vector for all the instances, in soft partition we have a matrix of memberships in which each instance has a membership vector that represents its belongingness to all clusters. Many ensemble methods have been proposed for the soft partitions [16,17,21], in which the most straightforward approach is conducting the fuzzy c-means over the membership matrices of different soft partitions. However, FCM uses squared Euclidean distance to measure the similarity of a membership vector to a cluster center. This is inappropriate for the situation when one data’s memberships to all clusters usually sum to one [21]. A better solution is to regard the membership vector as some discrete probability function and use the statistical distance like KL divergence as the similarity measure [24]. In this paper, we first propose an efficient fuzzy cluster ensemble method based on KL divergence . This algorithm is similar to the fuzzy c-means, differing only in the fact that it uses the KL divergence to handle the memberships like discrete probabilities. Theoretically, we have developed an optimization algorithm for the proposed . For image-segmentation problems, because it is well known that the comparative performance of different clustering methods can vary significantly across datasets, we first utilize heterogeneous center-based soft clustering algorithms to categorize the pixels in the image. The soft clustering results obtained by different methods guarantee the diversity of partitions. Then, the fusion of soft partitions is provided by applying . Although basically outperforms individual clustering methods, it still classifies noisy pixels in wrong segments sometimes. So, we further use the local spatial information in the calculation of membership values for to enhance the accuracy of image segmentation and propose the spatial (). Experimental results on synthetic and real image segmentation demonstrate that the proposed methods perform better than some widely used fuzzy clustering-based approaches. The remainder of this paper is organized as follows. In addition to several standard fuzzy clustering methods, Section 2 presents related ensemble clustering methodology that includes the ensemble clustering generator and consensus function. In Section 3, we propose the fuzzy cluster ensemble method based on KL divergence . For the image-segmentation problems, we utilize the local spatial information of the image to handle the membership values for the proposed clustering ensemble algorithm and propose . In Section 4, the numerical experiments demonstrate the good performance of the proposed algorithm for image segmentation. Section 5 gives the discussion of our methods. At last, the conclusion and future work are given in Section 6.

2. Related Work

2.1. Fuzzy C-Means

Fuzzy c-means (FCM) [12] divides a set of n datapoints into c clusters by minimizing the weighted summation of distances from the datapoints to the cluster centers: where denotes the Euclidean distance, is the fuzzification coefficient, which usually takes the value of 2, is the membership value of data to the i-th cluster center and , . The FCM algorithm iteratively updates and as follows: and until the stop criterion like maximum number of iterations is satisfied.

2.2. Local Spatial Fuzzy C-Means

To deal with noise and segment images better, one extension to FCM is incorporating the local spatial information into the standard FCM. In [25], the membership is updated by the weighted average value of its neighbors’ membership values, which exploits the spatial information as the following: where denotes a local square window centered on pixel in the spatial domain, and is the size of the neighborhood. On the one hand, clustering algorithms, such as FCM and its extensions, are effective for image segmentation. They could show good performance in some problems. On the other hand, cluster ensembles based on different clustering methods are more robust and stable. So, the proper combination of different fuzzy clustering algorithms could produce more reliable and accurate results.

2.3. Cluster Ensemble

Basically, there are two parts in the cluster ensembles. One is the ensemble clustering generator, and another is the consensus function. The first part concentrates on producing more diverse clustering results, while the second part concentrates on finding a good consensus function to improve the accuracy of the results. For the first part of cluster ensembles, Rathore et al. [17] proposed that multiple partitions can be obtained using the fuzzy c-means clustering algorithm on a randomly projected dataset. According to [18], Fred and Jain focused on running k-means several times by using different initializations to obtain various partitions. In [16], Zou et al. utilized different clustering algorithms to generate base partitions. For the second part of cluster ensembles, many methods have been proposed to fuse multiple partitions into a consensus one. The ensemble clustering method in [26] constructs the co-association matrix obtained by reliable data pairs in the same cluster among multiple partitions, and then applies the spectral clustering on the completed co-association matrix to obtain the best partition. In [27], the authors introduced an ensemble approach for categorical data by finding the best partition that minimizes an objective function. Similarly, ensemble clustering in [28] was cast into a problem that selects a consensus partition by maximizing the within-cluster similarity. Recently, Wu et al. proposed the utility function based on the fuzzified contingency matrix to measure the similarity multiple partitions. They established a framework for fuzzy consensus clustering by vertical and horizontal segmentation schemes to deal with big data clustering [23]. To solve the ensemble clustering problem, we can work on two basic operating units: the data objects and the basic clusters. Many effective algorithms have been proposed to handle the ensemble clustering problem at the object level [16,18,29] and cluster level [21,24]. In [24], each instance in a soft ensemble is represented by the concatenation of membership probability distributions. Then, a distance measure between two instances was defined using the Kullback–Leibler (KL) divergence. As the ensemble size becomes larger and the total number of clusters increases, the distance measure between clusters will be a computational burden. In contrast to cluster level, the data objects are used as basic operating units in this paper. We measure the similarly of a membership vector to a cluster center using the Kullback–Leibler (KL) divergence.

3. KL Divergence-Based Fuzzy Cluster Ensemble

In this section, we illustrate the problem of combining multiple clustering operations and propose an efficient fuzzy cluster ensemble method based on KL divergence. Then, when solving image-segmentation problems, we utilize the local spatial information of the image to handle the membership values for the proposed ensemble approach.

3.1. Formulation of the Fuzzy Cluster Ensemble Problem

In this paper, to build the ensemble, we first apply some heterogeneous center-based soft clustering algorithms to generate membership matrices as basic partitions , where r is the number of soft clustering algorithms, and is the transposed membership matrix obtained by the f-th clustering method. In the membership matrix, each entry denotes the degree of data belonging to a cluster. Here, the number of basic partitions is the same as the one of clustering algorithms, and any clustering algorithm that generates membership matrices can be applied. These partition matrices are then concatenated to . Here, n is the number of the data and s is the number of memberships derived for data from different algorithms. If each partition has c clusters, we have for brevity. However, this is not a necessity, and we can have different number of clusters in different partitions. Let for the normalization of matrix and we define as the entry of , where . Let , we have , , which implies each row of as a probability vector. is the input data of the algorithm. As an illustrated example, Table 1 demonstrates two soft basic partitions derived from two algorithms, and their concatenation and normalization . From this example, it is easy to know that we do not arrange lexicographically the different soft cluster solutions. In the proposed ensemble method, we simply concatenate and normalize all membership values of different fuzzy clustering methods and take them as the new representation of the data. The new representation sums to one and can be regarded as a discrete distribution. Therefore, the entropy-based KL divergence is a better measurement of a discrete distribution.
Table 1

Concatenation and normalization .

U1 U2 Ucon U^con
x1 0.70.20.1 0.10.70.2 0.70.20.10.10.70.2 0.350.100.050.050.350.10
x2 0.90.10.0 0.00.80.2 0.90.10.00.00.80.2 0.450.050.000.000.400.10
x3 0.20.60.2 0.10.10.8 0.20.60.20.10.10.8 0.100.300.100.050.050.40
x4 0.10.90.0 0.20.10.7 0.10.90.00.20.10.7 0.050.450.000.100.050.35
x5 0.10.20.7 0.60.20.2 0.10.20.70.60.20.2 0.050.100.350.300.100.10
We aggregate the soft clustering results by a fuzzy KL divergence-based objective function. For a new set of n probability vectors, is applied to divide them into a desired number of clusters again. Specifically, use KL divergence to measure the distance of a membership vector to a cluster center. Moreover, for the image-segmentation problems, we utilize the local spatial information for calculating the membership value of the proposed and developed . In the long run, we can get the ensemble clustering results by and . Figure 1 illustrates the framework of the two methods we proposed.
Figure 1

The framework of and .

3.2. Fuzzy Cluster Ensemble Based on KL Divergence ()

The KL divergence is used to measure variation in the discrete probability distributions of attributes P and Q. It is defined by Let . The divides a set of n probability vectors with s dimension into c clusters by minimizing the following objective function, subject to where presents the membership of the k-th probability vector in the i-th cluster , and denotes . The vectors close to the centroid of their clusters based on KL divergence are assigned high membership values, and low membership values are assigned to datapoints far from the centroid to minimize the objective function. Next, we take an iterative way for solving the modified membership and cluster centers here. Let the Lagrangian of formula (6) be Letting the first derivatives of with respect to u and o equal to zero yields the two necessary conditions for minimizing . Thus, we obtain respectively. Using (10), By formula (8), we have Hence, On the other hand, from (11), Therefore, by formula (7), it holds that Therefore, the algorithm iteratively updates (15) and (19) until the stop criterion, like the convergence of the objective function or the satisfaction of maximum iteration number.

3.3. Spatial Information-Based

The disadvantage of for image segmentation is the discarding of spatial information in images. Since neighboring pixels are highly correlated, to suppress the noise and acquire better segments, we include this property into the algorithm. The membership of k-th pixel in the i-th cluster is summed over the weighted average value of the membership values of the user-specified neighbors as following [25]: where denotes a local square window centered on pixel in the spatial domain, and is the size of the neighborhood. The modified memberships can be used to get the cluster center by Therefore, our spatial information-enhanced algorithm for image segmentation is presented in Algorithm 1. In the algorithm, the memberships and cluster centers are updated iteratively according to Equations (19)–(21). Make initialization of cluster centers, , ; ; repeat Calculate the membership by Equation (19); Utilize the spatial information in membership function of and calculate new membership with Equation (20); Update cluster centers by Equation (21); Calculate the objective function by (6); ; until or

4. Experiment Results

In this section, we compare the newly proposed , with SFCM [25], SSCM [25], NLSFCM [16] and NLSSCM [16] on several synthetic and real images. For our clustering ensemble approaches, to generate basic partitions, we directly use SFCM, SSCM, NLSFCM and NLSSCM methods. Since r is the number of above-mentioned heterogeneous center-based clustering algorithms, here we have . More specifically, the and incorporate membership matrices derived by SFCM, SSCM, NLSFCM and NLSSCM to get their soft partitions. The fuzzification coefficient m is set to 2 for all the algorithms in our experiments. The size of square window of is set to . More settings and their justifications are provided in the following discussion section. When the algorithms are tested on images with ground truth or reference partitions, the segmentation accuracy (SA) is calculated as The SA of algorithm i on class j is measured as where denotes the set of pixels belonging to class j that are found by algorithm i, and denotes the set of pixels in class j which is in the reference segmented image. It should be noted that when we have a soft partition based on a fuzzy approach, we need a defuzzification method to assign each pixel to a segment. After we obtain the membership matrix , we calculate the to obtain the final results. In other words, we finally classify one pixel or data into the category where it takes the largest membership value.

4.1. Synthetic Images

We use a synthetic two-value image, the image ‘Trin’, and the synthetic magnetic resonance (MR) images as testing images first. The synthetic two-cluster image shown in Figure 2b, with values 1 and 0, is similar to the image used in [14]. Its size is set to pixels. The synthetic image of ‘Trin’ contains four regions and its size is set to pixels. Both synthetic images add Gaussian noise and Rician noise. Figure 2 and Figure 3a present synthetic two-value images with 50% Gaussian noise and 50% Rician noise. Figure 4 and Figure 5a show ‘Trin’ with 15% Gaussian noise and 12% Rician noise, respectively. For two sorts of noised images, Table 2 and Table 3 show the segmentation accuracies (SAs) of six methods. We obtain these results by running all methods on the same image (no matter the size or noise) for fair comparison.
Figure 2

50% Gaussian-noised synthetic two-value image: (a) original; (b) ground-truth image and after segmentation into two regions with (c) SFCM; (d) SSCM; (e) NLSFCM; (f) NLSSCM; (g) ; (h) .

Figure 3

50% Rician-noised synthetic two-value image: (a) original; (b) ground-truth image and after segmentation into two regions with (c) SFCM; (d) SSCM; (e) NLSFCM; (f) NLSSCM; (g) ; (h) .

Figure 4

15% Gaussian-noised synthetic images of ‘Trin’: (a) original; (b) ground-truth image and after segmentation into four regions with (c) SFCM; (d) SSCM; (e) NLSFCM; (f) NLSSCM; (g) ; (h) .

Figure 5

12% Rician-noised synthetic images of ‘Trin’: (a) original; (b) ground-truth image and after segmentation into four regions with (c) SFCM; (d) SSCM; (e) NLSFCM; (f) NLSSCM; (g) ; (h) .

Table 2

Segmentation accuracies of different methods on noised two-value images.

SASFCMSSCMNLSFCMNLSSCM FCE_KL FCE_sKL
1% Gaussian 1.0000 1.0000 1.0000 1.0000 1.0000 1.0000
3% Gaussian 1.0000 1.0000 1.0000 1.0000 1.0000 1.0000
5% Gaussian 1.0000 1.0000 1.0000 1.0000 1.0000 1.0000
10% Gaussian0.99920.99920.99960.99960.9996 1.0000
15% Gaussian0.99840.99800.99920.99920.9992 1.0000
20% Gaussian0.99280.99320.99880.99880.9988 1.0000
30% Gaussian0.97320.97600.99760.99760.9956 1.0000
50% Gaussian0.91000.91480.99160.99160.9808 1.0000
1% Rician 1.0000 1.0000 1.0000 1.0000 1.0000 1.0000
3% Rician 1.0000 1.0000 1.0000 1.0000 1.0000 1.0000
5% Rician 1.0000 1.0000 1.0000 1.0000 1.0000 1.0000
10% Rician 1.0000 1.0000 1.0000 1.0000 1.0000 1.0000
15% Rician 1.0000 1.0000 1.0000 1.0000 1.0000 1.0000
20% Rician 1.0000 1.0000 1.0000 1.0000 1.0000 1.0000
30% Rician0.99560.99600.99880.99880.9988 0.9996
50% Rician0.86400.86640.97440.98280.9678 0.9968
Table 3

Segmentation accuracies of different methods on noised images of ‘Trin’.

SASFCMSSCMNLSFCMNLSSCM FCE_KL FCE_sKL
10% Gaussian0.98170.98780.99460.99510.9966 0.9968
12% Gaussian0.95630.97120.98930.98170.9922 0.9951
15% Gaussian0.89790.93480.98000.96700.9797 0.9907
18% Gaussian0.84200.88960.96240.91380.9631 0.9836
20% Gaussian0.82180.84570.92900.89370.9468 0.9785
25% Gaussian0.77290.77730.78470.78130.9026 0.9558
30% Gaussian0.68600.74460.73410.75390.7239 0.8569
10% Rician0.98390.99170.99190.99340.9954 0.9973
12% Rician0.87820.97680.98170.98880.9897 0.9951
15% Rician0.75780.71360.92090.96700.9758 0.9905
18% Rician0.71140.71120.78860.95260.9570 0.9832
20% Rician0.67990.70850.75980.84130.9377 0.9729
25% Rician0.66500.67750.76810.75310.8472 0.9263
30% Rician0.62870.64060.69140.68180.7832 0.8472
For the synthetic two-value image, Table 2 shows that the performance of and are as good as that of SFCM, SSCM, NLSFCM and NLSSCM, which means that all of them can easily remove the low-level noise from this image. When Gaussian noise reaches 30% and 50%, can segment the image better than SFCM and SSCM but worse than NLSFCM and NLSSCM. can eliminate the noise better than the other five methods for the high-level noised images, and this demonstrates its robustness. The fuzzy clustering methods, SFCM and SSCM, adjust the memberships with the local information of image. NLSFCM and NLSSCM update the membership values by the nonlocal information. Then, the four methods are all affected by the setting of neighborhood size and weight values when using the local or nonlocal information. In addition, NLSSCM and NLSFCM are affected much more by the initial values in the iteration than SFCM, SSCM and the proposed methods. Then, the simple setting of SFCM, SSCM and their robustness to the initialization may be the reason that some results of SFCM and SSCM in our experiments are the same as the ones in [25] coincidentally. On the other hand, our experiments show different results of NLSFCM and NLSSCM when compared to [16], for we may choose a different size of nonlocal window, a different initialization, or a different computation of the nonlocal weights which are affected much by the similar measurement of patches in the nonlocal window. Some of our results of NLSSCM and NLSFCM are better than the ones in [16], and some are not. However, the difference is not great. So, the superiority of the proposed approach demonstrated in Table 2 is still trustworthy. For the synthetic image “Trin”, Table 3 demonstrates that individual clustering methods (SFCM, SSCM, NLSFCM and NLSSCM) deteriorate dramatically with the increasing of noise rate. However, and can stop this trend and improve the performance. behaves better than individual clustering methods handling this image with Gaussian noise except 15% and 30%. removes all Rician noise better than SFCM, SSCM, NLSFCM and NLSSCM. It is obvious that can handle this image with heavy Gaussian and Rician noise more easily than the other five methods. In order to visually compare the performance, the segmentation results of all of the six methods on the noised two-value image and the image “Trin” are shown in Figure 2, Figure 3, Figure 4 and Figure 5. All of these figures illustrate that behaves better than the other five methods. The synthetic MR images of the human brain and their reference segmentations are provided by [30]. They are T1-weighted MR phantom with slice thickness of 1 mm, having various levels of Rician noise, without intensity inhomogeneity. The Rician noise rates range from 25% to 50% added to the synthetic MR images of the human brain. Since ground truth of the synthetic MR image is available, Table 4 shows the SAs of the six methods. Both and segment these images better than SFCM, SSCM, NLSFCM and NLSSCM. Furthermore, according to the SAs, the superiority of can be verified easily. Figure 6a presents the synthetic synthetic MR images with 32% Rician noise. We add the zoom of the image portion highlighted with a red rectangle. Segmentation results in enlarged red rectangles reveal that and can remove noise better and obtain smoother regions than the other four methods. behaves better than because the latter still produces several misclassified pixels.
Table 4

Segmentation accuracies of different methods on noised synthetic MR images.

SASFCMSSCMNLSFCMNLSSCM FCE_KL FCE_sKL
25% Rician0.85230.81250.84380.95430.9692 0.9700
27% Rician0.84730.81170.84110.85040.9669 0.9670
30% Rician0.84050.80700.83910.84110.9605 0.9632
32% Rician0.83730.80260.83930.83390.9559 0.9608
35% Rician0.83210.79660.83900.83170.9485 0.9566
40% Rician0.81600.79030.83950.82490.9361 0.9493
50% Rician0.78040.77840.86100.79110.9074 0.9298
Figure 6

32% Rician-noised synthetic MR images: (a) original; (b) ground-truth image and after segmentation into four regions with (c) SFCM; (d) SSCM; (e) NLSFCM; (f) NLSSCM; (g) ; (h) .

4.2. Real Images

The first real image is the MR brain image obtained from the Internet Brain Segmentation Repository (IBSR) database [31]. This image should be partitioned into three regions corresponding to cerebrospinal fluid (CSF), white matter (WM) and gray matter (GM). As mentioned in [32], one general method of MR brain image segmentation involves two parts: the classification and the identification of all of the voxels belonging to a specific structure. Since the CSF in the center of the brain is a continuous volume, it can be well segmented by some contour-mapping algorithm. Therefore, this paper focuses on the segmenting of WM and GM. Figure 7 illustrates one sample without noise. According to the ground truth of the IBSR image, the SAs of different methods on the images having 12%, 15% and 18% Rician noise are shown in the Table 5, Table 6 and Table 7, in which stands for SA for the cluster of WM, and is for the cluster of GM. The Table 5, Table 6 and Table 7 demonstrate excellent performance of in terms of various levels of noise.
Figure 7

(a) Original IBSR image; (b) WM; (c) GM.

Table 5

Segmentation accuracies of different methods on noised images of ‘IBSR12%’.

12% Rician NoiseSFCMSSCMNLSFCMNLSSCM FCE_KL FCE_sKL
SA0.72930.74000.71810.71630.7247 0.7526
SA1 0.74630.75300.72970.72470.7381 0.7663
SA2 0.70990.72560.70550.70730.7098 0.7372
Table 6

Segmentation accuracies of different methods on noised images of ‘IBSR15%’.

15% Rician NoiseSFCMSSCMNLSFCMNLSSCM FCE_KL FCE_sKL
SA0.70790.72370.70230.70930.7098 0.7442
SA1 0.72810.74310.72480.74210.7369 0.7679
SA2 0.68440.70120.67580.66700.6763 0.7150
Table 7

Segmentation accuracies of different methods on noised images of ‘IBSR18%’.

18% Rician NoiseSFCMSSCMNLSFCMNLSSCM FCE_KL FCE_sKL
SA0.67440.70140.68790.70280.6953 0.7307
SA1 0.70310.72680.72050.75840.7338 0.7653
SA2 0.63950.67080.64670.61390.6438 0.6841
The second real image is the positron emission tomography (PET) lung image of a dog. It is demonstrated in Figure 8a, with pixels. The reference segmentations of these images are not available. However, the segmentation results with six methods are illustrated in Figure 8b–g. Visually, the results obtained by SFCM and SSCM contain many flakes, while the results using NLSFCM and NLSSCM are more robust, such that some details of the lung are ignored. The Figure 8f,g reveals that and outperform the other four methods. behaves better than , for still produces several misclassified pixels.
Figure 8

Segmentation result of different methods on a PET image of the lung of dog: (a) original and after segmentation into three regions with (b) SFCM; (c) SSCM; (d) NLSFCM; (e) NLSSCM; (f) ; (g) .

Another medical image is shown in Figure 9b, which is a image of a healthy bone. We add 10% Gaussian noise to this image. From Figure 9, it is obvious that the performance of is better than the other five methods. can remove noise in the image and obtain more homogeneous regions.
Figure 9

10% Gaussian-noised bone images: (a) original; (b) noise-free image and after segmentation into two regions with (c) SFCM; (d) SSCM; (e) NLSFCM; (f) NLSSCM; (g) ; (h) .

The last real image is demonstrated in Figure 10a, which is a image of horses. From Figure 10, SFCM and NLSFCM contain lots of misclassified pixels. In contrast, the segmentation results obtained by and have less flakes in the lower-left corner than the other four methods. and can well segment the horses from the background.
Figure 10

Segmentation result of different methods on an image of a horse: (a) original and after segmentation into three regions with (b) SFCM; (c) SSCM; (d) NLSFCM; (e) NLSSCM; (f) ; (g) .

5. Discussion

In the previous section, compared with different clustering methods like SFCM, SSCM, NLSFCM and NLSSCM, both and have shown superior performance in segmentation of low-level noised synthetic and real images. The possible explanation is that for image-segmentation problems, the results obtained by SFCM and SSCM contain many flakes, while the results using NLSFCM and NLSSCM are much more robust such that some details of the images are ignored easily. The ensemble methods like and can avoid the weakness of a single clustering method. Experiment results also demonstrate that can eliminate the high-level noise better than all the other five methods in synthetic and real image segmentation. This is the result of ’s inclusion of spatial information for the purpose of noise suppression. For the parameters of different clustering algorithms used by the ensemble, various settings can be chosen. However, in this paper we applied some ad-hoc selections. For example, the fuzzy coefficient m has huge influence on fuzzy clustering, notably in FCM. In contrast to k-means clustering, in fuzzy c-means, the datapoint is not directly assigned to any cluster, but its fuzzy membership values of all clusters are given as the final result. When m –> 1, FCM is similar to the k-means algorithm. The larger m, the more clusters share their objects, and vice versa. The fuzzification coefficient m was set to 2 in many studies. We also set in this paper. This decision should not discourage the future study on applying different fuzzy coefficients in fuzzy cluster ensembles, and more discussion on the determining of parameters for fuzzy c-means cluster analysis can be found in [33,34]. Similarly, we do not carefully select the parameters used in noise suppression of clustering algorithms. For example, the size of the local square window of is set to in our experiments. If we choose a larger size of square window to include more local information and ignore the computing cost, the performance of may be better. In unsupervised learning, the goodness-of-fit measure related to the number of cluster centers can be obtained by many methods [35,36]. Because it is not the focus of our research, we set it as the prior knowledge in this paper. For image-segmentation problems, the number of segmentation regions is usually expected to be known or given by the image directly in the experiments [14]. Moreover, in practical applications, it is not mandatory to get the same number of clusters across all different clustering methods in the ensemble, especially in the process of obtaining the basic partitions by different clustering methods for our proposed method. That is to say, different numbers of cluster centers can be used by these different clustering methods to get the basic partitions. In Figure 1, the obtained membership matrices, , can be of different sizes in rows. However, in the experiments, we use the ground truth and the measurement SA to do fair comparisons. Because the number of clusters has great influence on the values of SA, the number of clusters must be set the same across all different clustering methods when dealing with the specified picture. Furthermore, we want to show the proposed ensemble methods outperform the other clustering methods in the case of the exact same parameter settings. So, the same number of clusters are used in all methods of the experiments.

6. Conclusions

In this paper, our main contribution is to propose an efficient fuzzy cluster ensemble method based on KL divergence, , considering that soft partitions are more suitable to be measured by KL divergence. In order to obtain ensemble partitions as diverse as possible, the data are classified using distinct clustering methods instead of a single method with different parameters. Theoretically, we have developed an optimization algorithm for the proposed . For image-segmentation problems, we further use spatial information in our method and propose . According to experimental results, the proposed methods perform better than many existing clustering methods in synthetic and real image-segmentation problems. In future work, we would explore more robust distance calculations instead of KL divergence in ensemble clustering. Currently, we handle each base clustering equally, which could overlook the different reliabilities of base clusterings. So, we will also explore certain methods to calculate weights of base clusterings for further consideration.
  10 in total

1.  An adaptive spatial fuzzy clustering algorithm for 3-D MR image segmentation.

Authors:  Alan Wee-Chung Liew; Hong Yan
Journal:  IEEE Trans Med Imaging       Date:  2003-09       Impact factor: 10.048

2.  Robust image segmentation using FCM with spatial constraints based on new kernel-induced distance measure.

Authors:  Songcan Chen; Daoqiang Zhang
Journal:  IEEE Trans Syst Man Cybern B Cybern       Date:  2004-08

3.  A simple and fast method to determine the parameters for fuzzy c-means cluster analysis.

Authors:  Veit Schwämmle; Ole Nørregaard Jensen
Journal:  Bioinformatics       Date:  2010-09-29       Impact factor: 6.937

4.  Combining multiple clusterings using evidence accumulation.

Authors:  Ana L N Fred; Anil K Jain
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  2005-06       Impact factor: 6.226

5.  Fuzzy c-means clustering with spatial information for image segmentation.

Authors:  Keh-Shih Chuang; Hong-Long Tzeng; Sharon Chen; Jay Wu; Tzong-Jer Chen
Journal:  Comput Med Imaging Graph       Date:  2005-12-19       Impact factor: 4.790

6.  Clustering ensembles: models of consensus and weak partitions.

Authors:  Alexander Topchy; Anil K Jain; William Punch
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  2005-12       Impact factor: 6.226

7.  Anatomic segmentation and volumetric calculations in nuclear magnetic resonance imaging.

Authors:  D N Kennedy; P A Filipek; V R Caviness
Journal:  IEEE Trans Med Imaging       Date:  1989       Impact factor: 10.048

8.  A multiple-kernel fuzzy C-means algorithm for image segmentation.

Authors:  Long Chen; C L Philip Chen; Mingzhu Lu
Journal:  IEEE Trans Syst Man Cybern B Cybern       Date:  2011-04-05

9.  Attributed relational graphs for cell nucleus segmentation in fluorescence microscopy images.

Authors:  Salim Arslan; Tulin Ersahin; Rengul Cetin-Atalay; Cigdem Gunduz-Demir
Journal:  IEEE Trans Med Imaging       Date:  2013-03-28       Impact factor: 10.048

10.  Multiscale Combinatorial Grouping for Image Segmentation and Object Proposal Generation.

Authors:  Jordi Pont-Tuset; Pablo Arbelaez; Jonathan T Barron; Ferran Marques; Jitendra Malik
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  2016-03-02       Impact factor: 6.226

  10 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.