Literature DB >> 29240756

Elastic K-means using posterior probability.

Aihua Zheng1, Bo Jiang1, Yan Li2, Xuehan Zhang1, Chris Ding1.   

Abstract

The widely used K-means clustering is a hard clustering algorithm. Here we propose a Elastic K-means clustering model (EKM) using posterior probability with soft capability where each data point can belong to multiple clusters fractionally and show the benefit of proposed Elastic K-means. Furthermore, in many applications, besides vector attributes information, pairwise relations (graph information) are also available. Thus we integrate EKM with Normalized Cut graph clustering into a single clustering formulation. Finally, we provide several useful matrix inequalities which are useful for matrix formulations of learning models. Based on these results, we prove the correctness and the convergence of EKM algorithms. Experimental results on six benchmark datasets demonstrate the effectiveness of proposed EKM and its integrated model.

Entities:  

Mesh:

Year:  2017        PMID: 29240756      PMCID: PMC5730165          DOI: 10.1371/journal.pone.0188252

Source DB:  PubMed          Journal:  PLoS One        ISSN: 1932-6203            Impact factor:   3.240


Introduction

Data clustering, a method of unsupervised learning and a common technique for statistical data analysis has currently been widely used in machine learning [1, 2], pattern recognition [3, 4], image analysis [5, 6] and bioinformatics [7-9]. As one of the most popular clustering methods, K-means has drawn lots of attention and been widely used for data clustering. Recent work has revealed that K-means can be represented by matrix factorization formulation, and thus can deduce many kind of K-means variants [10-15]. From the point of view of determinacy, K-means clustering approaches are hard clustering where a data point exactly belongs to one particular class. However, in many cases, this is not realistic since not all the data point distinctly belongs to one single class especially for the outliers which may have equivalent similarities to every/some classes. On the other hand, above K-means inspired methods generally deal with the attribute information (feature vectors) of data. In real world applications, besides attribute information, we have various pairwise relations between data points which are expressed as graph data. There exist many clustering methods that utilize graph data, such as Ratio Cut based methods [16, 17] and Normalized Cut based methods [18, 19] and Min and Max Cut based methods [20, 21]. In essence, these clustering methods first embed graph nodes in low-dimensional space using linear embedding method PCA and nonlinear method IsoMAP [22-24], Local linear Embedding (LLE) [25-27], Local Tangent Space Alignment [26, 28, 29] etc, where feature vector information is utilized to obtain the final clustering results. However, It’s one-sided to just take into account of one single information. Inspired by recent works on matrix factorization based K-means models, In this paper, we first propose a Elastic clustering model using posterior probability (named as Elastic K-means, EKM). The main observation of our Elastic K-means model is that the clustering indicator from K-means can be interpreted as the posterior cluster probabilities during the factorization procedure. The key point of the Elastic K-means is that each data point is assigned to the clusters scattered into several possible classes according to its posterior probabilities. Then we extend our EKM to graph EKM (gEKM), which simultaneously utilizes both feature vectors and pairwise relations between data points. Specifically, gEKM integrates the proposed Elastic K-means with Normalized Cut clustering. In order to evaluate the effectiveness of proposed EKM and gEKM, we implement them on six benchmark datasets. The promising experimental results demonstrate the benefit of the proposed Elastic K-means models.

Elastic K-means and related works

K-means

Given data matrix X = (x1, ⋯, x) ∈ ℜ denoting n data points. The objective of K-means clustering is to find the cluster centroids c, k = 1, ⋯ , K by minimizing the following cost function: Our work starts with the observation that the above K-means clustering objective can be reformulated as below: Proposition 1. The objective function Eq (1) can be rewritten as: where n, (i = 1, ⋯ , k) denotes the number of data points in each cluster. Note that ∥A∥ denotes the Frobenius norm of a matrix A. And G ∈ ℜ is the normalized cluster indicators, i.e., Proof of Proposition 1. Let us introduce the standard cluster indicator Then, Eq (1) can be equivalently written as: where C = (c1, c2, ⋯ , c) ∈ ℜ are cluster centroids. The orthogonality of G is preserved in a relaxation of class indicator matrix [21]. Now, the cluster centroids can be written as , where is the k-th column of , or equivalently, . Thus: where absorbs the unknown .

Elastic K-means

Proposition 1 provides an natural way to form the Elastic K-means. First, the orthogonal constraint G G = I ensures that one data point belongs to only one cluster as in [21, 30]. We relax this constraint, so that each data point can belong to multiple clusters. Because the formulation Eq (2) is a matrix decomposition form, the constraint G G = I will be satisfied approximately. Second, since the mixed signs of G may deviate the results severely from the true solution of clustering [21, 30], we relax the discrete constraint , to a simple nonnegativity constraint G ≥ 0. Thus, the final model is formulated as: Note that the input data X has mixed signs. We call this Elastic K-means (EKM). (1) This model is invariant w.r.t. X → βX, β ∈ ℜ is any constant. This is the same as the standard K-means. (2) Similar to K-means, the expected clustering centroids are adopted during the updating iterations. (3) different from the traditional K-means, the elastic indicator (interpreted as the posterior probability which is a fraction) performs a smooth convergency during updating. In fact, all the NMF-based methods we discussed above have K-means clustering interpretations when the factor G is orthogonal (G G = I), which in turn means they can be regarded as the relaxations of K-means.

Related works

Our goal is to study the case where a data point is not restricted to belong to a single cluster. Three related clustering models are (A) Gaussian mixtures [31, 32], (B) Fuzzy C-means [33-35], (C) Fuzzy K-means [12, 14, 36, 37]. In general, Gaussian mixtures works only for low-dimensional data, typically d ≤ 10, therefore they are not suitable for high dimensional data. Fuzzy C-means solves where u is the membership of data point x belonging to cluster k and ∑ u = 1. Our model differs from Fuzzy C-means on two aspects: (1) Cluster distribution of single data point x is implicit in our model, whereas it is explicit in Fuzzy C-means. (2) The requirement that the posterior probability ∑ u = 1 for any data point x, but the actual distribution used in Eq (8) is , m = 2 in most cases. Thus the normalization of the actual cluster distribution is and z is different for different data point x. This theoretical/conceptual drawback for Fuzzy C-means is not present in our Elastic K-means. Fuzzy K-means solves where w is the membership of data point x belonging to cluster k and ∑ w = 1. The main difference between Fuzzy K-means and our method is, the soft/fuzzy capability is achieved by exploring the corresponding weight w (or u in Fuzzy C-means) for each data point x, while our Elastic K-means achieves this via the soft/fuzzy cluster indicator G, which is more explicit. Note that, the concept of soft/fuzzy K-means (or C-means)has been mentioned in a group of literatures [38-42], which also have the fuzzy/soft capability for data clustering. However, these soft clustering methods are generally derived from Fuzzy C-means or Fuzzy K-means and their variants, which are essentially different from the proposed Elastic K-means model. Because in our model, G is nonnegative. In some sense, our model also relates to NMF. There exists a very broad category of work along NMF direction. We refer to a recent survey [43].

Computational algorithm and analysis

Algorithm

An effective updating algorithm can be derived to solve Elastic K-means problem. The algorithm iteratively updates the current solution as follows: The pseudo codes of EKM is illustrated as Algorithm 1. Algorithm 1: Elastic K-means Algorithm Input: data X 1 (1) Initialize G0. 2 Construct the indicator G: G = 1 if x belongs to cluster k. otherwise, G = 0 3 (2) Update G 4 while not converged 5 6 where 7 A = (|(X X)| + (X X))/2 8 B = (|(X X)| − (X X))/2 9 end The convergence of the proposed algorithm can be found in the supplementary.

Updating algorithms for quartic models

There are many matrix models involving matrix variables to the 4th power. The simplest is: where W = W is symmetric. A number of papers claim that the updating algorithm for this model is In fact, this updating algorithm does not guarrentee the monotonic decreasy of the objective function value. Using the inequality of quartic matrix of Eq (S15) in the supplementary, we can easily prove that the following updating algorithm; which guarrentees decrease of objective, and thus the convergence.

Benefit of Elastic K-means

One main drawback of standard K-means is that each data point is assigned to a single cluster (hard clustering), but in real data, many data points are somewhere in-between different clusters centers (such as points 1, 2, 3 shown in Fig 1). One clear benefit of the proposed Elastic K-means is that these ambiguous data points are assigned into several nearby clusters, i.e., their posterior probabilities of cluster assignment are nearly evenly distributed.
Fig 1

EKM clustering results on 2 dimensional 3 Gaussian clusters.

x and y axes denote the first and the second dimension respectively.

EKM clustering results on 2 dimensional 3 Gaussian clusters.

x and y axes denote the first and the second dimension respectively. Fig 1 shows the EKM results on one 2D dataset of 3 randomly generated Gaussian clusters. The EKM cluster results are indicated as red, green and blue stars respectively. Point 4, 5, 6 clearly belong to their assigned clusters. This fact is correctly encoded in the EKM model, as shown in the posterior probabilities in Fig 2: they are sharply concentrated on a single cluster.
Fig 2

An example of the posterior G of 3 ambiguous data points and 3 sharp data points on above EKM clustering results.

On the other hand, points 1, 2, 3 are ambiguous: they are in-between different cluster centers. A good ‘elastic’ clustering model should assign these points to nearby clusters with nearly-equal posterior probabilities. Indeed, this desirable fact is correctly encoded in the EKM model, as shown in the posterior probabilities in Fig 2: they are nearly evenly distributed on nearby clusters. Posterior Probability. In this and following sections, we normalize G to so that ∑ G = 1 for data point x. A natural question is: how to identify those ambiguous data points such points 1, 2, 3 in Fig 2. For this purpose, we need to see how the posterior probability for data point x is distributed. We use a useful quality to indicate approximately how sharply the distribution is peaked around the highest cluster. This is the gap in the posterior distribution: the difference between the highest and second highest peaks: where The idea is simple. (A) If a data point is well inside a cluster, such as points 4, 5, 6 in Fig 2, the highest peak is large and the second highest peak is small. Thus the gap is big. (B) If a data point is in-between different clusters, such as points 1, 2, 3, the difference between the highest peak and the second highest peak is small. Thus the gap is small. From these situations, the gap is useful indication of how sharply the posterior probability is distributed. Fig 3 shows the distribution of the gap Δ of the 300 data points shown in Fig 1. Noted that, the lager Δ, the sharper that the data point belongs to a single class. Obviously, there are many ambiguous data points which can be easily detected using the gap. These studies indicate that Elastic K-means is able to detect the fuzzy/soft characteristics of data clustering.
Fig 3

The gap in the posterior distribution: The difference between the highest and second highest peaks.

Integrating Elastic K-means and normalized cut

EKM uses only vector (sometimes called attribute) data X. In many applications, the input data consists of both vector data X and similarity data W (also called graph data since W represents the similarity between data points i and j). For this situation, we can easily extend EKM to utilize the similarity data W for more effective elastic clustering. We incorporate the similarity data through the Normalized Cut [18, 19] formalism. This extension form a single clustering formulation for input data with both vector and similarity data. The Ncut (Normalized Cut) is defined as: where D = diag(d1, ⋯ , d), d = ∑ W. Proposition 2. Optimization problem Eq (18) is equivalent to with the same constraints of Eq (18), where Proof. Because ∥S − QQ∥2 = Tr(S2 − 2Q SQ + QQ QQ), the first term and the last term are both constant due to Q Q = I. The minimization of −2Q SQ is the same as Eq (18). In the same manner as EKM in Eq (7), we relax the orthogonal constraint Q Q = I to be nonnegative, then Eq (19) can be rewritten as: Since Q is also the cluster indicators as G in EKM, we can integrate EKM and NCut by simplifying/approximating these constraints into G ≥ 0. The above model is now expressed in a simplified way: We call this model as graph EKM (simplified as gEKM in the following parts) since Ncut is a popular approach for graph data clustering. It naturally incorporates both the feature/attribute data X and pairwise relations W.

Analysis

Because the model is invariant with X → βX, let β = 1/∥X∥, which implies the input X is normalized: . The above objective can be written as: The first 2 terms are independent of G. Thus the above optimization becomes where X is normalized. This expression reveals an important insight that the gEKM model is a discrete combinatorial optimization.

Algorithm 2

The optimization problem of Eq (21) can be solved by the following iterative algorithm: where α1 = α * ∥X∥2, and A, B are defined in Algorithm 1. The proof of convergence can be similarly established as Algorithm 1.

Experiments

We perform experiments on following six benchmark datasets described as Table 1.
Table 1

Dataset names, number of data samples, data dimension and number of classes of the 6 benchmark datasets.

Dataset#Sample#Dimension#Class
AT&T400102440
USPS929825610
MNIST100078410
COIL201440102420
Isolet1156061726
BinAlph101432026
AT & T (http://www.cl.cam.ac.uk/research/dtg/attarchive/facedatabase.html) face dataset contains ten different images of each 40 distinct subjects. For some subjects, the images were taken at different times, varying the lighting, facial expression and facial details (glass/no glass). All images were taken against a dark homogeneous background with the subjects in an upright, frontal, position. We stretch each image into a vector. USPS (http://www.cs.nyu.edu/roweis/data.html) is a handwritten digit (1–10) database and we select 1000 images (100 images for every digit). MNIST (http://www.cs.nyu.edu/roweis/data.html) is a handwritten digit database. Each image is centered (according to the center of mass of the pixel intensities) on a 2828 grid. In our experiments, we randomly choose 1000 images (i.e., each digit has 100 images). We reshape each image into one vector. COIL20 (http://www.cs.columbia.edu/CAVE/software/softlib/coil-20.php) contains 20 objects. Each image of the same object is taken 5 degrees apart as the object is rotated on a turntable and each object has 72 images. The size of each image is 3232 pixels, with 256 grey levels per pixel. Each image is represented by a 1024 dimensional vector. Isolet1 (http://archive.ics.uci.edu/ml/datasets/ISOLET) dataset was generated by 150 subjects spoken the name of each letter of the alphabet twice. The speakers are grouped into sets of 30 speakers each, and are referred to as isolet1, isolet2, isolet3, isolet4, isolet5. The features include spectral coefficients, contour features, sonorant features, pre-sonorant features, and post-sonorant features. In our experiment, we utilize subset isolet1 only. BinAlpha (http://www.cs.nyu.edu/roweis/data.html) contains 26 binary hand-written alphabets and we select 30 images for every alphabet. We stretch each image into one vector. Clustering accuracy is used to measure the performances. Once the clustering solution is computed, the confusion matrix is calculated using the Hungarian algorithm [44] is employed to match obtained clusters with the ground truth classes. The accuracy of this matching is computed as the clustering accuracy. In all experiments, for each dataset, we first run 20 K-means clustering with random starts. We pick the best (lowest clustering objective function value) result of these 20 runs, record the corresponding clustering accuracy as the result of K-means. Starting from this K-means result, we run NMF [11], FCM [34], MinMaxCut [21], EKM, gEKM, and record the corresponding results. For Ncut [18], we run K-means in the eigenspace 20 times with random starts, and pick the best (lowest K-means clustering objective function value) result of these 20 runs.

Choice of parameter α

The clustering accuracy of gEKM depends on parameter α. If α > 1, the similarity part (2nd term of Eq (21)) is more important. If α < 1, the vector attribute part (first term of Eq (21)) is more important. For this reason, the proper range of choice for α is [1/10, 1/5, 1/2, 1, 2, 5, 10]. We show the clustering results against α on all six datasets in Fig 4 (here X is in full dimension).
Fig 4

Clustering accuracy for gEKM against α on six datasets in full dimension.

We observe that (1) The results are not very sensitive to α. (2) α = 1 generally gives best results, which in turn means the vector attribute and the similarity part have more or less the balanced contribution. From these observations, we fix α to 1 in the following experiments.

Convergency visualization

Fig 5 visually illustrates the changes of the objective functions of Eqs (7) and (21), which demonstrates the convergency of the objective functions.
Fig 5

Changes of the objective function of the EKM and gEKM in 100 iterations.

Results on clustering

We compare our methods, EKM and gEKM to K-means, NMF [11], FCM [34], Ncut [18] and MixMaxCut [21] on the six datasets. For EKM, Ncut [18] and MinMaxCut [21], we compute the graph similarity W as: where c = 0.7 and d is the average distances between each x and its 7 nearest neighbors. We set W = 0 and construct . Table 2 reports the clustering accuracy on the six datasets using K-means, NMF, FCM, Ncut, MinMaxCut, EKM and gEKM. One can see that (1) Ncut and MinMaxCut consistently outperform the existing NMF, FCM and Kmeans methods. (2) The proposed EKM beats Ncut and MinMaxCut in most of the cases. (3) MinMaxCut and Ncut perform nearly to proposed EKM and slightly better in some cases. (4) The integrated gEKM consistently outperforms the other six methods including MinMaxCut and Ncut which verifies the effectiveness of the integration. In order to elaborate the performance of proposed methods, we evaluate the clustering experiments in different PCA subspace with the dimension of p = 50, 100, 150, 200 on the six datasets. The obtained clustering accuracies are given in Tables 3 to 6. First of all, the results on different PCA subspaces are consistent as the full space as we discussed above. Furthermore, as shown in Fig 6, the clustering results on PCA subspace can achieve competitive or even slight better than the full space, which implies the effectiveness of subspace discovery.
Table 2

Clustering accuracy of six datasets in full dimension.

DatasetKmeansFCMNMFNcutMinMaxCutEKMgEKM
AT&T0.7150.7010.7210.7310.7300.7330.746
USPS0.6180.6260.6240.6440.6460.6210.656
MNIST0.4940.5010.5180.5310.5320.5340.545
COIL200.6250.6150.6310.6380.6350.6340.643
Isolet10.6260.6240.6280.6510.6450.6540.663
BinAlph0.4950.5110.5020.5110.5140.5120.532
Table 3

Clustering accuracy of six datasets in PCA subspace when p = 50.

DatasetKmeansFCMNMFNcutMinMaxCutEKMgEKM
AT&T0.7130.7130.7150.7360.7310.7340.748
USPS0.6040.6190.6210.6270.6320.6360.637
MNIST0.5210.5230.5240.5450.5370.5360.551
COIL200.6200.6310.6270.6330.6400.6520.658
Isolet10.6260.6270.6260.6410.6350.6460.653
BinAlph0.4980.5100.5180.5340.5330.5410.553
Table 6

Clustering accuracy of six datasets in PCA subspace when p = 200.

DatasetKmeansFCMNMFNcutMinMaxCutEKMgEKM
AT&T0.7120.7080.7020.7310.7250.7230.736
USPS0.6100.6160.6250.6290.6310.6410.643
MNIST0.5120.5030.5180.5250.5320.5390.545
COIL200.6230.6110.6210.6260.6300.6360.643
Isolet10.6190.6240.6250.6310.6290.6330.644
BinAlph0.4770.5120.5150.5220.5250.5280.537
Fig 6

Clustering accuracy of gEKM against the number of dimension of PCA subspace on six datasets.

where “full” on the x-axis denotes the full dimension of the datasets as indicated on Table 1.

Clustering accuracy of gEKM against the number of dimension of PCA subspace on six datasets.

where “full” on the x-axis denotes the full dimension of the datasets as indicated on Table 1.

Conclusion

In this paper, we firstly propose a Elastic K-means(EKM) framework which provides a elastic clustering solution and prove that the correctness and the convergency of the updating algorithm. Secondly, an integrated framework specified by the combination of EKM and Normalized Cut (gEKM) is proposed to take into account of both attribute data and pairwise relations and the choice of combination parameter is particularly analysed. The experimental results on six benchmark datasets demonstrate the proposed EKM leads to the better clustering accuracy than K-means as well as previous NMF-based algorithms and the popular Fuzzy C-means and Fuzzy K-means. The gEKM has been also been tested to have satisfied performance on data clustering.

Supplementary materials.

(PDF) Click here for additional data file.

Supplementary materials.tex file.

(TEX) Click here for additional data file.
Table 4

Clustering accuracy of six datasets in PCA subspace when p = 100.

DatasetKmeansFCMNMFNcutMinMaxCutEKMgEKM
AT&T0.7110.7010.7110.7370.7310.7380.749
USPS0.6080.6250.6250.6320.6220.6350.641
MNIST0.5160.5210.5250.5340.5350.5390.546
COIL200.6180.6250.6300.6330.6340.6370.650
Isolet10.6280.6230.6270.6370.6390.6450.656
BinAlph0.4810.5150.5180.5230.5220.5280.534
Table 5

Clustering accuracy of six datasets in PCA subspace when p = 150.

DatasetKmeansFCMNMFNcutMinMaxCutEKMgEKM
AT&T0.7170.7020.7020.7350.7200.7220.741
USPS0.6050.6120.6270.6310.6230.6370.638
MNIST0.5190.5110.5280.5370.5390.5410.551
COIL200.6070.6220.6290.6310.6330.6380.644
Isolet10.6270.6220.6310.6420.6400.6450.658
BinAlph0.4850.5120.5190.5260.5310.5360.538
  15 in total

1.  Nonlinear dimensionality reduction by locally linear embedding.

Authors:  S T Roweis; L K Saul
Journal:  Science       Date:  2000-12-22       Impact factor: 47.728

2.  A global geometric framework for nonlinear dimensionality reduction.

Authors:  J B Tenenbaum; V de Silva; J C Langford
Journal:  Science       Date:  2000-12-22       Impact factor: 47.728

3.  A Convergence Theorem for the Fuzzy ISODATA Clustering Algorithms.

Authors:  J C Bezdek
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  1980-01       Impact factor: 6.226

4.  Constrained Nonnegative Matrix Factorization for Image Representation.

Authors:  Haifeng Liu; Zhaohui Wu; Xuelong Li; Deng Cai; Thomas S Huang
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  2011-11-08       Impact factor: 6.226

5.  Spectral embedded clustering: a framework for in-sample and out-of-sample spectral clustering.

Authors:  Feiping Nie; Zinan Zeng; Ivor W Tsang; Dong Xu; Changshui Zhang
Journal:  IEEE Trans Neural Netw       Date:  2011-09-29

6.  Convex and semi-nonnegative matrix factorizations.

Authors:  Chris Ding; Tao Li; Michael I Jordan
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  2010-01       Impact factor: 6.226

7.  Regression reformulations of LLE and LTSA with locally linear transformation.

Authors:  Shiming Xiang; Feiping Nie; Chunhong Pan; Changshui Zhang
Journal:  IEEE Trans Syst Man Cybern B Cybern       Date:  2011-10

8.  Extending the Functionality of Behavioural Change-Point Analysis with k-Means Clustering: A Case Study with the Little Penguin (Eudyptula minor).

Authors:  Jingjing Zhang; Kathleen M O'Reilly; George L W Perry; Graeme A Taylor; Todd E Dennis
Journal:  PLoS One       Date:  2015-04-29       Impact factor: 3.240

9.  NMF-mGPU: non-negative matrix factorization on multi-GPU systems.

Authors:  Edgardo Mejía-Roa; Daniel Tabas-Madrid; Javier Setoain; Carlos García; Francisco Tirado; Alberto Pascual-Montano
Journal:  BMC Bioinformatics       Date:  2015-02-13       Impact factor: 3.169

10.  BiCAMWI: A Genetic-Based Biclustering Algorithm for Detecting Dynamic Protein Complexes.

Authors:  Amir Lakizadeh; Saeed Jalili
Journal:  PLoS One       Date:  2016-07-27       Impact factor: 3.240

View more
  1 in total

1.  Systematic analysis of HD-ZIP transcription factors in sesame genome and gene expression profiling of SiHD-ZIP class I entailing drought stress responses at early seedling stage.

Authors:  Maryam Mehmood; Muhammad Jadoon Khan; Muhammad Jawad Khan; Nadeem Akhtar; Fizza Mughal; Syed Tahir Abbas Shah; Muhammad Zeeshan Hyder; Sumaira Farrakh; Irfan Sadiq
Journal:  Mol Biol Rep       Date:  2022-01-07       Impact factor: 2.316

  1 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.