Literature DB >> 27552203

Performance Evaluation of Missing-Value Imputation Clustering Based on a Multivariate Gaussian Mixture Model.

Jing Xiao1, Qiongqiong Xu1, Chuanli Wu1, Yuexia Gao1, Tianqi Hua1, Chenwu Xu2.   

Abstract

BACKGROUND: It is challenging to deal with mixture models when missing values occur in clustering datasets. METHODS AND
RESULTS: We propose a dynamic clustering algorithm based on a multivariate Gaussian mixture model that efficiently imputes missing values to generate a "pseudo-complete" dataset. Parameters from different clusters and missing values are estimated according to the maximum likelihood implemented with an expectation-maximization algorithm, and multivariate individuals are clustered with Bayesian posterior probability. A simulation showed that our proposed method has a fast convergence speed and it accurately estimates missing values. Our proposed algorithm was further validated with Fisher's Iris dataset, the Yeast Cell-cycle Gene-expression dataset, and the CIFAR-10 images dataset. The results indicate that our algorithm offers highly accurate clustering, comparable to that using a complete dataset without missing values. Furthermore, our algorithm resulted in a lower misjudgment rate than both clustering algorithms with missing data deleted and with missing-value imputation by mean replacement.
CONCLUSION: We demonstrate that our missing-value imputation clustering algorithm is feasible and superior to both of these other clustering algorithms in certain situations.

Entities:  

Mesh:

Year:  2016        PMID: 27552203      PMCID: PMC4994954          DOI: 10.1371/journal.pone.0161112

Source DB:  PubMed          Journal:  PLoS One        ISSN: 1932-6203            Impact factor:   3.240


Introduction

Clustering analysis, as a multivariate statistical method, refers to the process of classifying a set of observations into subsets, called clusters, such that observations in the same cluster are similar in certain respects [1-3]. Clustering is widely used in medical sciences, for instance when clustering diseases or gene-expression profiles. Clustering methods usually fall into two categories: hierarchical clustering methods [4], which are used for clustering datasets with small size [5,6]; and dynamic clustering methods, such as K-means [7,8] and self-organizing maps [9], which begin with an initial partitioning of the individuals and iteratively move individuals from one cluster to another until the criterion of convergence is met. With dynamic clustering, the number of clusters must be specified in advance [10]. Dynamic algorithms are mostly heuristically motivated, and they do not require an underlying statistical model. Nevertheless, selection of the “correct” number of clusters and of the best clustering method remains a topic for discussion. Model-based clustering methods [11-13] are a type of dynamic clustering based on the hypothesis that the whole dataset is a finite mixture of the same type of distribution with different sets of parameters, such as a finite mixture of a multivariate Gaussian distribution. Compared with “heuristic” algorithms, one obvious advantage to model-based clustering is that objective statistical criteria—such as the Akaike information criterion (AIC) or the Bayesian information criterion (BIC) [14]—are used to determine the number of clusters. Missing data present a problem for medical research, given that the data are often supplied retrospectively and from various sources [15]. In particular, missing data are a frequent occurrence in microarray experiments and in medical research. Missing values are especially common in large-scale studies involving dozens of variables and hundreds of individuals. Indeed, many clustering methods require a full set of data. Individuals with missing values are either rejected or been estimated prior to the analysis. Consequently, several missing-value imputation methods have been developed [16-20], such as mean substitution, regression imputation, fuzzy c-means (FCM) clustering of incomplete data [21], and Gaussian mixture model-based missing-value imputation classification [22]. In this study, we propose a dynamic method for a model-based missing-value imputation clustering algorithm. With our proposed method, missing values are estimated iteratively until arithmetic convergence is reached during the process of clustering individuals. We used 12 simulated datasets. Each of these datasets had two versions: a complete version and a version where 10% of the individuals had at least one variable missing. We used these datasets to evaluate the clustering accuracy and the accuracy and precision of cluster-parameter estimators. We compared our proposed algorithm based on the missing-value imputation according to the maximum likelihood estimation (a “pseudo-complete” dataset) with a model-based clustering algorithm using the complete dataset and another two model-based clustering algorithms, one using a dataset with missing values deleted and the other imputing the missing values by mean replacement. In addition, we compared our algorithm with the FCM clustering method using real datasets—viz., Fisher’s Iris dataset, the Yeast Cell-cycle Gene-expression dataset, and the CIFAR-10 images dataset.

Methods

Multivariate Gaussian mixture model

The dataset is arranged as an n × k matrix denoted by Y, where n is the number of individuals and k is the number of variables. Let y be the observed value of the i-th individual in the j-th variable, for i = 1,2,…, n and j = 1,2,…, k. Let Y = (y,y,…,y)T be the i-th column in matrix YT—i.e., a k × 1 vector of the data for individual i under all variables. The value of y across all the k variables represents the expression level of the i-th individual. Under a finite multivariate Gaussian mixture model, each Y is assumed to follow a k-dimensional Gaussian mixture distribution. Mathematically, the mixture distribution for c clusters is as follows: where is the probability density function for the l-th k-dimensional Gaussian distribution with mean vector μ = (μ, μ,…, μ) and variance-covariance matrix , for l = 1,2,…,c. Moreover, P with , is the mixing proportion of cluster l. The mixing proportion p is defined as the proportion of individuals that belong to the l-th cluster. The joint log-likelihood function for n independent individual vectors is defined as follows:

Clustering algorithm for missing values

Most reports [23-25] consider only complete datasets (without missing values) for Gaussian mixture clustering. Indeed, it is challenging to estimate missing values accurately. An urgent problem concerns how to provide the optimal number of clusters c and how to cluster n individuals into the c clusters precisely. Therefore, we propose a missing-value imputation algorithm as follows. Suppose that there are r missing values in Y. Thus, y, y,…, y are missing. The Y, μ, and S are then divided into where Suppose further that Y is from the l-th k-dimensional Gaussian distribution and that Y is known. The conditional expectation of Y, which belongs to the l-th cluster, is derived as follows: We provide the initial value for μ, S, and S. Then, based on the criterion of the minimum mean-square deviations from Eq (3), the conditional expectation E(Y|Y) can be calculated, representing the best predicted function for Y [26]. Under the mixture distribution of the individual vector Y, the conditional expectation for Y based on Y is where E(Y|Y) denotes the estimators for the missing values Y. Subsequently, a “pseudo-complete” dataset can be constructed.

Iterative clustering algorithm

The proposed model-based clustering algorithm for datasets with missing values assigns each individual to one of the c clusters with a certain probability. We define this probability as p, which is that of the i-th individual belonging to the l-th cluster. An individual is assigned to the l-th cluster if p is greater than a certain pre-determined threshold. The probability can be calculated using the expectation-maximization (EM) algorithm [27] to achieve the maximum likelihood for the objective function derived with Eq (2). The EM algorithm begins with some initial parameters that are set in advance. Then, each parameter is iteratively updated in the algorithm until convergence to a minimizer is reached. An EM iteration includes the following steps: Initialize the prior probabilities of the cluster assignment and the cluster parameters: Calculate the missing values using Eqs (3) and (4) to generate the “pseudo-complete” dataset for . Update the posterior probabilities of the cluster assignment: Update the cluster proportions, mean vectors, and variance-covariance matrices: Repeat steps (1)–(4) until convergence is reached. The number of clusters c can also be treated as an unknown parameter and inferred by the BIC or AIC tests. The BIC is derived as follows: where q = c(k+1)(k+2)/2−1 is the number of independent parameters to be estimated in the model, is the likelihood of —i.e., the vector for the maximum likelihood estimation of the parameters—and n is the size of the dataset. The number of clusters c is determined by the maximum BIC value.

Results

Simulation analysis

A simulation was designed to evaluate the feasibility and accuracy of the proposed missing-values imputation algorithm. The individuals belonging to each cluster were known exactly and without subjective errors. Indeed, when analyzing real data, the misjudgment rate (MR) may reflect confounding errors in the experiment and the subjective errors. Without loss of generality, we simulated datasets with a two-dimensional Gaussian distribution and some missing values. All simulations were performed using statistical SAS (version 9.3; SAS/IML) software.

Design of the simulation

Suppose that there are 500 individuals derived from one of two (two-dimensional) Gaussian populations, denoted by MVN (μ1,S1) and MVN (μ2,S2), from which 10% of the individuals have at least one variable randomly removed. The entire dataset contains 1000 individuals. Each individual has two variables, and 100 individuals have at least one variable missing. The cluster’s mean vectors are denoted by μ1 and μ2, which were simulated at three different levels: A1: μ1 = (0, 0), μ2 = (2.5, 2.5), A2: μ1 = (0, 0), μ2 = (2.0, 2.0), and A3: μ1 = (0, 0), μ2 = (1.5, 1.5). The variance-covariance matrices are denoted by S1 and S2 (without loss of generality, supposing that S1 = S2 = S), and these were set at four levels: , , , and . The total number of treatment combinations (datasets) is therefore 3 × 4 = 12, that is, A1B1, A1B2, A1B3, A1B4, A2B1, A2B2, A2B3, A2B4, A3B1, A3B2, A3B3, and A3B4. Twenty replicated simulations were conducted for each of the twelve scenarios. Our missing-value imputation cluster algorithm (denoted by M-3, generating a “pseudo-complete” dataset) was compared with three other cluster algorithms: a cluster algorithm using the complete dataset (denoted by M-1, i.e., the above simulation dataset without any missing values); a cluster algorithm using a dataset from which all individuals with missing variables are deleted (denoted by M-2, i.e., a cropped dataset); and a cluster algorithm using missing-value imputation by mean replacement in which the missing values are replaced by the mean value of the whole dataset (denoted by M-4, another “pseudo-complete” dataset). Thus, we used these four algorithms to cluster simulated datasets. These algorithms are all based on the multivariate Gaussian mixture model, using the maximum likelihood estimation with the EM algorithm. As such, the clustering results could be fairly compared. We evaluated M-1, M-2, M-3, and M-4 using the following metrics: (1) the average convergence rate; (2) the accuracy and precision of their respective parameter estimates; and (3) the total misjudgment rate (MR), where MR is the ratio of all misjudged individuals to the total number of individuals in the 20 replicates and MR = 1 - accuracy. A chi-square test and multiple comparisons of Scheffé’s Confidence Interval were used to test the differences in MR among the four algorithms.

Simulation results

The average parameter estimates include the mean vectors (± se), variance-covariance matrices (± se), the likelihood value of the maximum-likelihood function, and the total MR for the four clustering algorithms. The results are presented in Tables 1–3. Our proposed missing-value imputation clustering algorithm (M-3) outperformed the other two algorithms (M-2 and M-4) in terms of the convergence rate for most of the simulation datasets. Indeed, its superiority was most obvious when the mean vectors of two clusters were near each other and the variance-covariance of each cluster was more disperse, such as with datasets A2B4, A3B3, and A3B4. Moreover, the proposed algorithm resulted in the most accurate and precise parameter estimations, closely approximating the true values.
Table 1

Average parameter estimates under 4 different simulation datasets A1B1-A1B4 in 20 replicates.

TreatmentIterative timeLikelihood valueProbability estimateMean vector estimateCovariance matrix estimateMR (%)
P^1P^2μ^1 ± seμ^2 ± seS^1 ± seS^2 ± se
A1B1M-151-1876.320.500.500.00±0.010.00±0.012.49±0.072.50±0.060.25±0.040.15±0.030.25±0.040.15±0.020.20a
0.15±0.030.26±0.040.15±0.020.25±0.04
M-258-1864.940.500.500.00±0.01-0.01±0.022.47±0.082.53±0.070.24±0.050.14±0.030.25±0.030.15±0.020.30a
0.14±0.030.26±0.040.15±0.020.24±0.04
M-364-1897.050.500.500.00±0.020.00±0.012.49±0.072.50±0.060.24±0.040.14±0.020.24±0.040.14±0.040.20a
0.14±0.020.26±0.030.14±0.040.25±0.04
M-455-1800.380.490.49-0.01±0.030.00±0.022.46±0.082.46±0.090.22±0.070.13±0.040.23±0.060.14±0.051.50b
0.13±0.040.23±0.060.14±0.050.21±0.07
A1B2M-170-2578.470.500.500.01±0.030.00±0.042.47±0.142.47±0.140.48±0.060.28±0.040.47±0.070.28±0.042.40a
0.28±0.040.53±0.080.28±0.040.48±0.06
M-280-2362.690.500.50-0.01±0.040.01±0.042.52±0.152.48±0.160.48±0.070.32±0.040.47±0.070.26±0.062.80a
0.32±0.040.54±0.080.26±0.060.46±0.08
M-387-2591.610.500.50-0.01±0.040.01±0.032.47±0.142.46±0.160.47±0.060.27±0.050.46±0.070.28±0.042.60a
0.27±0.050.53±0.070.28±0.040.48±0.06
M-493-2653.930.480.480.02±0.090.01±0.082.43±0.262.44±0.230.42±0.140.25±0.110.44±0.120.25±0.115.64b
0.25±0.110.43±0.130.25±0.110.44±0.12
A1B3M-1107-2837.520.490.510.02±0.07-0.01±0.082.47±0.212.47±0.210.68±0.160.42±0.100.71±0.140.43±0.095.90 ab
0.42±0.100.78±0.150.43±0.090.75±0.14
M-2142-2569.690.500.50-0.02±0.060.01±0.062.48±0.202.52±0.190.73±0.160.47±0.080.69±0.160.40±0.095.45a
0.47±0.080.81±0.150.40±0.090.67±0.17
M-3134-2822.260.490.51-0.01±0.07-0.02±0.062.45±0.212.46±0.200.65±0.170.39±0.080.73±0.140.42±0.076.12ab
0.39±0.080.76±0.120.42±0.070.68±0.15
M-4138-2859.620.450.55-0.05±0.150.04±0.122.40±0.372.43±0.320.63±0.210.40±0.160.70±0.190.39±0.118.94b
0.40±0.160.60±0.200.39±0.110.71±0.19
A1B4M-1180-3211.910.490.52-0.05±0.10-0.03±0.122.40±0.252.42±0.280.86±0.240.51±0.170.94±0.240.59±0.158.50a
0.51±0.171.05±0.260.59±0.151.02±0.23
M-2214-2852.780.480.52-0.04±0.10-0.06±0.142.38±0.302.37±0.320.84±0.250.55±0.160.97±0.190.56±0.169.20a
0.55±0.161.05±0.260.56±0.160.94±0.25
M-3215-3100.000.480.52-0.05±0.12-0.06±0.132.39±0.262.38±0.300.84±0.230.48±0.190.90±0.250.57±0.148.70a
0.48±0.191.02±0.210.57±0.140.99±0.20
M-4267-3289.320.430.57-0.07±0.20-0.08±0.252.17±0.482.22±0.520.77±0.320.43±0.280.82±0.270.48±0.1518.90b
0.43±0.281.09±0.230.48±0.151.07±0.25

M-1 indicates a complete data clustering algorithm; M-2 indicates a missing-data-deleted clustering algorithm; M-3 indicates our missing-value imputation clustering algorithm

M-4 indicates the clustering algorithm for missing-value imputation by mean replacement. These four algorithms are based on the multivariate Gaussian mixture model.

indicates the multiple comparisons of the differences in MR among the algorithms M-1, M-2, M-3, and M-4: having the different letters indicate that there is statistical significance between these two groups (P<0.05), and vice versa.

Table 3

Average parameter estimates under 12 different simulation datasets A3B1-A3B4 in 20 replicates.

TreatmentIterative timeLikelihood valueProbability estimateMean vector estimateCovariance matrix estimateMR (%)
P^1P^2μ^1 ± seμ^2 ± seS^1 ± seS^2 ± se
A3B1M-1129-1816.340.500.500.00±0.060.01±0.061.47±0.151.48±0.140.23±0.080.13±0.050.23±0.090.14±0.064.90a
0.13±0.050.26±0.090.14±0.060.25±0.10
M-2337-1637.810.510.490.01±0.070.01±0.061.53±0.151.47±0.160.23±0.080.14±0.060.23±0.100.12±0.065.35a
0.14±0.060.27±0.100.12±0.060.22±0.09
M-3134-1762.640.500.50-0.01±0.070.00±0.061.46±0.161.46±0.160.21±0.080.13±0.050.21±0.090.14±0.055.00a
0.13±0.050.26±0.080.14±0.050.25±0.08
M-4359-1800.320.520.48-0.02±0.090.03±0.081.44±0.251.45±0.260.20±0.140.13±0.070.21±0.130.14±0.069.20b
0.13±0.070.24±0.130.14±0.060.25±0.13
A3B2M-1312-2322.430.450.55-0.06±0.10-0.03±0.071.43±0.261.45±0.250.45±0.230.22±0.110.45±0.230.32±0.0911.50a
0.22±0.110.55±0.230.32±0.090.53±0.22
M-2444-2102.310.450.55-0.08±0.110.07±0.101.41±0.271.42±0.260.41±0.250.23±0.130.45±0.240.30±0.1013.45ab
0.23±0.130.47±0.250.30±0.100.54±0.23
M-3300-2296.720.460.54-0.06±0.10-0.03±0.081.42±0.261.43±0.250.44±0.230.25±0.100.46±0.220.30±0.0912.00a
0.25±0.100.54±0.220.30±0.090.54±0.21
M-4489-2358.990.430.57-0.09±0.250.07±0.261.39±0.331.40±0.350.41±0.280.25±0.140.43±0.290.30±0.1316.50 b
0.25±0.140.54±0.290.30±0.130.52±0.28
A3B3M-1584-2586.530.420.60-0.15±0.28-0.14±0.261.38±0.461.40±0.470.68±0.320.49±0.180.80±0.310.49±0.1618.00a
0.49±0.180.72±0.310.49±0.160.78±0.30
M-2603-2364.520.390.63-0.19±0.35-0.18±0.331.38±0.521.37±0.500.64±0.350.48±0.170.81±0.320.46±0.1823.30 b
0.48±0.170.65±0.340.46±0.180.75±0.32
M-3310-2578.770.440.58-0.10±0.25-0.10±0.261.40±0.451.39±0.450.65±0.310.44±0.170.74±0.290.47±0.1518.00a
0.44±0.170.78±0.320.47±0.150.78±0.29
M-4590-2600.620.370.63-0.18±0.38-0.17±0.411.38±0.551.37±0.540.63±0.390.42±0.250.82±0.390.48±0.2125.00b
0.42±0.250.68±0.400.48±0.210.75±0.39
A3B4M-1645-3013.370.350.65-0.13±0.46-0.17±0.381.30±0.721.30±0.700.92±0.380.50±0.201.14±0.400.68±0.2428.00b
0.50±0.201.08±0.340.68±0.240.97±0.39
M-2690-2633.700.200.80-0.22±0.74-0.15±0.621.00±0.991.20±0.980.70±0.570.38±0.401.26±0.621.01±0.3848.32c
0.38±0.400.79±0.511.01±0.381.23±0.58
M-3520-2892.210.400.60-0.11±0.340.15±0.311.34±0.661.37±0.630.93±0.360.62±0.181.07±0.370.69±0.2021.30a
0.62±0.181.07±0.390.69±0.200.96±0.41
M-4835-2901.330.100.90-0.58±0.74-0.55±0.800.90±1.040.97±1.070.61±0.720.21±0.581.26±0.770.89±0.4756.00d
0.21±0.580.62±0.740.89±0.471.23±0.77

M-1 indicates a complete data clustering algorithm; M-2 indicates a missing-data-deleted clustering algorithm; M-3 indicates our missing-value imputation clustering algorithm

M-4 indicates the clustering algorithm for missing-value imputation by mean replacement. These four algorithms are based on the multivariate Gaussian mixture model.

indicates the multiple comparisons of the differences in MR among the algorithms M-1, M-2, M-3, and M-4: having the different letters indicate that there is statistical significance between these two groups (P<0.05), and vice versa.

M-1 indicates a complete data clustering algorithm; M-2 indicates a missing-data-deleted clustering algorithm; M-3 indicates our missing-value imputation clustering algorithm M-4 indicates the clustering algorithm for missing-value imputation by mean replacement. These four algorithms are based on the multivariate Gaussian mixture model. indicates the multiple comparisons of the differences in MR among the algorithms M-1, M-2, M-3, and M-4: having the different letters indicate that there is statistical significance between these two groups (P<0.05), and vice versa. M-1 indicates a complete data clustering algorithm; M-2 indicates a missing-data-deleted clustering algorithm; M-3 indicates our missing-value imputation clustering algorithm M-4 indicates the clustering algorithm for missing-value imputation by mean replacement. These four algorithms are based on the multivariate Gaussian mixture model. indicates the multiple comparisons of the differences in MR among the algorithms M-1, M-2, M-3, and M-4: having the different letters indicate that there is statistical significance between these two groups (P<0.05), and vice versa. M-1 indicates a complete data clustering algorithm; M-2 indicates a missing-data-deleted clustering algorithm; M-3 indicates our missing-value imputation clustering algorithm M-4 indicates the clustering algorithm for missing-value imputation by mean replacement. These four algorithms are based on the multivariate Gaussian mixture model. indicates the multiple comparisons of the differences in MR among the algorithms M-1, M-2, M-3, and M-4: having the different letters indicate that there is statistical significance between these two groups (P<0.05), and vice versa. Thus, our proposed missing-value imputation clustering algorithm correctly clusters individuals with missing variables. The total MR resulting from the proposed algorithm (M-3) demonstrates that its performance is superior to that of M-4 in all simulation trials, with significant statistical difference (p < 0.05), and superior to that of M-2 in simulation trials A2B4, A3B3, and A3B4, with significant statistical difference (p < 0.05). Moreover, the clustering accuracy of our proposed algorithm was comparable to that of the complete dataset (M-1) (p > 0.05). In general, the more similar the mean vectors of two clusters or the more disperse their variance-covariance, the higher the MR of all clustering methods and the better the performance of our proposed algorithm compared to M-2 and M-4. Interestingly, however, our algorithm performed well in such cases. For example, in scenario A3B4, the speed of convergence with M-2 and M-4 was lower in both cases than for other scenarios and had a total MR of 48.32% and 56.00%, respectively, whereas the total MR with our proposed algorithm M-3 was 21.30%. In fact, this result was slightly better than M-1, which resulted in an MR of 28.00%.

Real data analysis

Iris dataset

Fisher’s Iris dataset (Fisher, 1936) [12] is perhaps the best-known database for comparing clustering algorithms. The dataset contains three clusters with 50 individuals each, where each cluster is a species of the genus Iris: namely, Iris setosa, Iris versicolor, and Iris virginica. These three species are thus segregated. The dataset contains four variables: sepal length, sepal width, petal length, and petal width. Fisher’s Iris dataset is available on the Internet (http://archive.ics.uci.edu/ml/datasets/Iris). We randomly removed one of the four variables in 10% of the individual observations of Iris data; the resulting dataset constituted a dataset with missing values. Table 4 lists the clustering results from five clustering algorithms: the above four clustering algorithms [the clustering algorithm with a complete dataset (M-1), the clustering algorithm in which individuals with missing variables are deleted (M-2), our proposed missing-value imputation algorithm (M-3), and the clustering algorithm for missing-value imputation by mean replacement (M-4)] and the FCM clustering method applied to the dataset with data missing from 10% of the individuals. In the M-4 and FCM methods, the missing values are replaced by the mean of the whole dataset. The superiority of our proposed method is obvious insofar as M-3 resulted in the lowest MR of the five algorithms. Because of Gaussian mixture distribution of the dataset, our algorithm makes full use of the known data based on the Gaussian mixture model to estimate missing values, the accuracy of our method was higher than those of the FCM method and the M-2 and M-4 algorithms, and it closely approximated the accuracy obtained using the complete-dataset clustering algorithm (M-1). Moreover, the model-based algorithms (M-1, M-2, M-3, and M-4) resulted in the highest BIC values in the three clusters, validating the effectiveness of using the BIC to determine the number of clusters.
Table 4

The comparison among four algorithms and FCM method for Fisher’s Iris dataset.

MethodnMisjudgment rate (MR)Iterative timeBIC value in three clusters
M-11503 (2.0%)37-307.848
M-21356 (4.4%)43-273.539
M-31503 (2.0%)36-305.211
M-41507 (4.7%)49-307.382
FCM1505 (3.3%)

M-1 indicates the complete data clustering algorithm; M-2 indicates missing data deleted clustering algorithm; M-3 indicates our missing-value imputation clustering algorithm; M-4 indicates clustering algorithm for missing-value imputation by mean replacement; FCM indicates Fuzzy C-Means clustering.

M-1 indicates the complete data clustering algorithm; M-2 indicates missing data deleted clustering algorithm; M-3 indicates our missing-value imputation clustering algorithm; M-4 indicates clustering algorithm for missing-value imputation by mean replacement; FCM indicates Fuzzy C-Means clustering.

Yeast Cell-cycle dataset

The Yeast Cell-cycle dataset [28] shows the fluctuation in the expression levels of the whole genome (6220 genes) over 17 time points taken at 10-minute intervals, covering nearly two cell cycles. The entire raw dataset is available at http://genome-www.stanford.edu/cellcycle/. Among the effectively expressed genes, 416 genes were found to have consistent periodic changes in expression level. Yeung et al. picked out 384 genes over 15 time points that reached their peaks at five (C = 5) cell phases on the basis of the size of the buds, after which five non-intersecting clusters were classified [29]. These 384 gene expressions, a subset with a moderate sample size, are available at http://www.cs.washington.edu/homes/kayee/model. All 384 genes were assigned to one of the five clusters by the original researchers [28,29]. Therefore, we can use this dataset to test the performance of the clustering algorithms introduced here and compare them with the performance of existing methods. We randomly removed one of the 15 variable values in 10% of the genes to create a dataset with missing values. Table 5 lists the clustering results from five clustering algorithms: the above four model-based clustering algorithms (M-1, M-2, M-3, and M-4) and the FCM clustering method. Further validating the results from Fisher’s Iris dataset, we found that our missing-value imputation algorithm had the lowest MR. In addition, because our missing-value imputation based on the Gaussian mixture distribution is compatible with the Yeast Cell-cycle dataset, the accuracy of our method was significantly higher than that of the FCM method, and it closely approximated the accuracy from using the complete-dataset clustering algorithm (M-1).
Table 5

The comparison among four algorithms and FCM method for Yeast Cell-cycle dataset.

MethodnMisjudgment rate (MR)Iterative timeBIC value in five clusters
M-138465 (16.9%)1265-6472.448
M-234662 (17.9%)1340-5892.953
M-338462 (16.1%)1260-6316.730
M-438478 (20.3%)1307-6490.706
FCM38480 (20.8%)

M-1 indicates the complete data clustering algorithm; M-2 indicates missing data deleted clustering algorithm; M-3 indicates our missing-value imputation clustering algorithm; M-4 indicates the clustering algorithm for missing-value imputation by mean replacement; FCM indicates Fuzzy C-Means clustering.

M-1 indicates the complete data clustering algorithm; M-2 indicates missing data deleted clustering algorithm; M-3 indicates our missing-value imputation clustering algorithm; M-4 indicates the clustering algorithm for missing-value imputation by mean replacement; FCM indicates Fuzzy C-Means clustering.

CIFAR-10 images dataset

The CIFAR-10 dataset [30] contains 60,000 32 × 32 color images (3072 variables) divided into ten classes, which represent different kinds of pictures. The classes are designed to be completely mutually exclusive. For example, neither automobile pictures nor truck pictures contain images of pickup trucks. Having a large sample size, each class contains 5000 training images and 1000 test images. The entire dataset was used to test the performance of test images according to training images in supervised classification, and it is available at http://www.cs.toronto.edu/~kriz/cifar.html. We used the training images to test the performance of our unsupervised clustering algorithm and compared it to existing clustering algorithms. We randomly removed 10 of the 3072 variable values in 10% of the color images to create a dataset with missing values. Table 6 lists the clustering results from five clustering algorithms: the above four model-based clustering algorithms (M-1, M-2, M-3, and M-4) and the FCM clustering method. We found that our missing-value imputation algorithm had a lower total MR than algorithms M-2 and M-4. The accuracy of our method closely approximated the accuracy obtained using the clustering algorithm with complete data (M-1) but was lower than that of the FCM method. (It should be noted that most clusters of the CIFAR-10 data have deviants from the normal distribution.) Thus, the performance of our model-based algorithm does not offer clustering that is more accurate than the FCM method. Moreover, the speed of computations was much lower in algorithms M-1, M-2, M-3, and M-4.
Table 6

The comparison among four algorithms and FCM method for CIFAR-10 dataset.

MethodMR
totalCluster 1Cluster 2Cluster 3Cluster 4Cluster 5Cluster 6Cluster 7Cluster 8Cluster 9Cluster 10
M-139.9%38.2%41.4%35.6%43.5%40.5%37.9%38.5%42.1%35.4%45.6%
M-242.4%42.6%40.8%35.2%45.8%42.0%46.3%40.4%46.7%38.4%45.4%
M-339.8%42.0%41.0%36.2%39.6%39.7%38.9%38.0%43.2%34.2%45.0%
M-444.0%45.3%43.6%37.4%45.9%43.7%48.5%42.8%47.9%41.3%43.9%
FCM35.9%35.8%35.6%35.5%36.7%34.1%37.6%32.6%37.6%35.7%37.8%

M-1 indicates the complete data clustering algorithm; M-2 indicates missing data deleted clustering algorithm; M-3 indicates our missing-value imputation clustering algorithm; M-4 indicates clustering algorithm for missing-value imputation by mean replacement; FCM indicates Fuzzy C-Means clustering.

M-1 indicates the complete data clustering algorithm; M-2 indicates missing data deleted clustering algorithm; M-3 indicates our missing-value imputation clustering algorithm; M-4 indicates clustering algorithm for missing-value imputation by mean replacement; FCM indicates Fuzzy C-Means clustering.

Discussion

Missing values are unavoidable in experimental data and data from field investigations [31]. Many imputation methods [16-20] have been proposed, such as the mean-substitution method, the regression imputation method, the EM estimation method, and the FCM method. Clustering is a widely used statistical method, especially in gene expression profiles. Most classical clustering methods, such as K-means clustering and hierarchical clustering, are unable to deal with missing values. Therefore, we developed a model-based clustering algorithm for imputing missing values. The proposed method utilizes information from existing data, and it is based on the Gaussian mixture model. With our imputation algorithm, missing values are estimated by calculating the conditional expectation, from which a “pseudo-complete” dataset can be generated and used for clustering. The performance of our proposed algorithm was tested and compared with other algorithms using both real and simulated data. We found that our proposed method outperformed other algorithms. In particular, our proposed method produced clustering results comparable to an algorithm that uses a complete dataset, and it outperformed an algorithm in which samples with missing data are removed and another algorithm in which missing values are imputed by mean replacement. The results are consistent with Hunt and Jorgensen’s reports [32]. Even with datasets in which the boundaries of the clusters overlap, such as datasets A3B3, A2B4, and A3B4, our algorithm performed well. Furthermore, we found that the proposed algorithm outperformed the FCM clustering method on Fisher’s Iris dataset and on the Yeast Cell-cycle dataset. The FCM clustering method, a heuristic approach, has been verified as a successful missing-value clustering method [33]. FCM clusters data based on the characteristics of the samples. The criteria for FCM clustering are inconsistent, however, because the characteristics of dynamic algorithms lead to uncertainties about the samples, degrading FCM’s ability to precisely cluster data [34-36]. Our study shows that the proposed method is more likely to result in a low MR when the dataset is exactly a multivariable Gaussian mixture distribution. That is, it is more accurate than other methods with both simulated data and real data, e.g., Fisher’s Iris dataset and the Yeast Cell-cycle dataset. One possible reason for this is that our method estimates missing values with a conditional expectation based on the Gaussian mixture distribution and generates a “pseudo-complete” dataset. In addition, our method combines dynamic clustering with discriminant analysis and uses the EM algorithm to achieve a maximum-likelihood function. It can thus better utilize the information from known data to discover the intrinsic and implicit properties of a dataset. Usually, model-based methods are better than heuristic approaches, even though the former are always more time consuming. Model-based methods depend on a probability model, which is usually proposed based on the experience and feasibility of the model. The convergence and accuracy of the EM algorithm have already been established [25]. The Gaussian mixture model is both convenient and robust. Moreover, a model-based method provides substantial information—e.g., the posterior probabilities of individuals and the parameters of each cluster—and this information can be exploited to more accurately impute missing values with iterative updates to the algorithm until convergence to a minimizer is reach. Our method uses the Bayesian posterior probability to cluster individuals; individuals are assigned to the cluster having the highest posterior probability. It is more reasonable to assign individuals to a cluster having a high posterior probability, such as 0.8 or 0.9, and this improves the clustering accuracy. Whereas traditional dynamic clustering methods cluster individuals directly to a certain category, this process is difficult when the boundaries of the clusters overlap. Another advantage to a model-based approach is that statistical criteria, such as AIC and BIC [14], can be used to discover the number of clusters that best suit the data. We confirmed the effectiveness of such criteria with results indicating three clusters for Fisher’s Iris dataset (a dataset with exactly three species of Iris) and five clusters for the Yeast Cell-cycle Gene-expression dataset. However, the reversible-jump Markov-chain Monte Carlo (MCMC) method was previously developed to infer the number of distributions in a mixture [37]. MCMC estimates the optimal number of clusters from a Bayesian perspective. Our future research shall compare MCMC with the BIC criteria used in this study. Finally, a prerequisite for raw data collected by experimenters is that they be normalized. Indeed, our method is more sensitive to data that have not been normalized. Therefore, either a log or some other form of data transformation should be applied in order to distribute the data in a normal fashion. Furthermore, because our proposal relies on the information from known individuals when imputing missing values, a large sample size is desirable for more accuracy when estimating missing values.

Conclusions

In this paper, we have presented a model-based missing-data imputation algorithm for clustering numerical data with missing values. We achieved a high convergence rate with our algorithm when it was tested on simulated and real datasets. In addition, the statistical power of our algorithm is high, and it can accurately estimate parameters and correctly cluster individuals with missing variables. Moreover, the results of our evaluation show a low MR, even when the mean vectors of two clusters are close to each other or when their variance-covariance is disperse. The algorithm was applied successfully to datasets (both simulated and real) having missing data. Its performance is superior to that of an algorithm in which samples with missing data were removed and another algorithm in which missing values were imputed by mean replacement. A comparative evaluation demonstrated that our proposed algorithm outperforms the FCM method when the data from each cluster fit a multivariate Gaussian distribution.
Table 2

Average parameter estimates under 4 different simulation datasets A2B1-A2B4 in 20 replicates.

TreatmentIterative timeLikelihood valueProbability estimateMean vector estimateCovariance matrix estimateMR (%)
P^1P^2μ^1 ± seμ^2 ± seS^1 ± seS^2 ± se
A2B1M-181-1903.640.500.50-0.01±0.040.00±0.032.00±0.122.00±0.130.25±0.050.15±0.040.24±0.040.14±0.041.64a
0.15±0.040.25±0.050.14±0.040.24±0.05
M-2117-1696.530.510.490.01±0.04-0.01±0.042.00±0.141.99±0.130.24±0.050.15±0.030.25±0.050.14±0.041.86a
0.15±0.030.26±0.040.14±0.040.24±0.06
M-392-1885.630.500.50-0.01±0.040.01±0.041.99±0.141.99±0.120.23±0.050.14±0.040.23±0.060.15±0.041.60a
0.14±0.040.26±0.040.15±0.040.24±0.05
M-4126-2008.470.480.52-0.02±0.070.01±0.061.95±0.191.96±0.200.22±0.100.12±0.070.21±0.100.11±0.052.26a
0.12±0.070.27±0.120.11±0.050.27±0.07
A2B2M-1120-2372.140.500.50-0.02±0.07-0.01±0.062.01±0.161.97±0.170.45±0.130.27±0.090.46±0.140.28±0.105.60a
0.27±0.090.52±0.110.28±0.100.50±0.14
M-2179-2128.490.510.49-0.02±0.080.02±0.061.95±0.181.96±0.180.48±0.110.27±0.080.47±0.140.25±0.096.30ab
0.27±0.080.53±0.130.25±0.090.45±0.13
M-3135-2462.730.510.490.01±0.070.02±0.061.95±0.191.97±0.170.46±0.110.28±0.080.44±0.090.26±0.095.40a
0.28±0.080.53±0.100.26±0.090.46±0.08
M-4266-2410.500.450.55-0.03±0.100.02±0.111.91±0.281.92±0.290.44±0.170.25±0.100.44±0.180.24±0.119.00 b
0.25±0.100.54±0.160.24±0.110.43±0.19
A2B3M-1219-2699.350.480.52-0.05±0.08-0.03±0.081.92±0.271.94±0.240.67±0.200.40±0.100.75±0.190.42±0.1110.90a
0.40±0.100.78±0.210.42±0.110.78±0.19
M-2387-2489.380.470.53-0.07±0.090.05±0.101.90±0.271.90±0.280.64±0.250.38±0.130.72±0.220.41±0.1211.40a
0.38±0.130.73±0.240.41±0.120.79±0.21
M-3242-2709.340.470.53-0.07±0.10-0.05±0.101.90±0.281.90±0.300.66±0.180.38±0.090.76±0.200.47±0.1111.00a
0.42±0.090.77±0.180.47±0.110.75±0.20
M-4450-2646.830.420.58-0.14±0.200.12±0.181.89±0.411.88±0.450.64±0.290.38±0.170.72±0.270.40±0.1516.96b
0.38±0.170.78±0.270.40±0.150.77±0.23
A2B4M-1360-3014.360.450.55-0.11±0.160.09±0.171.85±0.341.80±0.320.86±0.260.52±0.181.01±0.230.63±0.1712.80a
0.52±0.181.04±0.250.63±0.171.05±0.23
M-2486-2677.250.430.57-0.15±0.25-0.15±0.251.84±0.351.82±0.320.82±0.300.51±0.191.04±0.290.61±0.1818.00b
0.51±0.190.94±0.250.61±0.181.01±0.29
M-3342-2907.180.440.56-0.10±0.170.08±0.141.86±0.341.82±0.320.87±0.270.53±0.180.92±0.280.62±0.1712.30a
0.53±0.181.03±0.250.62±0.171.03±0.29
M-4440-2977.460.400.60-0.20±0.350.19±0.331.82±0.571.81±0.620.76±0.410.51±0.271.05±0.330.65±0.2721.80c
0.51±0.270.90±0.390.65±0.271.07±0.34

M-1 indicates a complete data clustering algorithm; M-2 indicates a missing-data-deleted clustering algorithm; M-3 indicates our missing-value imputation clustering algorithm

M-4 indicates the clustering algorithm for missing-value imputation by mean replacement. These four algorithms are based on the multivariate Gaussian mixture model.

indicates the multiple comparisons of the differences in MR among the algorithms M-1, M-2, M-3, and M-4: having the different letters indicate that there is statistical significance between these two groups (P<0.05), and vice versa.

  11 in total

1.  A hierarchical unsupervised growing neural network for clustering gene expression patterns.

Authors:  J Herrero; A Valencia; J Dopazo
Journal:  Bioinformatics       Date:  2001-02       Impact factor: 6.937

2.  Model-based clustering and data transformations for gene expression data.

Authors:  K Y Yeung; C Fraley; A Murua; A E Raftery; W L Ruzzo
Journal:  Bioinformatics       Date:  2001-10       Impact factor: 6.937

3.  Missing value estimation methods for DNA microarrays.

Authors:  O Troyanskaya; M Cantor; G Sherlock; P Brown; T Hastie; R Tibshirani; D Botstein; R B Altman
Journal:  Bioinformatics       Date:  2001-06       Impact factor: 6.937

Review 4.  Computational analysis of microarray data.

Authors:  J Quackenbush
Journal:  Nat Rev Genet       Date:  2001-06       Impact factor: 53.242

5.  Supervised cluster analysis for microarray data based on multivariate Gaussian mixture.

Authors:  Yi Qu; Shizhong Xu
Journal:  Bioinformatics       Date:  2004-03-25       Impact factor: 6.937

6.  Fuzzy c-means clustering of incomplete data.

Authors:  R J Hathaway; J C Bezdek
Journal:  IEEE Trans Syst Man Cybern B Cybern       Date:  2001

7.  Model-based clustering for RNA-seq data.

Authors:  Yaqing Si; Peng Liu; Pinghua Li; Thomas P Brutnell
Journal:  Bioinformatics       Date:  2013-11-04       Impact factor: 6.937

8.  Cluster analysis and display of genome-wide expression patterns.

Authors:  M B Eisen; P T Spellman; P O Brown; D Botstein
Journal:  Proc Natl Acad Sci U S A       Date:  1998-12-08       Impact factor: 11.205

9.  Unsupervised performance evaluation strategy for bridge superstructure based on fuzzy clustering and field data.

Authors:  Yubo Jiao; Hanbing Liu; Peng Zhang; Xianqiang Wang; Haibin Wei
Journal:  ScientificWorldJournal       Date:  2013-10-29

10.  A model-based clustering method for genomic structural variant prediction and genotyping using paired-end sequencing data.

Authors:  Matthew Hayes; Yoon Soo Pyon; Jing Li
Journal:  PLoS One       Date:  2012-12-27       Impact factor: 3.240

View more
  2 in total

1.  A Bayesian Failure Prediction Network Based on Text Sequence Mining and Clustering.

Authors:  Wenbing Chang; Zhenzhong Xu; Meng You; Shenghan Zhou; Yiyong Xiao; Yang Cheng
Journal:  Entropy (Basel)       Date:  2018-12-03       Impact factor: 2.524

2.  Using Machine Learning to Identify Biomarkers Affecting Fat Deposition in Pigs by Integrating Multisource Transcriptome Information.

Authors:  Huatao Liu; Kai Xing; Yifan Jiang; Yibing Liu; Chuduan Wang; Xiangdong Ding
Journal:  J Agric Food Chem       Date:  2022-08-11       Impact factor: 5.895

  2 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.