Literature DB >> 28045950

A Rough Set Bounded Spatially Constrained Asymmetric Gaussian Mixture Model for Image Segmentation.

Zexuan Ji1, Yubo Huang1, Quansen Sun1, Guo Cao1, Yuhui Zheng2.   

Abstract

Accurate image segmentation is an important issue in image processing, where Gaussian mixture models play an important part and have been proven effective. However, most Gaussian mixture model (GMM) based methods suffer from one or more limitations, such as limited noise robustness, over-smoothness for segmentations, and lack of flexibility to fit data. In order to address these issues, in this paper, we propose a rough set bounded asymmetric Gaussian mixture model with spatial constraint for image segmentation. First, based on our previous work where each cluster is characterized by three automatically determined rough-fuzzy regions, we partition the target image into three rough regions with two adaptively computed thresholds. Second, a new bounded indicator function is proposed to determine the bounded support regions of the observed data. The bounded indicator and posterior probability of a pixel that belongs to each sub-region is estimated with respect to the rough region where the pixel lies. Third, to further reduce over-smoothness for segmentations, two novel prior factors are proposed that incorporate the spatial information among neighborhood pixels, which are constructed based on the prior and posterior probabilities of the within- and between-clusters, and considers the spatial direction. We compare our algorithm to state-of-the-art segmentation approaches in both synthetic and real images to demonstrate the superior performance of the proposed algorithm.

Entities:  

Mesh:

Year:  2017        PMID: 28045950      PMCID: PMC5207730          DOI: 10.1371/journal.pone.0168449

Source DB:  PubMed          Journal:  PLoS One        ISSN: 1932-6203            Impact factor:   3.240


Introduction

As one of the classical problems in image processing, image segmentation has been extensively studied, which can be treated as a classification problem [1-5] for the target image. Various image segmentation algorithms have been developed such as active contour models [6, 7], graph based methods [8, 9] and clustering techniques [10-12]. Over the last decades, model-based techniques [13, 14] have been widely used in image segmentation, where the standard Gaussian mixture model (GMM) [15, 16] is a well-known method because of its simplicity and ease of implementation [17]. The parameters involved in GMM can be efficiently estimated by expectation maximization (EM) algorithm [18]. However, the standard GMM still suffers from the following limitations: sensitivity to noise, less flexibility to fit the shape of the data and unbounded distributions [19]. In order to reduce noise sensitivity for segmentation, (hidden) Markov random fields ((H)MRF) based mixture models have been widely utilized for pixel labels [20-22], where (H)MRF is acted on neighboring labels, and the clustering result for each pixel depends on the neighboring pixels [23]. On the other hand, in order to impose spatial constraints among neighboring pixels, another group of mixture models with MRF has been proposed by modeling the joint distribution of the priors for each pixel [24-27]. For example, in [26], Displaros et al. proposed a generative GMM model by introducing a pseudo-likelihood quantity to incorporate the spatial smoothness constraints based on Kullback-Leibler (KL) divergence. In [27], Nikou et al. proposed a novel spatial constraint that can adaptively select spatial directions. In order to directly apply the EM algorithm to estimate the involved parameters, Nguyen and Wu [23] proposed a robust spatially constrained GMM by introducing a spatial factor into the prior distribution. Although the forementioned algorithms can reduce the impact of noise in the image, most (H)MRF based algorithms are still not sufficiently robust with respect to different noise types and levels. Because of the utilization of Gaussian distribution in GMM, the distribution tail is often shorter for many applied problems [17], which means that the Gaussian distribution is not sufficiently flexible to fit data [19]. In order to improve the flexibility for the data fitness, the Student’s-t distribution, Laplace distribution and generalized Gaussian distribution are used to replace the Gaussian distribution in mixture model. Therefore, the Student’s-t mixture model (SMM) [28, 29], Laplace mixture model (LMM) [30, 31], and generalized Gaussian mixture model (GGMM) [32, 33] have been proposed. On the other hand, using only one distribution for each component in the mixture model is not sufficiently satisfactory for many practical applications. Therefore, another solution for fitting data with different shapes is using multiple distributions for each component. For example, Zhang et al. [34] proposed a modified GMM by incorporating local spatial and intensity information. The conditional probability for each pixel is constructed based on the probabilities of neighboring pixels [34]. Browne et al. [35] proposed a mixture of mixture model by combining a multivariate Gaussian distribution and a multivariate uniform distribution together to model the component density [35]. Nguyen et al. [19] proposed an asymmetric mixture model by modeling the component with multivariate Gaussian distributions [19]. Moreover, the distributions in most mixture models are unbounded with a supporting range of (−∞, +∞), which is not consistent with the practical application where the practical data generally fall in a bounded region [19]. In [36], a bounded GMM (BGMM) was proposed for speech processing. In [37], a bounded generalized GMM was proposed that included GMM, LMM, GGMM, and BGMM as special cases. Nguyen et al. [17, 19, 38] proposed various bounded mixture models to fit different data shapes. However, the above mentioned approaches still suffer from the following limitations: (1) Without considering any spatial information, the mixture of mixture model [35] and the bounded asymmetric mixture model (BAMM) [19] are still sensitive to noise, although both of these types of models are more flexible. (2) For the bounded mixture models [19, 36–38], the bounded support regions of observed data should be predefined. Moreover, the indicator function of the bounded support region is a binary function that cannot easily manage uncertainty, vagueness, and incompleteness in data. Motivated by the aforementioned observations, in this paper, we propose a rough set bounded asymmetric Gaussian mixture model with spatial constraint for image segmentation. First, in our previous work [39], based on the rough set theory [40], we proposed a generalized rough fuzzy c-means (GRFCM) algorithm, where, for each cluster, an image is automatically partitioned into three rough regions with two adaptively computed thresholds. In this paper, we utilize these two thresholds to partition the target image into three rough regions, i.e., the positive, boundary and negative regions [41]. Second, a new bounded indicator function is proposed to determine the bounded support regions of the observed data. The bounded indicator of a pixel that belongs to each sub-region is estimated with respect to the rough region where the pixel lies. Only those pixels in the positive and boundary regions have non-zero indicators. Therefore, because of the benefits of rough set theory, the proposed bounded indicator function can further manage uncertainty in data. Third, to further overcome the impact of noise and reduce over-smoothness for segmentations, two novel prior factors are proposed to introduce the spatial information. The proposed prior factors can be treated as the within- and between-cluster spatial constraints with spatial direction. Finally, to further improve the robustness of the model, for each component, the posterior probabilities of within- and between-cluster for each pixel are estimated with respect to the rough regions. The proposed algorithm is compared to several state-of-the-art segmentation algorithms on simulated and real images to demonstrate its superior performance.

Finite Mixture Model

The notations used throughout this paper are as follows. The target image is denoted as X = {x, i = 1, 2, …, N}, where x with dimension D is the intensity values for the ith pixel. The neighborhood of the ith pixel is denoted as ∂, and the labels are denoted as (Ω1, Ω2, …, Ω). In order to segment an image with N pixels into K labels, the density function of the finite mixture model [42] is given by: where Π = {π}, i = {1, 2, …, N}, k = {1, 2, …, K} are the prior probabilities, and satisfy the constraints 0 ≤ π ≤ 1 and . In GMM [15, 16], the component p(x|Ω) is the Gaussian distribution Φ(x|μ, Σ) that can be written in the form: where μ is the mean vector with D dimension, Σ is the covariance matrix with D × D dimension, and |Σ| is the determinant of Σ. In order to address the issue that the observed data generally fall within the bounded support regions in practical applications, in [19, 36, 37], the bounded support region in ℜ is defined as ∂Ω for each label Ω, and the indicator function can be written as With the above indicator function H(x|Ω) and distribution p(x|Ω), a bounded distribution can be defined as For additional analysis details, please refer to [19, 36] and [37]. However, the major disadvantage of indicator function H(x|Ω) is that it is a binary function that cannot easily manage uncertainty in data. To improve the noise robustness, the spatial information is generally incorporated through MRF distribution: where U(Π) is the smoothing prior, and Z and T are two constants. Based on Bayes’rules, the probability density function can be written as: Most MRF-based mixture models have been successfully applied to image segmentation by adopting different energy functions U(Π). Nguyen and Wu [23] pointed out that the M-step of EM cannot be applied directly to the prior distribution π due to the complexity of the log-likelihood function. Thus, the resulting algorithms are computationally complex and have to utilize large amounts of computational power to solve the constrained optimization problem of the prior distribution π [23]. To overcome these disadvantages, they introduced a novel factor G by defining a multiplication of both posterior probability and prior distributions as follows. where z is the posterior probability and β is the balance parameter to control the smoothing prior. The main advantage of G is the ease of implementation and incorporation of the spatial relationships amongst neighborhood pixels in a simpler metric. Then the smoothing prior U(Π) is given by: However, the energy U(Π) can cause over-smoothing for segmentation and loss of details, especially for regions with abundant textures.

Proposed Model

In order to fit different data shapes, Nguyen et al [19] defined a new distribution p(x|Ω) to model the component density. Motivated by the bounded asymmetric distribution, in this paper we modify distribution p(x|Ω) to allow the model to easily incorporate the spatial information, which can be defined as: where L is the number of bounded multivariate Gaussian distribution Ψ(x|μ, Σ), and η is the weighting factor and satisfies the constraints 0 ≤ η ≤ 1 and . The bounded Gaussian distribution Ψ(x|μ, Σ) is defined as: where Φ(x|μ, Σ) is the Gaussian distribution defined in Eq (2) and H(x|Ω) is the indicator function for the bounded support region defined in Eq (3). ∫∂ Φ(x|μ, Σ)dx is the normalization constant. Therefore, in this paper, we propose a rough set bounded asymmetric Gaussian mixture model with spatial constraint for image segmentation. First, based on the rough set theory, we utilize our previous work [39] to partition the target image into three rough regions with two adaptively computed thresholds. Second, a new bounded indicator function is proposed to determine the bounded support regions of the observed data based on the above rough regions. Third, to further overcome the impact of noise and reduce over-smoothness for segmentations, two novel prior factors with spatial direction are constructed based on the prior and posterior probabilities of the within- and between-clusters. Finally, to further improve model robustness, for each component, the posterior probability is re-estimated based on the adaptively determined rough regions.

Determination of rough set region

Rough set theory can manage the uncertainty with lower and upper approximations. Specifically, let U ≠ ∅ be a universe of discourse, and R be an equivalence relationship that can lead to a U partition. By denoting U/R = {X1, X2, …, X}, where X is an equivalence class for R, the lower and upper approximations of subset X are defined as: The lower approximation is a set where all categories certainly belong to X, and the upper approximation is a set where all categories possibly belong to X. Based on the approximations, three rough regions of X, i.e., the R-positive Po region, R-negative Ne region, and R-boundary Bo region, can be defined as follows [41]: Fig 1 illustrates the definition of the three rough regions. The figure shows that positive region Po equals the lower approximation of Ω, whereas negative region Ne equals the complement of the upper approximation of Ω.
Fig 1

Illustration of three rough regions.

In our previous work [39], based on the distance between each pixel value x and intensity level g, two thresholds are adaptively computed to determine the rough regions: where ∂ is the neighborhood of pixel i with n pixels. In this paper, the size of neighborhood ∂ is set as 3 × 3. J and J are the maximum and minimum intensity values of the image, respectively, and J is the number of intensity levels in the image. Thus, we construct distance vector d = {d(g1), d(g2), …, d(g)} for each pixel i. The two thresholds can be estimated based on mean value d and minimum value d of vector d: Based on the distance between each pixel i and the mean value of class k using Eq (13), we can determine the rough regions for each cluster Ω as follows: For more analysis and discussion details, please refer to [39].

Construction of bounded support region

As mentioned before, the bounded support regions of observed data were predefined in [19, 36–38], and the indicator function of the bounded support region is a binary function that cannot easily manage uncertainty in data. Motivated by the aforementioned observations, in this paper, we propose a new bounded indicator function based on the rough regions. For each label Ω, the bounded support region in ℜ is defined as ∂Ω, and the new indicator function can be written as: With the new indicator function and distribution p(x|Ω) in Eq (2), a bounded multivariate Gaussian distribution can be defined as: where is the normalization constant, and it is identified as the share of Φ(x|μ, Σ) that belongs to support region ∂Ω [19]. Similar to BAMM [19], each component density in our model is constructed with multiple bounded asymmetric distribution. The corresponding distribution p(x|Ω) is defined as: where L is the number of bounded multivariate Gaussian distribution , and η is the weighting factor that satisfies the constraints 0 ≤ η ≤ 1 and . It should be noted that the above distribution always satisfies the conditions of the probability density [14]: Therefore, the log-likelihood function of the proposed model can be written as In order to maximize the above likelihood function, two variables z and y are defined as follows: The values z and y always satisfy the conditions and , respectively. It is worth mentioning that, based on Bayes’rules, we can treat both variables z and y as the posterior probability. Variable z indicates the relationship between pixels and clusters that can be treated as the between-cluster relationship. Meanwhile, variable y indicates the relationship between pixels and distribution components that can be treated as the within-cluster relationship because each cluster is modeled with multiple distributions. Consequently, variable z can be treated as the posterior probabilities of the between-cluster, whereas variable y can be treated as the posterior probabilities of the within-cluster. To further improve model robustness, for each component, the posterior probability of each pixel is re-estimated with respect to the rough region where the pixel lies. Thus, the new hidden variables and can be rewritten as:

Construction of prior factor

As we mentioned before, the prior factor G in [23] plays a role as an average filter on both posterior probability and prior distributions for smoothing noisy images, which may cause over smoothing for the segmentation and lose the details especially for the regions with abundant textures. The smoothing prior in [27] can adaptively select spatial directions, which introduces additional training complexity. To reduce the complexity of the smoothing prior and preserve more details for the segmentations, we propose two novel prior factors E and F that consider spatial direction based on the prior and posterior probabilities of the between- and within-cluster in this paper. Both prior factors are defined as: where is the neighborhood of pixel i at direction for cluster k, which contains pixels, and is given by: where dist(x, μ) is the Euclidean distance between point i and cluster center μ, and is the neighborhood of pixel i at direction s. In this paper, we set S = 4, i.e., four directions (horizontal, vertical and two diagonal) are considered. More explicitly, taking 3 × 3 neighborhood pixels as example, the filters for each neighboring direction can be constructed as shown in Fig 2. During the algorithm procedure, we only need to operate the convolution over the prior and posterior probabilities with these four predefined filters, and then adaptively select the satisfactory direction based on the differences among the intensity values along each direction. Therefore, the proposed prior factors can efficiently preserve more details.
Fig 2

An example of spatial filters that considers four directions.

From left to right: horizontal, vertical and two diagonal directions.

An example of spatial filters that considers four directions.

From left to right: horizontal, vertical and two diagonal directions. Then, we incorporate the proposed prior factors into the smoothing prior: MRF distribution p(Π) can be given by: where Z and T are set as 1. Therefore, the log-likelihood function of the proposed algorithm can be written in the form: Finally, we can optimize the involved parameters by maximizing the above log-likelihood function.

Parameter estimation

In order to determine label Ω for each pixel x, we need to estimate parameters Π = {π, η} and Θ = {μ, Σ} by maximizing log-likelihood function L(Π, Θ|X). By monotonically increasing the logarithm character, maximizing log-likelihood function L(Π, Θ|X) can lead to minimizing objective function J(Π, Θ|X) as follows: By applying the complete data condition in [24], minimizing the negative log-likelihood function in Eq (31) can also lead to minimizing the objective function E(Π, Θ|X) as follows: Then we can apply the EM algorithm to minimize Eq (32) by considering the derivation of function E(Π, Θ|X) with respect to each variable. Finally, we can obtain the following updating function for each variable. Please refer to the Appendix for a detailed derivation of the proposed algorithm. Mean value estimation: Covariance matrix estimation: Prior probability estimation: Consequently, the proposed algorithm for image segmentation is summarized as follows. Initialize parameters Π = {π, η} and Θ = {μ, Σ} with the K-means algorithm. Determine the rough regions for each cluster with Eqs (13)–(15). E-step Update posterior probabilities and with Eqs (21)–(24). M-step Update prior factors E and F with Eqs (25) and (26). Update means μ with Eq (33). Update covariance values Σ with Eq (34). Update prior probabilities π and η with Eqs (35) and (36). Check the convergence. Stop the iteration if the convergence criterion is satisfied; otherwise, go to Step 2. It should be noted that the convergence criterion is generally the distance between the values of objective functions or variables (i.e. means or covariance values) from two successive iterations. In this paper, we utilize the total distance between the mean values obtained from two successive iterations. When this distance becomes smaller than a user specified threshold, which was set to 10−5 in all algorithms for this study, we think the algorithm converges and stop the iteration.

Experimental Results

In this paper, we compare the proposed algorithm with four algorithms, i.e., a Bayesian bounded asymmetric mixture model (BAMM) [19], a bounded generalized Gaussian mixture model (BGGMM) [37], a spatially constrained generative model and EM algorithm (SCGM-EM) [26], and a fast and robust spatially constrained GMM (FRSCGMM) [23]. Unless otherwise specified, the parameters of the proposed algorithm are set as follows: The window size for prior factor construction is 3 × 3. The number of bounded multivariate Gaussian distribution is L = 3. A summary of the parameter settings for each comparison algorithm is listed in Table 1. Please see the corresponding references for more details. It should be noted that for fair comparison, all algorithms, including the proposed model, use the same initializations generated by the k-means algorithm for each testing image. All algorithms were implemented using MATLAB 7.8 platform and tested on a PC (Intel Core i7-4790 CPU, 3.60GHz, 16GB RAM, and 64-bit Windows 8).
Table 1

Summary of parameter setting for each comparison algorithm in experiment.

AlgorithmsParameter setting
SCGM-EMTemperature value β = 0.5
Size of neighborhood 5 × 5
FRSCGMMTemperature value β = 12
Size of neighborhood 5 × 5
BAMMThe number of Gaussian distribution Kj = 3
The number of random variables M = 106
The bounded support regions ∂Ωj ∈ (0, 255)
BGGMMThe number of random variables M = 106
The bounded support regions ∂Ωj ∈ (0, 255)
The algorithms are compared using synthetic, synthetic and real brain MR, and color images. Fig 3 shows an example of each type of testing image.
Fig 3

Examples of testing images.

From left to right: synthetic image, simulated T1-weighted brain MR, real T1-weighted brain MR and natural images.

Examples of testing images.

From left to right: synthetic image, simulated T1-weighted brain MR, real T1-weighted brain MR and natural images. (1) The synthetic images used in this paper, of size 128 × 128, contain different labels with luminance values. For example, the image shown at the left of Fig 3 contains four labels with luminance values [0, 1/3, 2/3, 1]. Then we add to the synthetic images different types of noise with different levels for each comparison experiment. The segmentation results for synthetic images are quantitatively evaluated by correct classification ratio (CCR) [27], which is defined as: where gt is the ground truth for cluster k, seg describes the pixels classified by the algorithm to cluster k and . CCR ranges from 0 to 1 with a higher value representing a better segmentation result. (2) The brain MR images are selected from two open sources: the BrainWeb (http://www.bic.mni.mcgill.ca/brainweb) [43] and IBSRv2.0 (https://www.nitrc.org/projects/ibsr) [44] databases. The objective for brain MR image segmentation is to partition the image into three tissue labels: gray matter (GM), white matter (WM), and cerebrospinal fluid (CSF). The Dice coefficient (DC) [45] is utilized to quantitatively evaluate the performance for segmenting each type of brain tissue, which is the ratio between the intersection and union of segmented volume S1 and ground truth volume S2 The DC value ranges from 0 to 1, with a higher value representing a more accurate segmentation result. (3) The natural images are selected from the open source Berkeley Segmentation Dataset 500 (BSDS500, https://www2.eecs.berkeley.edu/Research/Projects/CS/vision/bsds/) [46]. The probabilistic rand index (PRI) [47] is used to assess the segmentation performances on natural images. PRI between segmentation map S to be evaluated and a set of M ground truth images S = {S1, …, S} is given by: where c = 1 if pixels i and j belong to the same cluster, and c = 0 if pixels i and j belong to different clusters. PRI takes values between 0 and 1, with a higher value representing a more accurate segmentation result.

Illustration of proposed algorithm

In order to show the advantages and performance of the proposed algorithm in detail, we first segmented a brain MR image that is the 90-th slice of the 3D brain MR image with 9% noise selected from BrainWeb. The component number for the mixture model is K = 3. Meanwhile, we set the number of bounded multivariate Gaussian distribution as L = 3. The original noisy brain MR image, intermediate illustrations of the proposed algorithm, and final output segmentation are shown in Fig 4. Three rough regions for each cluster, i.e., CSF (k = 1), GM (k = 2), and WM (k = 3) are shown in the first part of the intermediate illustrations, where the positive, boundary, and negative regions are illustrated in bright, gray, and dark respectively. This reveals that the proposed algorithm can construct rough regions appropriately. The corresponding posterior probability of the between and within-cluster, i.e., and , are depicted in the second and third parts of the intermediate illustrations, respectively. From the nine images of variables , we find that variables provide different types of segmentation results for each cluster/tissue, including over and insufficient segmentations, which can further improve the accuracy of the posterior probability estimation by combining with the automatically computed prior probabilities (weights for each distribution). Comparing the last two images on the left of Fig 4, we find that the proposed algorithm can effectively overcome the impact of noise, and the corresponding segmentation result is consistent with the ground truth.
Fig 4

Illustrations of proposed algorithm.

Moreover, to further explain the limitations of GMM and the advantages of the proposed distributions, two examples with image histogram and corresponding estimated distributions are illustrated in Figs 5 and 6. Because BAMM [19] has already proven its superior performance over GMM, SMM, and GGMM, in this section, we only present the estimated distributions obtained by employing ϕ(x|μ, Σ) (i.e., the Gaussian distribution in the GMM model, Eq (2)), Ψ(x|μ, Σ) (i.e., the bounded Gaussian distribution in BAMM model, Eq (10)), and (i.e. the proposed distribution, Eq (17)).
Fig 5

Illustrations of estimated distributions on synthetic image.

Fig 6

Illustrations of estimated distributions on natural image.

In Fig 5, the noisy image with size 128 × 128 shown in Fig 5(a) is used to compare the performance of the proposed algorithm with GMM and BAMM. This image contains three labels (K = 3). The number of bounded multivariate Gaussian distributions for BAMM and the proposed algorithm is set as L = 3. The bounded region for BAMM is set as [0, 255] for each cluster. The ground truth of this image is shown in Fig 5(b), which presents the estimated density function obtained by GMM, BAMM, and the proposed algorithm based on the histogram of the observed data. Fig 5(d)–5(f) show the ground truth distributions for each class along with the estimated distributions of each corresponding cluster. From the histogram of observed data and corresponding ground truth distributions, we find that the image data are non-Gaussian, non-symmetric and bounded support data. Without the bounded constraint, GMM performance is not sufficiently satisfactory because the estimated density function cannot fit well with the histogram of observed data. Compared with GMM, both BAMM and the proposed algorithm better fit the observed data because of the introduction of the bounded multivariate Gaussian distribution. It can also be visualized that distribution in Eq (17) allows flexibility for a better fit of the observed data compared with distribution Ψ(x|μ, Σ) of BAMM in Eq (10). Similar to Figs 5 and 6 shows a gray natural image of size 481 × 321 that is used to compare the performance of the proposed algorithm with GMM and BAMM. This image contains three labels (K = 3). In comparison, the estimated distribution of the proposed algorithm is much better than that of the others.

Segmentation of synthetic images

In the second experiment, a synthetic image (image size: 128 × 128) as shown in Fig 7(a), is used to compare the performances among different algorithms. The image contains four labels with luminance values [0, 1/3, 2/3, 1]. The noisy image with Gaussian noise (0 means, 0.07 variance) is shown in Fig 7(b). From Fig 7(c)–7(g), we present the segmentation results obtained by employing the proposed algorithm, SCGM-EM, FRSCGMM, BAMM and BGGMM, respectively. Without considering any spatial information, the segmentation accuracies of BAMM and BGGMM are quite poor. The anti-noise ability for FRSCGMM is limited. Both SCGM-EM and the proposed algorithm obtain better performances. Nevertheless, the proposed method obtains higher CCR, especially for the pixels around the boundaries.
Fig 7

Experimental results on synthetic image with Gaussian noise (image size: 128 × 128).

(a) Original image, (b) Noisy image with Gaussian noise (0 mean, 0.07 variance); segmentation results by applying (c) proposed algorithm (CCR = 0.9954), (d) SCGM-EM (CCR = 0.9942), (e) FRSCGMM (CCR = 0.9906), (f) BAMM (CCR = 0.9131), (g) BGGMM (CCR = 0.9152).

Experimental results on synthetic image with Gaussian noise (image size: 128 × 128).

(a) Original image, (b) Noisy image with Gaussian noise (0 mean, 0.07 variance); segmentation results by applying (c) proposed algorithm (CCR = 0.9954), (d) SCGM-EM (CCR = 0.9942), (e) FRSCGMM (CCR = 0.9906), (f) BAMM (CCR = 0.9131), (g) BGGMM (CCR = 0.9152). Then, the synthetic image with luminance values [0, 1/4, 2/4, 3/4, 1] shown in Fig 8(a) is utilized to test the performance of different algorithms on different noise type. As shown in Fig 8(b), the original image is corrupted with multiplicative noise (speckle noise with mean 0 and variance 0.04.). Fig 8(c)–8(g) present the segmentation results obtained by employing the proposed algorithm, SCGM-EM, FRSCGMM, BAMM and BGGMM, respectively. Among these methods, the proposed algorithm obtains better preservation of image details.
Fig 8

Experimental results on synthetic image with speckle noise (image size: 128 × 128).

(a) Original image, (b) Noisy image with speckle noise (0 mean, 0.04 variance), segmentation results by applying (c) proposed algorithm (CCR = 0.9956), (d) SCGM-EM (CCR = 0.9802), (e) FRSCGMM (CCR = 0.9835), (f) BAMM (CCR = 0.8174), (g) BGGMM (CCR = 0.7456).

Experimental results on synthetic image with speckle noise (image size: 128 × 128).

(a) Original image, (b) Noisy image with speckle noise (0 mean, 0.04 variance), segmentation results by applying (c) proposed algorithm (CCR = 0.9956), (d) SCGM-EM (CCR = 0.9802), (e) FRSCGMM (CCR = 0.9835), (f) BAMM (CCR = 0.8174), (g) BGGMM (CCR = 0.7456).

Segmentation of brain MR images

Because of the utilization of our previous work [39], in this section, we supplement the generalized rough fuzzy c-means (GRFCM) algorithm [39] as a comparison method. It should be noted that GRFCM can overcome the impact of intensity inhomogeneity in brain MR images, but all other comparison algorithms can only overcome the impact of noise. Therefore, in this experiment, we apply all algorithms to segment the synthetic T1-weighted 1 mm brain MR images selected from BrainWeb, which only contain different noise levels. For a fair comparison, the intensity inhomogeneity estimation part in GRFCM is removed. Three sample brain MR images with 9% noise (80-th axial, sagittal and coronal slice), along with their segmentation results and ground truths, are shown in Fig 9. Similar to the proposed algorithm, both SCGM-EM and FRSCGMM construct spatial information based on the posterior and prior probabilities. However, without considering any directional differences, these two algorithms can obtain over-smoothness segmentations and lose details, especially for the CSF tissue. Meanwhile, both BAMM and BGGMM cannot well distinguish noisy pixels without considering any spatial information. GRFCM is not robust to noise, and the corresponding segmentations are not sufficiently smooth. By comparing the ground truth with the segmentations obtained with all algorithms, we see that the proposed algorithm visually obtains better results.
Fig 9

Illustration of three simulated T1-weighted brain MR images with 9% noise and corresponding segmentation results obtained by each algorithm.

In each subfigure, the images from left to right show: original image, segmentation results obtained by SCGM-EM, FRSCGMM, BAMM, BGGMM, GRFCM, proposed algorithm, and ground truth.

Illustration of three simulated T1-weighted brain MR images with 9% noise and corresponding segmentation results obtained by each algorithm.

In each subfigure, the images from left to right show: original image, segmentation results obtained by SCGM-EM, FRSCGMM, BAMM, BGGMM, GRFCM, proposed algorithm, and ground truth. Segmentation accuracy for each tissue is measured in terms of DC values, and the results are listed in Table 2, which further demonstrates superior performance of the proposed algorithm. It is worth mentioning that the DC values of axial and sagittal CSF segmentation for BAMM and BGGMM are slightly higher than other algorithms because of the low percentage and abundant texture details of the CSF tissue. For any spatially constrained algorithm, the introduced spatial information can improve segmentation accuracy for images occupied by noise, but this can lead to smoothing the texture details in the image. Therefore, it is a dilemma to balance the trade-off between the anti-noise ability and over-smoothness for image textures. By comparing with two other spatially constrained algorithms (SCGM-EM and FRSCGMM), we see that the proposed algorithm achieves better trade-off and balance.
Table 2

DC values of each tissue for the segmentations shown in Fig 9.

AlgorithmsTissueAxialSagittalCoronalAverage
SCGM-EMGM0.88540.86540.91620.8890
WM0.90900.90800.93080.9159
CSF0.87430.90320.90940.8956
FRSCGMMGM0.86920.86620.90750.8810
WM0.88100.90140.91370.8987
CSF0.84170.90330.90160.8822
BAMMGM0.83520.82680.85080.8376
WM0.84960.85220.86000.8539
CSF0.88140.90610.87070.8876
BGGMMGM0.83870.82410.85860.8405
WM0.85240.85540.86600.8579
CSF0.88140.90610.89210.8932
GRFCMGM0.89880.91350.90850.9069
WM0.88770.91400.93030.9107
CSF0.88530.91010.89640.8962
Proposed algorithmGM0.91060.89340.92780.9106
WM0.92610.92350.93690.9288
CSF0.87460.90400.91260.8971
To statistically show the significance of the proposed algorithm, we apply the previous six algorithms to segment 10 axial, 10 sagittal and 10 coronal (from slice 86th to 95th) MR images for each noise level, where the level ranges from 3% to 9%. The statistical results (means and standard deviations of DC values) are shown in Fig 10(a)–10(c). Moreover, the average CCR value over the entire segmentation results is show in Fig 10(d). From Fig 10, we can observe that the proposed algorithm produces more accurate segmentation (higher means) and has better robustness to noise (lower standard deviations). As indicated before, without considering any spatial information, BAMM and BGGMM produce satisfactory results when the noise level in the image is low (3%). SCGM-EM, FRSCGMM, and the proposed algorithm sacrifice segmentation accuracy for image details in orders to achieve better anti-noise ability. Therefore, the DC values of BAMM and BGGMM for the CSF tissue are slightly higher than the other three algorithms. By increasing the noise level, the performances of BAMM and BGGMM might decrease dramatically, especially for GM and WM tissues. Table 3 lists the statistical results (p-value of paired-sample t-test) between those five methods and the proposed algorithm on the testing images utilized in Fig 10. It is observed that, considering 0.05 as the level of significance, the proposed algorithm provides significantly better segmentation results with respect to both DC and CCR on the BrainWeb dataset.
Fig 10

DC values for: (a) GM segmentation, (b) WM segmentation, (c) CSF segmentation, (d) CCR values over the entire images obtained by applying six segmentation algorithms to simulated brain MR images with increasing noise levels.

Table 3

Paired-sample t-test results (p-value) of DC and CCR values between the five comparison methods and the proposed algorithm on all the testing images used in Fig 10.

MethodsSCGM-EMFRSCGMMBAMMBGGMMGRFCM
DC for GM4.5e-052.1e-033.8e-202.7e-181.7e-02
DC for WM1.9e-067.0e-052.8e-281.0e-252.9e-04
DC for CSF4.1e-041.4e-021.7e-162.8e-156.3e-03
CCR4.5e-061.8e-044.3e-313.9e-283.3e-03
DC values for: (a) GM segmentation, (b) WM segmentation, (c) CSF segmentation, (d) CCR values over the entire images obtained by applying six segmentation algorithms to simulated brain MR images with increasing noise levels. In the next experiment, we apply all algorithms on the IBSR v2.0 data set [44], which contains 18 3D images. It is worth mentioning that BAMM [19] has been verified to outperform widely used algorithms, i.e., EMS [45] and SPM [48]. Therefore, in this comparison experiment, we only need to compare BAMM with the proposed algorithm in order to demonstrate the superior performance of the latter. Fig 11 shows a 3D slice view of the real dataset (IBSR04). The image shown in Fig 11(b) is the ground truth of the original image. Fig 11(c)–11(h) show the results obtained by implementing the proposed method, GRFCM, SCGM-EM, FRSCGMM, BAMM, and BGGMM. It is obvious that case IBSR04 contains low contrast between the GM and CSF tissues. BAMM and BGGMM cannot well distinguish the GM and CSF tissues when low contrast occurs. SCGM-EM and FRSCGMM lead to over-smooth segmentations. Without estimation of the intensity inhomogeneity, GRFCM also fails to distinguish the tissues with low contrast. By compareing with these methods, we see that the effect of noise and low contrast on the final segmentation of our algorithm is small and has the most similarity with the ground truth.
Fig 11

3D slice view of the real dataset (IBSR04), corresponding ground truth and segmentations by applying the proposed method, GRFCM, SCGM-EM, FRSCGMM, BAMM, and BGGMM.

Fig 12 shows the tissue surfaces of the segmentation results obtained by six algorithms (SCGM-EM, FRSCGMM, BAMM, BGGMM, GRFCM, and the proposed algorithm) on case IBSR12, and the corresponding ground truth. Fig 12(a) and 12(h) show the ground truth of the GM and WM surfaces, respectively. Fig 12(b)–12(g) show the GM surface obtained by SCGM-EM, FRSCGMM, BAMM, BGGMM, GRFCM, and the proposed method, respectively. Fig 12(i)–12(n) show the WM surface obtained by SCGM-EM, FRSCGMM, BAMM, BGGMM, GRFCM, and the proposed method, respectively. By comparing with the ground truth, it can be observed that our method obtains more accurate segmentation result.
Fig 12

Example of tissue surfaces for case IBSR12.

(a) and (h) show ground truth of GM and WM surfaces, respectively. (b) to (g) show GM surface obtained by SCGM-EM, FRSCGMM, BAMM, BGGMM, GRFCM, and the proposed method, respectively. (i) to (n) show the WM surface obtained by SCGM-EM, FRSCGMM, BAMM, BGGMM, GRFCM, and the proposed method, respectively.

Example of tissue surfaces for case IBSR12.

(a) and (h) show ground truth of GM and WM surfaces, respectively. (b) to (g) show GM surface obtained by SCGM-EM, FRSCGMM, BAMM, BGGMM, GRFCM, and the proposed method, respectively. (i) to (n) show the WM surface obtained by SCGM-EM, FRSCGMM, BAMM, BGGMM, GRFCM, and the proposed method, respectively. Then we tested all those algorithms on 18 cases in the IBSR v2.0 dataset. The segmentation results were assessed in term of DC, and the variation of DC values was depicted in Fig 13. The statistical results, including mean, standard deviation (STD), and p-value of the t-test, of those methods on 18 3D brain images (from IBSR01 to IBSR18) were listed in Table 4. According to the obtained mean and STD, we can tell that the proposed algorithm steadily outperforms other five approaches. Based on the p-values, we find that, considering 0.05 as the level of significance, the proposed algorithm provides significantly more accurate segmentations on the IBSR v2.0 dataset than other five algorithm.
Fig 13

Performance of six segmentation algorithms on 18 benchmark data sets.

Table 4

Statistics of DC values (mean, standard deviation (STD) and p-value) obtained by applying six algorithms to 18 cases from the IBSR v2.0 dataset.

AlgorithmsStatisticsGMWMCSF
SCGM-EMMean0.80770.84700.5917
STD0.05310.02390.1051
p-value3.6e-051.0e-089.0e-07
FRSCGMMMean0.83400.79860.6198
STD0.03730.02240.0975
p-value4.2e-033.6e-081.4e-05
BAMMMean0.80140.86440.5738
STD0.04550.03130.0952
p-value6.6e-071.9e-023.4e-07
BGGMMMean0.80370.86170.5826
STD0.05310.02390.1051
p-value8.3e-061.0e-021.8e-06
GRFCMMean0.78490.86530.5270
STD0.03950.02920.1204
p-value6.8e-092.4e-033.7e-08
Proposed algorithmMean0.85770.86620.6843
STD0.03360.02900.0925
p-value

Segmentation of color images

In this section, we test all comparison methods on color images in the Lab color space selected from Berkeley dataset. In Fig 14, we compare the segmentation results of three real-world color images. The first row shows the original images with image ID “105019,”, “100007,” and “28083,” from left to right, and the corresponding number of clusters is 2, 3, and 4. The segmentation results obtained by SCGM-EM, FRSCGMM, BAMM, BGGMM and the proposed SCGAGMM algorithm are shown from the second to sixth row.
Fig 14

Comparison of color image segmentations.

The image IDs are: (a) 105019, (b) 100007, (c) 28083. Images from second to sixth row show segmentation results obtained by SCGM-EM, FRSCGMM, BAMM, BGGMM, and proposed algorithm.

Comparison of color image segmentations.

The image IDs are: (a) 105019, (b) 100007, (c) 28083. Images from second to sixth row show segmentation results obtained by SCGM-EM, FRSCGMM, BAMM, BGGMM, and proposed algorithm. The first image (ID: 105019) is segmented into two classes: “lions” and background. As shown in the first row of Fig 14, SCGM-EM, FRSCGMM, BAMM and BGGMM cannot extract the “lions” accurately from the background. Parts of the background region are misclassified as the target. In comparison, the proposed algorithm can successfully extract the target from the background. We attempted to segment the second image (ID: 100007) into three classes, and the proposed method obtains better classification results with more detail. The third image (ID: 374067) consists mainly of four color components: “grassland”, “trees”, “mountain”and “sky”. All comparison methods cannot distinguish well the “trees”or “mountain”from “grassland”. In comparison, our method can successfully segments all objects, especially for the “mountain”region, and does not result in obvious misclassifications. Finally, a set of color images is tested to evaluate the performance of the proposed algorithm against the SCGM-EM, FRSCGMM, BAMM and BGGMM methods. Table 5 lists the PRI values obtained with all methods for 30 real world images. The statistical results (SR), including mean values, standard deviation (STD), and p-value for the t-test, are listed at the bottom of Table 5. From t-test results, we find that considering 0.05 as the level of significance, the proposed algorithm provides significantly better segmentations with respect to the PRI index. It is evident from the results that the proposed algorithm outperforms other methods with higher PRI values in most cases.
Table 5

PRI values of image segmentation results on Berkeley’s color image dataset.

KImage IDSCGM-EMFRSCGMMBAMMBGGMMProposed
230630.85840.85490.83720.83210.8961
30960.87700.87470.77100.87200.8779
800900.74450.63680.71820.73440.7850
800990.68820.88100.80410.83140.8912
1050190.57980.68140.78710.67890.8789
1080730.63340.56030.55310.64510.7408
1240840.77810.73590.74670.69400.7818
1350690.97460.97270.98290.98290.9841
1470910.85010.83400.78960.77520.8548
1590910.70640.71290.69620.74140.7793
3510840.69340.68340.65880.68460.6950
760020.75270.75210.76670.73770.7723
1130090.66830.66860.63470.61080.6725
1340520.58190.56570.56360.56670.7526
1630140.71080.71580.71880.70140.7280
1760390.83450.79180.81550.80570.8393
1760510.74940.74310.75370.60260.7744
1830550.78060.73390.78430.79340.8032
2490610.88470.87940.91510.78390.9149
2530550.96410.90860.96190.79340.9725
4140370.76340.82370.80580.80390.8245
970100.89850.89400.88400.81350.9014
1060250.79410.78860.79430.83400.8223
1170250.78360.77200.78740.78520.8112
1630040.75380.71110.73190.74130.7623
1970170.87070.86940.84870.86360.9086
1980040.75870.63720.74080.72100.7922
2320380.83110.81250.83820.80300.8321
2410040.79540.75960.79230.79240.8042
3610840.80900.80670.81160.80020.8273
SRMean0.77900.76870.77650.76090.8227
STD0.09760.10150.09700.08800.0750
p-value0.00142.0e-051.4e-052.7e-06

Comparison of computational complexity

The computational complexity of those five algorithms was compared in Table 6, where T is the number of iterations when the algorithm converges, N is number of pixels in an image, D is the dimension of each pixel, K is the number of clusters, L is the number of distributions and N∂ is the number of pixels in the neighborhood ∂. It should be noted that the computational complexity for the EM algorithm is of the order O(NKD2) for each iteration [49].
Table 6

Computational complexity, converging time, number of iterations and per iteration time (average ± standard deviation, UNIT: Second) by applying five algorithms on BrainWeb dataset.

AlgorithmsSCGM-EMFRSCGMMBAMMBGGMMProposed
Computational complexityO(NKD2TNi2)O(NKD2TNi2) O(NKD2TL)O(NKD2TL)O(NKD2TLNi2)
Converging time0.3804 ± 0.10900.9808 ± 0.35853.4747 ± 0.69143.4901 ± 1.47384.4978 ± 0.6397
Number of iterations41.25 ± 11.90109.70 ± 39.58160.90 ± 25.0638.40 ± 15.9437.60 ± 19.29
Per iteration time0.0092 ± 0.00010.0088 ± 0.00010.0207 ± 0.00050.0780 ± 0.00120.1391 ± 0.0519
Table 6 also gives the converging time, number of iterations and time-cost per iteration of five algorithms obtained by applying each of them to 100 2D brain MR images, which have a size of 217 × 181 and were selected from the BrainWeb dataset. In this comparative experiment, we checked for the convergence of the parameter values, set the stopping criteria to ε = 10−5 and executed each algorithm with 100 iterations (Intel Core i7-4790 CPU, 3.60GHz, 16GB RAM, 64-bit Windows 8, and Matlab Version 7.8). It is worth mentioning that the proposed algorithm was performed in the MATLAB environment without any particular code optimization.

Conclusion

To overcome the limitations involved in most GMM-based algorithms, in this paper, we proposed a rough set bounded asymmetric Gaussian mixture model with spatial constraint for image segmentation. Based on the rough set theory, a new bounded indicator function was proposed to determine the bounded support regions of the observed data. The bounded indicator and posterior probability of a pixel that belongs to each sub-region were estimated based on the rough regions. The within- and between-cluster spatial constraints were introduced by incorporating the spatial information with adaptively selected direction in order to reduce over-smoothness for segmentations. Experimental results demonstrated that the proposed algorithm is flexible to fit the data shapes, and robust to noise, which makes our method be capable of producing more accurate segmentation results comparing with several state-of-the-art algorithms. Future work will be devoted to reducing the complexity of the proposed algorithm.

Appendix

Mean value estimation

Considering the derivation of function E(Π, Θ|X) in Eq (32) with respect to μ, we have where the term is the expectation of function under distribution Φ(x|μ, Σ), which can be approximated as [14, 19, 36]: where s ∼ Φ(x|μ, Σ) is the random vector drawn from distribution Φ(x|μ, Σ), and M is the number of random vectors s [19]. In this paper, we set M = 106 for all experiments. Similarly, the term can be approximated as [14, 19, 36]: Based on Eqs (41) and (42), ∂E(Π, Θ|X)/∂μ from Eq (40) can be rewritten as: The solution ∂E(Π, Θ|X)/∂μ = 0 yields the updating function for μ during the iterations:

Covariance matrix estimation

Considering the derivation of function E(Π, Θ|X) in Eq (32) with respect to Σ, we have Similar to Eq (41), the term in the Eq (45) can be approximated as: By setting ∂E(Π, Θ|X)/∂Σ = 0, we obtain the updating function for Σ during the iterations:

Prior probability estimation

Because the prior distribution of between-cluster π satisfies constraint , Lagrange’s multiplier τ is used to enforce these constraints for each data point: Constraint allows: Similarly, we obtain the updating function for the prior distribution of within-cluster η under constraint :
  20 in total

1.  Generalized rough fuzzy c-means algorithm for brain MR image segmentation.

Authors:  Zexuan Ji; Quansen Sun; Yong Xia; Qiang Chen; Deshen Xia; Dagan Feng
Journal:  Comput Methods Programs Biomed       Date:  2011-11-15       Impact factor: 5.428

2.  Mixtures of Shifted AsymmetricLaplace Distributions.

Authors:  Brian C Franczak; Ryan P Browne; Paul D McNicholas
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  2014-06       Impact factor: 6.226

3.  A spatially constrained mixture model for image segmentation.

Authors:  K Blekas; A Likas; N P Galatsanos; I E Lagaris
Journal:  IEEE Trans Neural Netw       Date:  2005-03

4.  Random walks for image segmentation.

Authors:  Leo Grady
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  2006-11       Impact factor: 6.226

5.  A class-adaptive spatially variant mixture model for image segmentation.

Authors:  Christophoros Nikou; Nikolaos P Galatsanos; Aristidis C Likas
Journal:  IEEE Trans Image Process       Date:  2007-04       Impact factor: 10.856

6.  Incremental learning for ν-Support Vector Regression.

Authors:  Bin Gu; Victor S Sheng; Zhijie Wang; Derek Ho; Said Osman; Shuo Li
Journal:  Neural Netw       Date:  2015-04-06

7.  A Nonsymmetric Mixture Model for Unsupervised Image Segmentation.

Authors:  Thanh Minh Nguyen; Q M Jonathan Wu
Journal:  IEEE Trans Cybern       Date:  2013-03-07       Impact factor: 11.448

8.  Structural Minimax Probability Machine.

Authors:  Bin Gu; Xingming Sun; Victor S Sheng
Journal:  IEEE Trans Neural Netw Learn Syst       Date:  2016-04-14       Impact factor: 10.451

9.  A Robust Regularization Path Algorithm for $\nu $ -Support Vector Classification.

Authors:  Bin Gu; Victor S Sheng
Journal:  IEEE Trans Neural Netw Learn Syst       Date:  2016-02-24       Impact factor: 10.451

10.  Multiplicative intrinsic component optimization (MICO) for MRI bias field estimation and tissue segmentation.

Authors:  Chunming Li; John C Gore; Christos Davatzikos
Journal:  Magn Reson Imaging       Date:  2014-04-30       Impact factor: 2.546

View more
  3 in total

1.  Detect differentially methylated regions using non-homogeneous hidden Markov model for methylation array data.

Authors:  Linghao Shen; Jun Zhu; Shuo-Yen Robert Li; Xiaodan Fan
Journal:  Bioinformatics       Date:  2017-12-01       Impact factor: 6.937

2.  Similarity measure and domain adaptation in multiple mixture model clustering: An application to image processing.

Authors:  Siow Hoo Leong; Seng Huat Ong
Journal:  PLoS One       Date:  2017-07-07       Impact factor: 3.240

3.  Semantic segmentation of cerebrospinal fluid and brain volume with a convolutional neural network in pediatric hydrocephalus-transfer learning from existing algorithms.

Authors:  Florian Grimm; Florian Edl; Susanne R Kerscher; Kay Nieselt; Isabel Gugel; Martin U Schuhmann
Journal:  Acta Neurochir (Wien)       Date:  2020-06-25       Impact factor: 2.216

  3 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.