Literature DB >> 33286178

The Fisher-Rao Distance between Multivariate Normal Distributions: Special Cases, Bounds and Applications.

Julianna Pinele1, João E Strapasson2, Sueli I R Costa3.   

Abstract

The Fisher-Rao distance is a measure of dissimilarity between probability distributions, which, under certain regularity conditions of the statistical model, is up to a scaling factor the unique Riemannian metric invariant under Markov morphisms. It is related to the Shannon entropy and has been used to enlarge the perspective of analysis in a wide variety of domains such as image processing, radar systems, and morphological classification. Here, we approach this metric considered in the statistical model of normal multivariate probability distributions, for which there is not an explicit expression in general, by gathering known results (closed forms for submanifolds and bounds) and derive expressions for the distance between distributions with the same covariance matrix and between distributions with mirrored covariance matrices. An application of the Fisher-Rao distance to the simplification of Gaussian mixtures using the hierarchical clustering algorithm is also presented.

Entities:  

Keywords:  Fisher–Rao distance; Gaussian mixture simplification; information geometry; multivariate normal distributions

Year:  2020        PMID: 33286178      PMCID: PMC7516881          DOI: 10.3390/e22040404

Source DB:  PubMed          Journal:  Entropy (Basel)        ISSN: 1099-4300            Impact factor:   2.524


1. Introduction

A proper measure to determine the dissimilarity between probability distributions has been approached in many problems and applications. The Fisher–Rao distance is a very special metric for statistical models of probability distributions. This distance is invariant by reparametrization of the sample space and covariant by reparameterization of the parameter space [1]. Moreover, the Fisher–Rao metric is preserved under Markov morphisms and under centain conditions it is, up to a scaling factor, the unique Riemannian metric satisfying this condition [2,3]. Markov morphisms are associated with the notion of statistical sufficiency which express the criterion of passing from one statistical model to another with no loss of information [4,5,6]. Therefore it is natural to require the invariance of the geometric structures of statistical models under Markov morphisms. Between finite sample size simplex model , a Markov morphism is a linear map , where , with , is a matrix with non-negative entries such that every row sums to 1 and every column has precisely one non-zero element. The mapping corresponds to probabilistic refining of the event space where the refinement occurs with probability [7]. Chentsov [8,9] has proved the Fisher–Rao uniqueness invariance property under Markov morphisms for the finite sample spaces. The extension of this result to more general statistical models requires careful formulations of statistical sufficiency and Markov morphisms and has been evolved since then [3,10]. More recently in [5,6] it is shown this uniqueness of the Fisher–Rao metric under an assumption of strong continuity of the information metric. After previous papers [11,12,13] connecting geometry and statistics, C. R.Rao in an independent landmark paper [14] considered statistical models with the metric induced by the information matrix defined by R. Fisher in 1921 [15]. This work encouraged several authors to calculate the Fisher–Rao metric distance between other probability distributions [16,17,18] as well as stimulated approaches to other dissimilarity measures such as Kullback-Leibler divergence [19], total variation and Wasserstein distances [20]. Amari [3,4,21] unified the information geometry theory by organizing and introducing other concepts regarding statistical models [2]. An explicit form for the Fisher–Rao distance in the univariate normal distribution space is known via an association with the classical model of the hyperbolic plane [14,16,18,22]. It was applied to quantization of hyperspectral images [23] and to the space of projected lines in the paracatadioptric images [24]. This Fisher–Rao model was used to simplify Gaussian mixtures through the k-means method [25] and a hierarchical clustering technique [26]. An expression for the geodesic curve (initial value problem) in the multivariate normal distributions space was derived in [27] and in [28]. However, the calculus of the Fisher–Rao distance requires solving non-trivial differential equations under boundary conditions to find the geodesic connecting two distributions and then to calculate the integral along the geodesic. A closed form for this distance in the general case is still an open problem. Expressions for the distance are known only in special cases [16,17,18]. The Fisher–Rao distance between multivariate normal distributions in specific cases, such as distributions with a common mean, was considered in diffusion tensor image analysis [29,30,31], in color texture discrimination in several classification experiments [32], in the problem of distributed estimation fusion with unknown correlations [33], and in the machine learning technique [34]. In [35,36], the authors described shapes representing landmarks by a Gaussian model with diagonal covariance matrices and used the Fisher–Rao distance to quantify the difference between two shapes. In [17], this model was applied to statistical inference. Bounds for the Fisher–Rao distance were used to track quality monitoring [37]. This paper is organized as follows. In Section 2, we gather known results (closed forms for special cases and bounds) for the Fisher–Rao distance between multivariate normal distributions. In Section 3, we describe a closed form for the Fisher–Rao distance between distributions with the same covariance matrix and a non-linear system to find the distance between distributions with mirrored covariance matrices. An application of the Fisher–Rao distance to the simplification of Gaussian mixtures using the hierarchical clustering algorithm is presented in Section 4. Some conclusions and perspectives are drawn in Section 5.

2. The Fisher–Rao Distance in the Multivariate Normal Distribution Space: Special Submanifolds and Bounds

In this section, as in [38], we summarize previous results regarding the Fisher–Rao distance in the space of multivariate normal distributions including closed forms for this distance restricted to submanifolds and general bounds. Given a statistical model , a natural Riemannian structure [21] can be provided by the Fisher information matrix : where is the expected value with respect to the distribution . This matrix can also be viewed as the Hessian matrix of the Shannon entropy (concave function) [39], and is used to establish connections between inequalities in information theory and geometrical inequalities. The Fisher–Rao distance, , between two distributions and in , identified with their parameters and , is given by the shortest length of a curve in the parameter space connecting these distributions, , where . Note that this is in fact a metric, since for any , , and in , we have: (i) and if only if ; (ii) ; (iii) . A curve that provides the shortest length is called a geodesic and is given by the solutions of the differential equations where are the Christoffel symbols, and is the inverse matrix of the Fisher information matrix. We consider here the space of the multivariate normal distributions given by: where is the variable vector, is the mean vector, and is the covariance matrix in , the space of order n positive definite symmetric matrices. In this case, the model is a statistical -dimensional manifold. In this case, the model is a statistical manifold of dimension . Considering a parametrization of the model , the Fisher information matrix is given by [40] The metric provided by this matrix is invariant with respect to affine transformations. In other words, for any , where is the group of non-singular n-square matrices, the mapping: is an isometry in [16]. Consequently, the Fisher–Rao distance between and in satisfies: for any . In particular, for and , , the Fisher–Rao distance admits the form: where , is the n-order identity matrix, and is the null vector. The geodesic equations in can be expressed as [17]: and could be partially integrated [27]: where , , and B is a symmetric matrix. The initial conditions for this problem can be taken as: Eriksen [27] and Calvo and Oller [28], in independent works, solved this initial value problem. An explicit solution to the geodesic curve in [28] is: where is an n-order identity matrix, , and is the generalized inverse square matrix of G, that is . Due the fact that the geodesic curve has constant velocity at any point, given in the tangent space of , the Fisher–Rao distance between and is: where is the standard Euclidean norm. Note that the above expression provides the Fisher–Rao distance between two distributions only if we can determine the initial value problem from the boundary conditions, which usually is very difficult. Han and Park in [31] presented a numerical shooting method for computing the minimum geodesic distance between two normal distributions, through parallel transport of a vector field defined along the geodesic curve given in Equation (15). A closed form for the Fisher–Rao distance between two normal distributions in is still an important open question. Next, we present closed forms for this distance in some submanifolds of .

2.1. Closed Forms for the Fisher–Rao Distance in Submanifolds of

In this subsection, we consider submanifolds with the distance induced by the Fisher–Rao metric in . It is important to remark that, in general, given two distributions and in , the distance between and when restricted to a submanifold is bigger than the distance between and in , that is . This is due to the fact that to get , we consider the minimum length of restricted curves, which are the ones contained in the submanifold . We say that is totally geodesic if only if , for any , which means that the geodesic in connecting and is contained in .

2.1.1. The Submanifold Where Is Constant

In the n-dimensional manifold composed by multivariate normal distributions with common covariance matrix , , the Fisher–Rao distance between two distributions and is [18]: This distance is equal to the Mahalanobis distance [11], which is equal to the Euclidean distance between the image of and under the transformation , where is the Cholesky decomposition [18]. This distance was one of the first dissimilarity measures between datasets with some correlation. Note that this submanifold is not totally geodesic, as it can be seen even in the space of univariate normal distributions [22] and in Example 1 in the next section. A geodesic curve in connecting and can be provided by:

2.1.2. The Submanifold Where Is Constant

A totally geodesic submanifold of is given by of dimension composed by distributions that have the same mean vector . The Fisher–Rao distance in was studied by several authors in different contexts [16,18,30,41] and for and is given by: where are the eigenvalues of . An expression for the geodesic curve connecting these two distributions is [30]:

2.1.3. The Submanifold Where Is Diagonal

Let , the submanifold of composed by distributions with a diagonal covariance matrix. If we consider the parameter , it can be shown [22] that the metric in the parametric space of is equal to the product metric: where is the Fisher–Rao distance in the univariate case given by [22]: In this space, a curve is a geodesic if, and only if, is a geodesic curve in the univariate case, for all . The geodesic curves in the univariate normal distributions space (upper half plane ) are half-vertical lines and half-ellipses centered at , with eccentricity [22]. It is important to note that is not totally geodesic. The submanifold of composed only by normal distributions with covariance matrices which are multiples of the identity (round normals) is totally geodesic [22]. In fact, this submanifold of round normals is also contained in the totally geodesic submanifold described next.

2.1.4. The Submanifold Where Is Diagonal and Is an Eigenvector of

Let be the -dimensional submanifold composed by distributions with the mean vector for some (the canonical basis of ) and diagonal covariance matrix , and without loss of generality, we shall assume that . An analytic expression for the distance in is: We proved in [42] that this submanifold is totally geodesic.

2.2. Bounds for the Fisher–Rao in

As mentioned, a closed form for the Fisher–Rao distance between two general normal distributions is not known. In this subsection, we present some bounds for this distance.

2.2.1. A Lower Bound

Calvo and Oller [43] derived a lower bound for the Fisher–Rao distance through an isometric embedding of the parametric space into the manifold of the positive definite matrices. [ where We note that this bound satisfies the distance proprieties in . In [44], through a similar approach, a lower bound for the Fisher–Rao distance was obtained in the more general space of elliptical distributions, restricted to normal distributions, is the above bound.

2.2.2. The Upper Bound

In [45], we proposed an upper bound based on an isometry (8) in the manifold and on the distance in the non-totally geodesic submanifold (21), as follows: [ where Λ given by the eigenvalues of

2.2.3. The Upper Bounds and

Considering the Fisher–Rao distance in the totally geodesic submanifold and the triangular inequality, we propose another upper bound [42]. Given and , we consider the Fisher–Rao distance between and as in Equation (10). Let ; by the triangular inequality, it follows that: To calculate this bound, we choose appropriately. For , note that . Let P be an orthogonal matrix such that and a diagonal matrix. We will consider and . By the isometry , given in Equation (9), for and , it follows: Then, combining Inequality (27) and Equation (28), the left side of the equation below is an upper bound for the Fisher–Rao distance between and , In [42], we derived the upper bound: through a numerical minimization process by considering the diagonal elements of D as a vector that minimizes , We also derive an analytic upper bound by minimizing of the distance . By expressing this distance in terms of the parameters , we can show that it reaches the minimum at: The lower bound of Section 2.2.1 and the upper bounds of Section 2.2.2 and Section 2.2.3 are summarized in Table 1.
Table 1

The lower bound and the upper bounds , and for the Fisher–Rao distance, , between distributions and in . is the distance between univariate normal distributions given in Equation (22).

LB(θ1,θ2)=12i=1n+1[log(λi)]2 Si=Σi+μitμiμitμi1;λiaretheeigenvaluesofS11/2S2S11/2.
UB1(θ1,θ2)=i=1ndF2((0,1),(μi,λi)) Σ1(1/2)Σ2Σ1(1/2)=QΛQt;λiarethediagonaltermsofthematrixΛ;μiarethecoordinatesofμ=QtΣ1(1/2)(μ2μ1).
UB2(θ1,θ2)=12i=1n[log(λi)]2+ dF2((0,1),(|μ3|,d¯1))+i=2mdF2((0,1),(0,d¯i)) μ3=Σ1(1/2)(μ2μ1);d¯iisgivenby (31);Σ3=Σ1(1/2)Σ2Σ1(1/2);P isanorthogonalmatrixsuchthat                     Pμ=(|μ3|,0,,0);Σ¯=P1Σ3Pt;λiaretheeigenvaluesofΣ¯(1/2)Σ3Σ¯(1/2).
UB3(θ1,θ2)=12i=1n[log(λi)]2+ dF(0,1),|μ3|,|μ3|+22 μ3=Σ1(1/2)(μ2μ1);Σ3=Σ1(1/2)Σ2Σ1(1/2);P isanorthogonalmatrixsuchthat                Pμ=(|μ3|,0,,0);Σ¯=P1Σ3Pt;λiaretheeigenvaluesofΣ¯(1/2)Σ3Σ¯(1/2).
Upper and lower bounds have been used to estimate the Fisher–Rao distance in applications such as [37].

2.2.4. Comparisons of the Bounds

In this section, as in [42], we illustrate comparisons between the bounds presented previously. We consider the bivariate normal distributions model () and distributions and , where: From (10), we can see that there always exists an isometry that converts any two pairs of bivariate distributions into a pair of distributions as above. We present next a comparison between the lower bound “LB” (25), the upper bounds (26), (31), and (32), and the numerical solution given by the geodesic shooting algorithm (GS) [31] in specific situations. In Figure 1, we consider the eigenvalues , , and to be fixed and varying from zero to . We note that the upper bound is very near the lower bound and to the numerical solution . The other upper bounds are bigger than the bound . In Figure 2, it is considered and the previous eigenvalues. Now, the best performance is of bounds and , which are similar. In Figure 3a, we again keep the eigenvalues; the rotation angle is fixed ; and varies from zero to 10. We can see similar performances of and , which are better than for larger values of .
Figure 1

A comparison between the bounds , , , , and . (, , and are fixed, and varies from zero to ).

Figure 2

A comparison between the bounds , , , , and . (, , and are fixed, and varies from zero to ).

Figure 3

(a) A comparison between the bounds , , , , and . (, , and the rotation angle are fixed, and varies from zero to 10). (b) A comparison between the bounds , , and . (, , and the rotation angle are fixed, and varies from zero to 10).

We may also consider the upper bound: Figure 3b displays the comparison between , , and for the same data of the comparison in Figure 3a.

3. Fisher–Rao Distance Between Special Distributions

In this section, we describe the Fisher–Rao distance in the full space between special kinds of distributions.

3.1. The Fisher–Rao Distance Between Distributions with Common Covariance Matrices

The Fisher–Rao distance between distributions with common covariance matrices given in Section 2.1.1 was restricted to non-totally geodesic submanifold . We show next that using the isometry given in (8) and the distance in the submanifold , it is possible to find a closed form for the distance between two distributions with the same covariance matrix, in the full manifold . Given two distributions where U is an upper triangular matrix with all diagonal entries equal to one and D is a diagonal matrix. The Fisher–Rao distance between By considering the isometries and and the decomposition given by Equation (35), it follows from Equation (9) that: Since the distributions and belong to the submanifold , we conclude that: □ Consider two bivariate normal distributions

3.2. The Fisher–Rao Distance Between Mirrored Distributions

We consider here two mirrored normal distributions; that is, without loss of generality, if we consider up rotation, the line connecting and as parallel to the -axis, and the covariance matrices and satisfying: This condition implies also the same eigenvalues for both matrices. For bivariate normal distributions, we then should have: see Figure 5.
Figure 5

Example of level curves of mirrored distributions where and are given by Equation (40).

After several experiments using the algorithm geodesic shooting for the and , we have observed that for the geodesic curve connecting these distributions (, with and ), satisfies where , , and are real values; see Figure 6.
Figure 6

Approximation of the geodesic curve connecting and via the geodesic shooting algorithm. The level curve of is the dashed one.

The focus here is the “shape” of these distributions. Note that at , the distribution appears as , which has a diagonal covariance matrix, and the tangent vector appears as , which is composed by a mean vector with the second entry equal to zero and by a symmetric covariance matrix with a null diagonal. This observation inspired us to get an explicit expression for the geodesic connecting two mirrored distributions. Starting with the bi-dimensional case again, we will prove that in fact we have equality in Expressions (41) and (42). Let , , and the geodesic curve in connecting and , and consider that and . Given the isometry , we define: Then: and: Applying the natural changing of parameters: it follows that: Then, given that , That is, at , the tangent vector is equal to the tangent vector in (46). Furthermore, the distributions and are also mirrored . In fact, and by similar arguments, we obtain: Figure 7 illustrates the distributions , , and .
Figure 7

Contour curves of distributions and .

Conversely, by considering: in the initial value problem given in Equations (13) and (14), it follows that the matrix is diagonal. Therefore, the geodesic curve with initial value and tangent vector given in Equation (15) can be simplified as follows: From the parity of the functions and , it is possible to show that, given , the distributions: are also mirrored. By the above discussion, we conclude that it is possible to calculate the geodesic curve connecting and making and . That is, we need to find the values of , , and of the isometry and the values of such that: Since the two equations above are equivalent, it is enough to solve the equation: This is equivalent to solving the system: where . The above non-linear system has five equations and five variables and can be solved by an iterative method. With the solution of this system, we can determine the geodesic curve connecting the distributions and . Moreover, by Equation (16), the Fisher–Rao distance is: We also remark that the curve of the means (and therefore, ) satisfies the equation of a hyperbola; in fact: Summarizing the above discussion, we have: Expression (57) provides a closed form for the Fisher–Rao distance between two mirrored bivariate normal distributions, based on the solutions of the non-linear system (56). The plane curve given by the coordinates of the mean vector in the geodesic connecting two of these distributions is a hyperbola. Table 2 shows a time comparison between the numerical method proposed here and the geodesic shooting to obtain the Fisher–Rao distance. The distributions used in this experiment were: for different values of .
Table 2

A time comparison between the numerical method proposed here and the geodesic shooting to calculate the distance between two mirrored distributions.

μ dF(θ1,θ2) Time Systems (s)Time G.Shooting (s)
12.773950.0468754.70313
23.670270.0468755.60938
34.529330.06257.10938
45.260930.0781259.17188
55.874800.04687512.5313
66.394390.062518.4219
76.840430.078125492.563
87.229030.0625574.422
97.572210.046875917.859
107.878960.0468751007.13
The method proposed here uses a non-linear system for the calculus of the Fisher–Rao distance, so it is faster the geodesic shooting algorithm. Furthermore, we remark that for , the geodesic shooting requires additional adaptation to convergence. Next, we generalize the results of Proposition 4 to pairs of general multivariate normal mirrored distributions. Without loss of generality, we may assume: with as in (39), that is: The Fisher–Rao distance between a pair of multivariate mirrored normal distributions where: The values x and where Let , , be the geodesic curve in connecting and . The proof is similar to the bivariate case, by considering , and where . with . Table 3 collects the results in Section 2.1 and the new results of this section.
Table 3

Closed forms for the Fisher–Rao distance in submanifolds of and the distance in between pairs of special distributions.

Distance in Non-totally Geodesic Submanifolds
SubmanifoldDistance
MΣ=pθ;θ=(μ,Σ),Σ=Σ0Pn(R)constant,θi=(μi,Σ0) dΣ(θ1,θ2)=(μ1μ2)tΣ01(μ1μ2)
MD=pθ;θ=(μ,Σ),Σ=diag(σ12,σ22,,σn2),θi=(μ1i,σ1i,μ2i,σ2i,,μni,σni) dD(θ1,θ2)=i=1ndF2((μ1i,σ1i),(μ2i,σ2i))
Distance in Totally Geodesic Submanifolds
Mμ=pθ;θ=(μ,Σ),μ=μ0Rncontant,θi=(μ0,Σi)dF(θ1,θ2)=12i=1n[log(λi)]2, where λi are the eigenvalues of Σ11/2Σ2Σ11/2
MDμ={pθ;θ=(μ,Σ),μisaneigenvectorofΣ=diag(σ12,σ22,,σn2)},θi=(μ1i,σ1i,σ2i,,σni) dDμ(θ1,θ2)=(dF2μ11,σ11,μ21,σ21+             i=2ndF20,σ1i,0,σ2i)1/2
Distance Between Special Distributions in M
  Distributions with Common Covariance Matrices, θi=(μi,Σ0) dF(θ1,θ2)=dDμ(0,D),(|μ2μ1|e1,D),where P is an orthogonal matrix such thatP(μ2μ1)=|μ2μ1|e1 and PΣPt=UDUt
Mirrored Distributions,θ1=(μ1e1,Σ1) and θ2=(μ2e1,Σ2),with Σ2=M1Σ1M1dF(θ1,θ2)=2l=1n1bl2+x2,where x and bi are obtained bythe solution of Equation (63)

4. Hierarchical Clustering for Diagonal Gaussian Mixture Simplification

A parameterized Gaussian mixture model f is a weighted sum of m multivariate normal distributions, that is, where , , , are normal distributions and , , are mixture, . In this paper, we call the diagonal Gaussian mixture model (DGMM) the mixture composed only by distributions with diagonal covariance matrices. Gaussian mixture models (GMM) are used in modeling datasets: image processing, signal processing, and density estimation problems [46,47,48]. In many applications involving mixture models, the computational requirements are of a very high level due to the large number of mixture components. This can be handled if we reduce the number of components of the mixture: given a mixture f of m components, we want to find a mixture g of l components, , such that g is a good approximation of f with respect to a similarity measure [49]. Gaussian mixture simplification was considered in statistical inference in [50] and to decode low-density lattice codes [51]. In [49] was proposed a hierarchical clustering algorithm to simplify an exponential family mixture model based on Bregman divergences. This section describes an agglomerative hierarchical clustering method based on the Fisher–Rao distance in the submanifold (21) to simplify DGMM, and we present an application to image segmentation, complementing what was developed in [52]. We start by introducing the concept of the centroid for a set of distributions in .

4.1. Centroids in the Submanifold

In [53], Galperin described centroids in the two-dimensional Minkowski model, which can be translated also to the Klein disk and Poincare half-plane models. Given a set of points in the Minkowski model, with associated weights , the centroid is computed and normalized as: To calculate the centroid of a subset of points , , the isometries presented in [25] and the relation between the media × standard deviation plane of parameters of univariate normal distributions and the Poincare half-plane given in [22] are used. Given a dataset , where are distributions in , the centroid of is: where , , is the centroid of given in Equation (66).

4.2. Hierarchical Clustering Algorithm

Let a DGMM f with parameters . In order to apply the hierarchical clustering algorithm, we need to consider the distance between two subsets A and B. The three most common distances are called linkage criteria and are given by [54]: Single linkage: Complete linkage: Group average linkage: where is the distance in the submanifold and is the number of elements of a set X. A summary of the hierarchical clustering algorithm (Algorithm 1) [49] using one of these distances is given next. Form m clusters with one element. Find the two closest clusters, and , with respect to a distance D, and merge them into asingle cluster . Compute distances between the new cluster and each of the old clusters. Repeat Steps 2 and 3 until all items are clustered into a single cluster of size n. The simplified DGMM: of l components is built from the l subsets , …, remaining after the iteration of the hierarchical clustering algorithm. In this work, we choose the parameters of in two ways: as the centroid in the submanifold (Fisher–Rao hierarchical clustering) and as the Bregman left-sided centroid [49] (Bregman–Fisher–Rao hierarchical clustering) of the subset with weights . As remarked in [49], the hierarchical clustering algorithm allows introducing a method to learn the optimal number of components in the simplified mixture g. Thus, g must be as compact as possible and reach a minimum prescribed quality , where is the Kullback–Leibler divergence.

4.3. Experiments in Image Segmentation

We can apply the Fisher–Rao and the Bregman–Fisher–Rao hierarchical clusterings to simplify a mixture of exponential families in the context of clustering-based image segmentation as was done in [49] for the Bregman hierarchical clustering. Given an input color image I, we adapt the Bregman soft clustering algorithm to generate a DGMM f of 32 components, which models the image pixels. We point out that the restriction considered in this paper (only DGMM) is also used in many applications due its much lower computational cost. We consider here a pixel as a point in , where , , and are the RGB color information. For image segmentation, we can say that the image pixel belongs to the class when: Thus, the segmented image is illustrated by replacing the color value of the pixel by the mean of the Gaussian . Using the the Fisher–Rao and the Bregman–Fisher–Rao hierarchical clusterings, we simplify the mixture f into mixtures g of l components with . Each mixture gives one image segmentation. The linkage criterion used here was the complete linkage (68), which has presented better results in our simulations. Figure 8 shows the segmentation of the Baboon, Lena, and Clown input images given by the Bregman–Fisher–Rao hierarchical clustering. The number of colors in each image is equal to the number of components in the simplified mixture g.
Figure 8

Illustration of the mixture simplification using the Fisher–Rao clustering, where l is the number of components of the mixture (the last column is the original figure).

The quality of the segmentation was analyzed as a function of l through the Kullback–Leibler divergence estimated by the Monte Carlo method, since there was no closed form for this measure (five thousand points were randomly drawn to estimate ). Figure 9, Figure 10 and Figure 11 show the evolution of the simplification quality as a function of the number of components l for the Baboon, Lena, and Clown images, using the Bregman, the Fisher–Rao, and the Bregman–Fisher–Rao hierarchical clustering algorithms. We observed that the image quality increased ( decreased) with l, as expected, and the behavior was similar in all clustering algorithms. In general, the Bregman–Fisher–Rao hierarchical clustering algorithm presented better results. Considering the constraint , the learning process provided, for the Bregman–Fisher–Rao hierarchical clustering, mixtures of 19, 21, and 21 as optimal simplifications for the images of the Baboon, Lena, and Clown, respectively.
Figure 9

Illustration of the simplification quality of the mixture modeling Baboon image.

Figure 10

Illustration of the simplification quality of the mixture modeling Lena image.

Figure 11

Illustration of the simplification quality of the mixture modeling Clown image.

5. Concluding Remarks

The Fisher–Rao distance was approached here in the space of multivariate normal distributions. Initially, as in [38], we summarized some known closed forms for this distance in submanifolds of this model and some bounds for the general case. A closed form for the Fisher–Rao distance between distributions with the same covariance matrix was obtained in Proposition 3, and we also have derived a non-linear system characterizing the distance between two distributions with mirrored covariance matrices in Proposition 5. Some perspectives for future research related to this topic include deriving new bounds for the Fisher–Rao distance in the general case, by using these special distributions, to characterize as non-linear systems the distances between other types of distributions and to extend the closed forms and bounds presented here to the space of elliptical distributions. Finally, we have extended the analysis of the Bregman–Fisher–Rao hierarchical clustering algorithm to simplify Gaussian mixtures in the context of clustering-based image segmentation given in [52] with comparative results that encourage the use of the Fisher–Rao distance in other clustering or classification algorithms.
  2 in total

1.  Simplifying mixture models using the unscented transform.

Authors:  Jacob Goldberger; Hayit K Greenspan; Jeremie Dreyfuss
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  2008-08       Impact factor: 6.226

2.  Simplifying mixture models through function approximation.

Authors:  James T Kwok; Kai Zhang
Journal:  IEEE Trans Neural Netw       Date:  2010-02-22
  2 in total
  1 in total

Review 1.  Parameter estimation and uncertainty quantification using information geometry.

Authors:  Jesse A Sharp; Alexander P Browning; Kevin Burrage; Matthew J Simpson
Journal:  J R Soc Interface       Date:  2022-04-27       Impact factor: 4.293

  1 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.