Next Article in Journal
The Effects of Mo and Nb on the Microstructures and Properties of CrFeCoNi(Nb,Mo) Alloys
Next Article in Special Issue
The Relationship between the US Economy’s Information Processing and Absorption Ratios: Systematic vs Systemic Risk
Previous Article in Journal
Tensile Behavior and Evolution of the Phases in the Al10Co25Cr8Fe15Ni36Ti6 Compositionally Complex/High Entropy Alloy
Previous Article in Special Issue
Entropy and Geometric Objects
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

On the Geodesic Distance in Shapes K-means Clustering

by
Stefano Antonio Gattone
1,*,
Angela De Sanctis
2,
Stéphane Puechmorel
3 and
Florence Nicol
3
1
Department of Philosophical, Pedagogical and Economic-Quantitative Sciences, University “G. d’Annunzio” of Chieti-Pescara, 66100 Chieti, Italy
2
Department of Business Economics, University "G. D'Annunzio" of Chieti-Pescara, 65127 Pescara, Italy
3
Ecole Nationale de l’aviation Civile (ENAC), Université Fédérale de Toulouse, FR-31055 Toulouse CEDEX, France
*
Author to whom correspondence should be addressed.
Entropy 2018, 20(9), 647; https://doi.org/10.3390/e20090647
Submission received: 31 July 2018 / Revised: 22 August 2018 / Accepted: 27 August 2018 / Published: 29 August 2018

Abstract

:
In this paper, the problem of clustering rotationally invariant shapes is studied and a solution using Information Geometry tools is provided. Landmarks of a complex shape are defined as probability densities in a statistical manifold. Then, in the setting of shapes clustering through a K-means algorithm, the discriminative power of two different shapes distances are evaluated. The first, derived from Fisher–Rao metric, is related with the minimization of information in the Fisher sense and the other is derived from the Wasserstein distance which measures the minimal transportation cost. A modification of the K-means algorithm is also proposed which allows the variances to vary not only among the landmarks but also among the clusters.

1. Introduction

Shapes clustering is of interest in various fields such as geometric morphometrics, computer vision and medical imaging. In the clustering of shapes, it is important to select an appropriate measurement of distance among observations. In particular, we are interested in classifying shapes which derive from complex systems as expression of self-organization phenomenon. We consider objects whose shapes are based on landmarks [1,2,3]. These objects can be obtained by medical imaging procedures, curves defined by manually or automatically assigned feature points or by a discrete sampling of the object contours.
Since the shape space is invariant under similarity transformations, that is translations, rotations and scaling, the Euclidean distance on such a space is not really meaningful. In Shape Analysis [4], to apply standard clustering algorithms to planar shapes, the Euclidean metric has to be replaced by the metric of the shape space. Examples were provided in References [5,6], where the Procrustes distance was integrated in standard clustering algorithms such as the K-means. Similarly, Lele and Richtsmeier [7] applied standard hierarchical or K-means clustering using dissimilarity measures based on the inter-landmark distances. In a model-based clustering framework, Huang and Zhu [8] and Kume and Welling [9] developed a mixture model of offset-normal shape distributions.
In Shape Analysis, it is common to assume that the landmark coordinates have an isotropic covariance structure [4]. To relax the isotropic assumption, a bivariate Gaussian model was proposed to describe the landmarks of a planar shape [10,11], where the means are the landmark geometric coordinates and capture uncertainties that arise in the landmark placement while the variances derive from the natural variability across the population of shapes. The novelty of this shape representation is given by the fact that variances are considered as additional coordinates for the landmarks of a shape. According to Information Geometry, the space of bivariate Gaussian densities is considered as a statistical manifold [12,13] with the local coordinates given by the model parameters. In this way, distances between landmarks can be defined using the geodesic distances induced by different Riemannian metrics.
In this paper, we consider the Fisher–Rao and the Wasserstein metrics as Riemannian metrics on the statistical manifold of the Gaussian densities. The geodesic distance induced by the Fisher–Rao metric is related to the minimization of information in the Fisher sense while the Wasserstein distance is related to the minimal transportation cost. Applications of geodesics to shape clustering techniques have also been provided in a landmark-free context [14,15].
As is well known, any hierarchical clustering algorithm uses as input the pairwise distances of all possible pairs of objects under study. Using the geodesic distances induced by Wasserstein and Fisher–Rao metrics, in References [10,11], a hierarchical clustering algorithm which allows the variances to vary among the landmarks was proposed.
In this paper, the discriminative power of these shapes distances is evaluated in the setting of shapes K-means clustering which is easier to implement and computationally faster. Furthermore, a modification of the K-means algorithm is proposed which allows the variances to vary not only among the landmarks but also among the clusters. The simulation results show that the proposed algorithm is able to cope with the effects of anisotropy in the landmark variances across different clusters.

2. Geometrical Structures for a Manifold of Probability Distributions

We call “manifold” a geometric object which is locally Euclidean then described by local coordinates. Manifolds can be used to study patterns from complex systems. Since pattern recognition essentially relies on quantitative assessment of the proximity of points, for the comparison of patterns, we need a well-suited similarity measure (distance or divergence). From Differential Geometry, we know that a Riemannian metric on a differential manifold X is induced by a metric matrix g, which defines an inner product on every tangent space of the manifold as follows: u , v = u T g i j v with associated norm u = u , u . Then, the distance between two points P , Q of the manifold is given by the minimum of the lengths of all the piecewise smooth paths γ joining these two points. Precisely, the length of a path is calculated using the inner product,
Length   of   γ = γ ( t ) d t
thus
d ( P , Q ) = m i n γ { Length   of   γ } .
A curve that encompasses this shortest path is called a Riemannian geodesic and the previous distance is named geodesic distance. We remark that in general the concept of geodesic is related to connections defined on a manifold. If a connection is not Riemannian, then a geodesic is different from a shortest path.
Probability theory, in the presence of non-deterministic phenomena, provides a natural description of the raw data. Each measurement x is regarded as a sample from an underlying probability distribution of the measurement characterized by its probability density function p ( x / θ ) . Measurements described by the distribution parameters, θ , may contain more information than a measurement expressed as a value and an associated error bar. Therefore, we apply pattern recognition methods directly in the space of probability distributions. Let P be a family of probability density functions p ( x θ ) parameterized by θ R k . It is well known that we can endow it with a structure of manifold, called statistical manifold, whose local coordinates are the parameters of the family. As an example, we consider the family of p-variate Gaussian densities:
f ( x θ = ( μ , Σ ) ) = ( 2 π ) p 2 ( det Σ ) 1 2 exp { 1 2 ( x μ ) T Σ 1 ( x μ ) }
where x = ( x 1 , x 2 , , x p ) T , μ = ( μ 1 , μ 2 , μ p ) T is the mean vector and Σ the covariance matrix. Note that the parameter space has dimension k = p + p ( p + 1 ) 2 . In particular, we are interested in the case p = 2 .
Two geometrical structures have been extensively studied for a manifold of probability distributions. One is based on the Fisher information metric (Fisher–Rao metric), which is invariant under reversible transformations of random variables, while the other is based on the Wasserstein distance of optimal transportation, which reflects the structure of the distance between random variables.
In the statistical manifold of bivariate Gaussian densities, we consider these two different Riemannian metrics which in turn induce two types of geodesic distances.

Fisher–Rao Metric for Gaussian Densities

The geometry of the Gaussian manifold endowed with the Fisher–Rao metric was intensively studied in References [16,17]. To avoid considering manifolds with boundaries, it is convenient to assume that all densities are non-degenerate, thus the covariance matrices are invertible. In this case, one can define the manifold of n-dimensional Gaussian densities as the set R n × R n ( n + 1 ) / 2 = R n + n ( n + 1 ) / 2 with local charts given by the obvious identification N n ( μ , Σ ) ( μ i , i = 1 n , σ i j , i = 1 n , j n ) , where the σ i j are the elements of the matrix Σ . A tangent vector at a point ( μ , Σ ) of the manifold is just a vector from R n + n ( n + 1 ) / 2 . While quite tractable, this choice of parameterization does not give any insight about the structure of the manifold. A more enlightening approach is obtained by considering groups of transformations, as detailed below.
Let symm + ( n ) be a group of symmetric positive definite matrices of size n × n endowed with the product [18]:
( A , B ) A B = A 1 / 2 B A 1 / 2
and let us denote, using a common abuse of notation, the group of translations of R n also by R n .
Now, define the group G ( n ) as the semi-direct product:
G ( n ) = symm + ( n ) ρ R n
where the action ρ of symm + ( n ) on R n is given by left multiplication with the square root of the matrix, namely:
ρ ( A ) u = A 1 / 2 u , A symm + ( n ) , u R n
In the sequel, we are dropping the ρ subscript in the semi-direct product and assume it implicitly.
An element in G ( n ) can be represented as a couple ( A , u ) with A symm + ( n ) , u R n . The group product is obtained from the action ρ as ( A , u ) · ( B , v ) = ( A 1 / 2 B A 1 / 2 , A 1 / 2 v + u ) .
The inverse of an element ( A , u ) is given by ( A 1 , A 1 / 2 u ) . The group G ( n ) is a Lie group with Lie algebra g ( n ) = symm + ( n ) R n with symm + ( n ) the vector space of symmetric matrices. Finally, the left translation by an element ( A , u ) is the mapping:
( B , v ) L ( A , u ) ( B , v ) = ( A , u ) · ( B , v )
Being an affine map, its derivative is its linear part. The Frobenius inner product on the space of square matrices of dimension n, defined as A , B = tr A t B = tr A B t , jointly with the standard euclidean inner product on R n , induces a left invariant metric by:
( X , η ) , ( Y , ξ ) ( A , u ) = K tr A 1 / 2 X A 1 Y A 1 / 2 + η 1 t A 1 η 1
where ( X , η ) , ( Y , ξ ) are tangent vectors to G ( n ) at ( A , u ) and K > 0 is a fixed scaling factor that may be arbitrary chosen to balance the relative contributions of the matrix part and the translation part.
It turns out that the metric obtained that way is exactly the Fisher–Rao metric on the manifold of multivariate Gaussian densities. Using the notations of Skovgaard [16], the length element of the Fisher–Rao metric g F is:
d s 2 = 1 2 tr Σ 1 X Σ 1 X + η t Σ 1 η
with ( X , η ) a tangent vector at ( Σ , μ ) .
The expression of d s 2 is the one of a warped product metric [19], which allows some simplifications when computing the geodesics between two densities with same means.
A closed form for the geodesic distance between two densities with diagonal covariance matrices may also be obtained as follows [17]:
d F ( θ , θ ) = 2 i = 1 2 ln | ( μ i 2 , σ i ) ( μ i 2 , σ i ) | + | ( μ i 2 , σ i ) ( μ i 2 , σ i ) | | ( μ i 2 , σ i ) ( μ i 2 , σ i ) | | ( μ i 2 , σ i ) ( μ i 2 , σ i ) | 2
where θ = ( μ , Σ ) with μ = ( μ 1 , μ 2 ) and Σ = diag ( σ 1 2 , σ 2 2 ) , θ = ( μ , Σ ) with μ = ( μ 1 , μ 2 ) and Σ = diag ( ( σ 1 ) 2 , ( σ 2 ) 2 ) .
For general Gaussian densities with Σ any symmetric positive definite covariance matrix, a closed form for the geodesic distance is not known and one has to solve numerically a system of differential equations:
D t t μ D t Σ Σ 1 D t μ = 0
D t t Σ + D t μ D t μ t D t Σ Σ 1 D t Σ = 0
where the expression D t (respectively, D t t ) stands for derivative (respectively, second derivative) with respect to t. A geodesic between two densities can be found by a shooting approach, which starts with one density as an initial condition to the system in Equation (8) and iteratively adjusts the initial speed vector of the curve so as to reduce the distance to the target density until the desired accuracy is reached. A collocation algorithm can also be used, and is a common choice for solving ordinary differential equations with boundary conditions. It is generally more stable than the shooting method, but may require more computations. In both cases, a tricky part of the process is to ensure that the Σ matrix remains positive definite. A rewrite of Equation (8) with the Cholesky decomposition Σ = L t L K allows this condition to be satisfied by design and is the preferred choice. Another option to get an approximate value is to use Equation (7) after diagonalizing the covariance matrices.
In regard to the Riemannian metric g w which induces the Wasserstein distance [20], for Gaussian densities, the explicit expression of the distance is the following:
d W ( θ , θ ) = | μ μ | + t r ( Σ ) + t r ( Σ ) 2 t r ( Σ 1 2 Σ Σ 1 2 )
where | · | is the euclidean norm and Σ 1 2 is defined for a symmetric positive definite matrix Σ so that Σ 1 2 · Σ 1 2 = Σ . We remark that, if Σ = Σ , the Wasserstein distance reduces to the Euclidean distance.
Otto [20] proved that, with respect to the Riemannian metric which induces the Wasserstein distance, the manifold of Gaussian densities has non-negative sectional curvature. We deduce that the Wasserstein metric is different from the Fisher–Rao metric. Indeed, for example, in the univariate case, the statistical manifold of Gaussian densities with the Fisher–Rao metric can be regarded as the upper half plane with the hyperbolic metric, which has negative curvature as it is well known.
Once a distance is defined, it can be used for clustering on a manifold. It is proven that the distance induced from Fisher–Rao metric and Wasserstein distance are in the more general class of Bregman divergences defined by a convex function [21]. For this class, a theorem states [22] that the centroid for a set of n points θ i , i = 1 , 2 , , n in the statistical manifold of the Gaussian densities is the Euclidean mean 1 n i = 1 n θ i . We use this result in the next section where a K-mean shapes clustering algorithm is defined using geodesic distances.

3. Clustering of Shapes

We consider only planar objects, as for example a flat fish or a section of the skull. The shape of the object consists of all information invariant under similarity transformations, that is translations, rotations and scaling [4]. Data from a shape are often realized as a set of points. Many methods allow to extract a finite number of points, which are representative of the shape and are called landmarks. One way to compare shapes of different objects is to first register them on some common coordinate system for removing the similarity transformations [2,23]. Alternatively, Procrustes methods [24] may be used in which objects are scaled, rotated and translated so that their landmarks lie as close as possible to each other with respect to the Euclidean distance.
Suppose we are given a planar shape configuration, S, consisting of a fixed number K of labeled landmarks
S = μ 1 , μ 2 , , μ K
with generic element μ k = μ k 1 , μ k 2 , for k = 1 , , K . Following Gattone et al. [10], the k-th landmark, for k = 1 , , K , may be represented by a bivariate Gaussian density as follows:
f ( x θ k = ( μ k , Σ k ) ) = ( 2 π ) 1 ( det Σ k ) 1 2 exp { 1 2 ( x μ k ) T Σ k 1 ( x μ k ) }
with x being a generic 2-dimensional vector and Σ k given by
Σ k = diag ( σ k 1 2 , σ k 2 2 )
where σ k 2 = ( σ k 1 2 , σ k 2 2 ) is the vector of the variances of μ k .
We remark that, in the previous representation, the means represent the geometric coordinates of the landmark and capture uncertainties that arise in the landmark placement. The variances are hidden coordinates of the landmark and reflect its natural variability across a population of shapes. Equation (11) allows assigning to the kth landmark the coordinates θ k = ( μ k , σ k ) on the four-dimensional manifold which is the product of two upper half planes.
Let S and S two planar shapes registered on a common coordinate system using Procrustes method. We parameterize them as follows: S = ( θ 1 , , θ K ) and S = ( θ 1 , , θ K ) .
The distances between landmarks allow defining a distance of the two shapes S and S . Precisely, a shape metric for measuring the difference between S and S can be obtained by taking the sum of the geodesic distances between the corresponding landmarks, according to the following definition:
D ( S , S ) = k = 1 K d ( θ k , θ k )
Please note that this expression is not the geodesic distance on the product manifold that one would have expected from the landmark model. This last distance is given by:
D ( S , S ) = k = 1 K d ( θ k , θ k ) 2
and is a L 2 distance instead of Equation (13) that is L 1 . It turns out that, according to simulations done, the L 1 approach is more robust and gives all the time better clusterings.
Then, a classification of shapes, using in turn, as distance d, the distance d F induced from Fisher–Rao metric and the Wasserstein distance d W , can be done following the standard methodology. In particular, the K-means clustering procedure allows the variances to vary step by step in each cluster fitting better real shape data.

4. K-Means Clustering Algorithm

The proposed shape distances are implemented in two different K-means algorithms: Type I and Type II. While in the Type I algorithm the landmark coordinates variances are assumed isotropic across the clusters, in Type II the variances are allowed to vary among the clusters.
Our task is clustering a set of n shapes, S 1 , S 2 , , S n into G different clusters, denoted as C 1 , C 2 , , C G .

4.1. Type I Algorithm

  • Initial step:
    Compute the variances of the k-th landmark coordinates σ k 2 = ( σ k 1 2 , σ k 2 2 ) , for k = 1 , , K .
    Randomly assign the n shapes, S 1 , S 2 , , S n into G clusters, C 1 , C 2 , , C G .
    For g = 1 , , G , calculate the cluster center c g = ( θ 1 g , , θ K g ) with k-th component θ k g = ( μ g k , σ k 2 ) obtained as θ k g = 1 n g i C g θ k i , where n g is the number of elements in the cluster C g and θ k i is the k-th coordinate of S i given by θ k i = ( μ i k , σ k 2 ) .
  • Classification:
    For each shape S i , compute the distances to the G cluster centers c 1 , c 2 , , c G .
    The generic distance between the shape S i and the cluster center c g is given by:
    D ( S i , c g ) = k = 1 K d ( θ k i , θ k g )
    where the distance d could be the distance d F induced from Fisher–Rao metric or the Wasserstein distance d W .
    Assign S i to cluster h that minimizes the distance:
    D ( S i , c h ) = min g D ( S i , c g ) .
  • Renewal step:
    Compute the new cluster centers of the renewed clusters c 1 , , c G .
    The k-th component of the g-th cluster center c g is defined as θ k g = 1 n g i C g θ k i .
  • Repeat Steps 2 and 3 until convergence [22].

4.2. Type II Algorithm

  • Initial step:
    Randomly assign the n shapes, S 1 , S 2 , , S n into G clusters, C 1 , C 2 , , C G .
    In each cluster, compute the variances of the k-th landmark coordinates σ g k 2 = ( σ g k 1 2 , σ g k 2 2 ) , for k = 1 , , K and g = 1 , , G .
    Calculate the cluster center c g = ( θ 1 g , , θ K g ) with k-th component θ k g = ( μ g k , σ g k 2 ) obtained as θ k g = 1 n g i C g θ k i for g = 1 , , G , where n g is the number of elements in the cluster C g and θ k i = ( μ i k , σ g k 2 ) for i C g .
  • Classification:
    For each shape S i , compute the distances to the G cluster centers c 1 , c 2 , , c G .
    The generic distance between the shape S i and the cluster center c g is given by:
    D ( S i , c g ) = k = 1 K d ( θ k i , θ k g )
    where the distance d could be the distance d F induced from Fisher–Rao metric or the Wasserstein distance d W .
    Assign S i to cluster h that minimizes the distance:
    D ( S i , c h ) = min g D ( S i , c g ) .
  • Renewal step:
    Update the variances of the k-th landmark coordinates in each cluster by computing σ g k 2 = ( σ g k 1 2 , σ g k 2 2 ) , for k = 1 , , K and for g = 1 , , G .
    Calculate the new cluster centers of the renewed clusters c 1 , , c G .
    The k-th component of the g-th cluster center c g is defined as θ k g = 1 n g i C g θ k i .
  • Repeat Steps 2 and 3 until convergence [22].

5. Numerical Study

The purpose of the simulation study was to evaluate the cluster recovery of the proposed shape K-means algorithm and to test its sensitiveness with respect to different shape distances defined on the manifold of the probability distributions. The shapes were simulated according to a Gaussian perturbation model where the ith configuration is obtained as follows:
X i g = ( μ g + E i ) Γ i + 1 K γ i T
where
  • E i are zero mean K × 2 random error matrices simulated from the multivariate Normal distribution with covariance structure Σ E ;
  • μ g is the mean shape for cluster g;
  • Γ i is an orthogonal rotation matrix with an angle θ uniformly produced in the range [ 0 , 2 π ] ; and
  • γ i T is a 1 × 2 uniform translation vector in the range [ 2 , 2 ] .
Three types of covariance structures are considered:
  • Isotropic with Σ E = σ I K σ I 2
  • Heteroscedastic with Σ E = diag [ σ 1 , σ 2 , σ K ] σ I 2
  • Anisotropic with Σ E = σ I K diag [ σ x , σ y ] with σ x σ y
The data were generated from the model in Equation (15) with sample size n = 100 and the number of clusters equal to G = 2 . The mean shapes in each cluster were taken from the rat calvarial dataset [1] corresponding to the skull midsagittal section of 21 rats collected at ages of 7 and 14 days. In the isotropic case, σ was equal to 13. In the heteroscedastic case, the values of σ 1 , σ 2 , , σ K were equal to 13 for 3 randomly chosen landmarks and equal to 1.3 for the remaining 5 landmarks of each shape. Finally, the anisotropic case was simulated by setting σ x = 13 and σ y = 1.3 in one cluster and σ x = 1.3 and σ y = 13 in the other cluster.
Examples of the simulated data under the different covariance structures are provided in Figure 1, Figure 2 and Figure 3. In the isotropic case (Figure 1), all landmark coordinates exhibit the same independent spherical variation around the mean. In the heteroscedastic case (Figure 2), the spherical variation is allowed to vary between the landmarks and the clusters. Finally, Figure 3 shows the anisotropic case where the variability of the landmark coordinates is different in the horizontal and vertical directions.
The shape K-means algorithm is implemented by using:
  • The Fisher–Rao distance under the round Gaussian model representation (dr) where the variance is considered as a free parameter, which is isotropic across all the landmarks
  • The Fisher–Rao distance under the diagonal Gaussian model representation (dd1 (Type I K-means algorithm) and dd2 (Type II K-means algorithm))
  • Wasserstein distance (dp)
For each covariance structure, we simulated 150 samples and, for each sample, we computed the adjusted Rand index [25] of each clustering method. The adjusted Rand index is a measure of agreement between two partitions. It ranges from about 0 when the compared partitions are completely random to 1 when they are equal. The index is generally used as a measure of cluster recovery, the closer the index to 1 the better the clustering results.
Figure 4, Figure 5 and Figure 6 display the boxplots of the adjusted Rand index over 150 simulated samples for each clustering method. Outliers are plotted individually with the + symbol. When the covariance structure is isotropic (Figure 4), all distances show a similar behavior. In particular, the Fisher–Rao distance with round Gaussian distribution (dr) and the Wasserstein distance (dw) yield the best clustering results with median values of the adjusted Rand index both equal to 0.96 versus 0.85 and 0.88 obtained by the diagonal Gaussian distribution with Type I (dd1) and Type II (dd2) algorithms, respectively. In the heteroscedastic setting (Figure 5), both the Fisher–Rao with the round Gaussian distribution (median adjusted Rand index equal to 0.54 ) and the Wasserstein distance (median adjusted Rand index equal to 0.43 ) perform poorly in comparison to the Fisher–Rao distance based on the diagonal distribution. As expected, the models which take into account different landmark variances (dd1-Type I algorithm) and also differences in the variances between the clusters (dd2-Type II algorithm) show a very good behavior with median values of the adjusted Rand index equal to 0.84 and 1, respectively. A very similar pattern is observed when anisotropy is also added in the covariance structure (Figure 6). As expected, the Type II algorithm significantly increases the computational time needed to meet the convergence condition since in each iteration both the means and the variances of the cluster centers have to be updated. On average, the Type II algorithm’s running time is approximately 20 times longer than Type I.

6. Conclusions

In this study, Information Geometry was used as a useful tool in the area of shape clustering. We first described a shape representing each landmark by a Gaussian model using the mean and the variance as coordinates, reflecting the geometrical shape of the configuration and the variability across a family of patterns, respectively. Within this framework, we considered the Fisher–Rao and the Wasserstein metric for quantifying the difference between two shapes.
Two version of the Fisher–Rao metric were proposed, depending on how the variances in the data are employed. In one case (round Gaussian distribution model), the variance was considered a free parameter that is isotropic across all the landmarks. In the second case, the isotropic assumption was relaxed allowing the variances to vary among the landmarks (diagonal Gaussian distribution model).
The results of the numerical study have shown that the violation of the isotropic assumption on the landmarks variability may cause a severe loss in the clustering recovery. Indeed, this assumption is rarely satisfied in practice where it is regularly seen that landmarks have different variances. In such a case, the relative importance among landmarks must be taken into account in the similarity measure adopted in the clustering algorithm. The proposed geodesic distance under the diagonal Gaussian model representation is able to face this problem. A further assumption that may be violated is that in all clusters the landmarks coordinates have a common covariance matrix. To cope with this issue, a new K-means shape algorithm was implemented that allows for differences among the clusters in the landmark coordinates variability.
Other extensions of the current work deserve further investigation, for example, the use of geodesics in the case of the general multivariate Gaussian model and considering more general shape measures, such as α -divergences.

Author Contributions

Although all authors contributed equally to the research, in particular A.D.S. proposed the idea and the methodology, S.P. and F.N. improved the theoretical derivation and S.A.G. implemented the algorithms and performed the numerical calculations.

Funding

This research received no external funding.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Bookstein, F.L. Morphometric Tools for Landmark Data: Geometry and Biology; Cambridge University Press: Cambridge, UK, 1991. [Google Scholar]
  2. Kendall, D.G. Shape manifolds, Procrustean metrics and complex projective spaces. Bull. Lond. Math. Soc. 1984, 16, 81–121. [Google Scholar] [CrossRef]
  3. Cootes, T.; Taylor, C.; Cooper, D.H.; Graham, J. Active shape models-their training and application. Comput. Vis. Image Underst. 1995, 61, 38–59. [Google Scholar] [CrossRef]
  4. Dryden, I.L.; Mardia, K.V. Statistical Shape Analysis; John Wiley & Sons: London, UK, 1998. [Google Scholar]
  5. Stoyan, D.; Stoyan, H. A further application of D.G. Kendall’s Procrustes Analysis. Biom. J. 1990, 32, 293–301. [Google Scholar] [CrossRef]
  6. Amaral, G.; Dore, L.; Lessa, R.; Stosic, B. K-means Algorithm in Statistical Shape Analysis. Commun. Stat. Simul. Comput. 2010, 39, 1016–1026. [Google Scholar] [CrossRef]
  7. Lele, S.; Richtsmeier, J. An Invariant Approach to Statistical Analysis of Shapes; Chapman & Hall/CRC: New York, NY, USA, 2001. [Google Scholar]
  8. Huang, C.; Martin, S.; Zhu, H. Clustering High-Dimensional Landmark-based Twodimensional Shape Data. J. Am. Stat. Assoc. 2015, 110, 946–961. [Google Scholar] [CrossRef] [PubMed]
  9. Kume, A.; Welling, M. Maximum likelihood estimation for the offset-normal shape distributions using EM. J. Comput. Gr. Stat. 2010, 19, 702–723. [Google Scholar] [CrossRef]
  10. Gattone, S.A.; De Sanctis, A.; Russo, T.; Pulcini, D. A shape distance based on the Fisher–Rao metric and its application for shapes clustering. Phys. A Stat. Mech. Appl. 2017, 487, 93–102. [Google Scholar] [CrossRef]
  11. De Sanctis, A.; Gattone, S.A. A Comparison between Wasserstein Distance and a Distance Induced by Fisher-Rao Metric in Complex Shapes Clustering. Proceedings 2018, 2, 163. [Google Scholar] [CrossRef]
  12. Amari, S.; Nagaoka, H. Translations of Mathematical Monographs. In Methods of Information Geometry; AMS & Oxford University Press: Providence, RI, USA, 2000. [Google Scholar]
  13. Murray, M.K.; Rice, J.W. Differential Geometry and Statistics; Chapman & Hall: London, UK, 1984. [Google Scholar]
  14. Srivastava, A.; Joshi, S.H.; Mio, W.; Liu, X. Statistical Shape analysis: Clustering, learning, and testing. IEEE Trans. PAMI 2005, 27, 590–602. [Google Scholar] [CrossRef] [PubMed]
  15. Mio, W.; Srivastava, A.; Joshi, S.H. On Shape of Plane Elastic Curves. Int. J. Comput. Vis. 2007, 73, 307–324. [Google Scholar] [CrossRef]
  16. Skovgaard, L.T. A Riemannian geometry of the multivariate normal model. Scand. J. Stat. 1984, 11, 211–223. [Google Scholar]
  17. Costa, S.; Santos, S.; Strapasson, J. Fisher information distance: A geometrical reading. Discret. Appl. Math. 2015, 197, 59–69. [Google Scholar] [CrossRef] [Green Version]
  18. Pennec, X.; Fillard, P.; Ayache, N. A Riemannian Framework for Tensor Computing. Int. J. Comput. Vis. 2006, 66, 41–66. [Google Scholar] [CrossRef] [Green Version]
  19. Bishop, R.L.; O’Neill, B. Manifolds of negative curvature. Trans. Am. Math. Soc. 1969, 145, 1–49. [Google Scholar] [CrossRef]
  20. Takatsu, A. Wasserstein geometry of Gaussian measures. Osaka J. Math. 2011, 48, 1005–1026. [Google Scholar]
  21. Amari, S. Applied Mathematical Sciences. In Information Geometry and Its Applications; Springer: Berlin, Germany, 2016. [Google Scholar]
  22. Banerjee, A.; Merugu, S.; Dhillon, I.; Ghosh, J. Clustering with Bregman Divergence. J. Mach. Learn. Res. 2005, 6, 1705–1749. [Google Scholar]
  23. Bookstein, F.L. Size and shape spaces for landmark data in two dimensions. Stat. Sci. 1986, 1, 181–242. [Google Scholar] [CrossRef]
  24. Goodall, C.R. Procrustes methods in the statistical analysis of shape. J. R. Stat. Soc. 1991, 53, 285–339. [Google Scholar]
  25. Hubert, L.; Arabie, P. Comparing Partitions. J. Classif. 1985, 2, 193–218. [Google Scholar] [CrossRef]
Figure 1. Independent spherical variation around each mean landmark.
Figure 1. Independent spherical variation around each mean landmark.
Entropy 20 00647 g001
Figure 2. Heteroscedastic variation around each mean landmark.
Figure 2. Heteroscedastic variation around each mean landmark.
Entropy 20 00647 g002
Figure 3. Anisotropy in the x and y directions around each mean landmark.
Figure 3. Anisotropy in the x and y directions around each mean landmark.
Entropy 20 00647 g003
Figure 4. Isotropic case: Boxplots of the adjusted Rand index over 150 simulated samples for each clustering method; aRand index median values are 0.96 (dr), 0.85 (dd1), 0.88 (dd2), 0.96 (dw).
Figure 4. Isotropic case: Boxplots of the adjusted Rand index over 150 simulated samples for each clustering method; aRand index median values are 0.96 (dr), 0.85 (dd1), 0.88 (dd2), 0.96 (dw).
Entropy 20 00647 g004
Figure 5. Heteroscedastic case: Boxplots of the adjusted Rand index over 150 simulated samples for each clustering method; aRand index median values are 0.54 (dr), 0.84 (dd1), 1.00 (dd2), and 0.43 (dw).
Figure 5. Heteroscedastic case: Boxplots of the adjusted Rand index over 150 simulated samples for each clustering method; aRand index median values are 0.54 (dr), 0.84 (dd1), 1.00 (dd2), and 0.43 (dw).
Entropy 20 00647 g005
Figure 6. Anisotropic case: Boxplots of the adjusted Rand index over 150 simulated samples for each clustering method; aRand index mean values are 0.51 (dr), 0.79 (dd1), 1 (dd2), and 0.44 (dw).
Figure 6. Anisotropic case: Boxplots of the adjusted Rand index over 150 simulated samples for each clustering method; aRand index mean values are 0.51 (dr), 0.79 (dd1), 1 (dd2), and 0.44 (dw).
Entropy 20 00647 g006

Share and Cite

MDPI and ACS Style

Gattone, S.A.; De Sanctis, A.; Puechmorel, S.; Nicol, F. On the Geodesic Distance in Shapes K-means Clustering. Entropy 2018, 20, 647. https://doi.org/10.3390/e20090647

AMA Style

Gattone SA, De Sanctis A, Puechmorel S, Nicol F. On the Geodesic Distance in Shapes K-means Clustering. Entropy. 2018; 20(9):647. https://doi.org/10.3390/e20090647

Chicago/Turabian Style

Gattone, Stefano Antonio, Angela De Sanctis, Stéphane Puechmorel, and Florence Nicol. 2018. "On the Geodesic Distance in Shapes K-means Clustering" Entropy 20, no. 9: 647. https://doi.org/10.3390/e20090647

APA Style

Gattone, S. A., De Sanctis, A., Puechmorel, S., & Nicol, F. (2018). On the Geodesic Distance in Shapes K-means Clustering. Entropy, 20(9), 647. https://doi.org/10.3390/e20090647

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop