Next Article in Journal
On a New Result on the Ratio Exponentiated General Family of Distributions with Applications
Next Article in Special Issue
Asymptotic Results in Broken Stick Models: The Approach via Lorenz Curves
Previous Article in Journal
Riccati Technique and Asymptotic Behavior of Fourth-Order Advanced Differential Equations
Previous Article in Special Issue
Theoretical Aspects on Measures of Directed Information with Simulations
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Partitions Subspace Clustering

by
Vincent Vandewalle
1,2
1
Biostatistics Department, Univ. Lille, CHU Lille, ULR 2694—METRICS: Évaluation des Technologies de Santé et des Pratiques MéDicales, F-59000 Lille, France
2
Inria Lille—Nord Europe, 59650 Villeneuve d’Ascq, France
Mathematics 2020, 8(4), 597; https://doi.org/10.3390/math8040597
Submission received: 28 February 2020 / Revised: 6 April 2020 / Accepted: 10 April 2020 / Published: 15 April 2020
(This article belongs to the Special Issue Probability, Statistics and Their Applications)

Abstract

:
In model based clustering, it is often supposed that only one clustering latent variable explains the heterogeneity of the whole dataset. However, in many cases several latent variables could explain the heterogeneity of the data at hand. Finding such class variables could result in a richer interpretation of the data. In the continuous data setting, a multi-partition model based clustering is proposed. It assumes the existence of several latent clustering variables, each one explaining the heterogeneity of the data with respect to some clustering subspace. It allows to simultaneously find the multi-partitions and the related subspaces. Parameters of the model are estimated through an EM algorithm relying on a probabilistic reinterpretation of the factorial discriminant analysis. A model choice strategy relying on the BIC criterion is proposed to select to number of subspaces and the number of clusters by subspace. The obtained results are thus several projections of the data, each one conveying its own clustering of the data. Model’s behavior is illustrated on simulated and real data.

1. Introduction

In exploratory data analysis, the statistician often uses clustering and visualization in order to improve his knowledge on the data. In visualization he looks for some principal components explaining some major characteristics of the data. For example in principal component analysis (PCA) the goal is to find a linear combination of the variables explaining the major variability of the data. In cluster analysis, the goal is to find some clusters explaining the major heterogeneity of the data. In this article we suppose that the data can contain several clustering latent variables, that is, we are in the multiple partition setting, and we are simultaneously looking for clustering subspaces, that is, linear projections of the data each one related to some clustering latent variable thus the developed model is later called multi-partitions subspace clustering. A solution to perform multi-partition subspace clustering is to use a probabilistic model on the data such as a mixture model [1], it allows to perform the parameters estimation, and model selection such as the choice of the number of subspaces and the number of clusters per subspace using standard model choice criteria such as BIC [2]. Thus the main fields related to our work are model based subspace clustering and multi-partitions clustering.
In the model based subspace clustering framework, let first notice that PCA can be re-interpreted in a probabilistic way by considering a parsimonious version of a multivariate Gaussian distribution [3] and that the k-means algorithm can be re-interpreted as a particular parsimonious Gaussian mixture model estimated using a classification EM algorithm [4]. A re-interpretation of the probabilistic PCA has also been used in clustering by Bouveyron et al. [5] in order to cluster high-dimensional data. Although the proposed high dimensional mixture does not performs dimension reduction, it rather operates a class per class dimension reduction which does not allow to have a global model-based data visualization. Thus Bouveyron and Brunet [6] proposed the so called Fisher-EM algorithm which simultaneously performs clustering and dimension reduction. This is performed through a modified version of the EM algorithm [7] by including a Fisher step between the E and the M step. This approach allows the same projection to be applied to all data, but does not guarantee the increasing of the likelihood at each iteration of the algorithm.
In the context of multi-partitions clustering, Galimberti and Soffritti [8] assumed that the variables can be partitioned into several independent blocks, each one following a full-covariance Gaussian mixture model. The model selection was done by maximizing the BIC criterion by a forward/backward approach. Then, Galimberti et al. [9] generalized thier previous work by relaxing the assumption of block independence. The proposed extension takes into account three types of variables, classifying variables, redundant variables and non-classifying variables. In this context, the choice of the model is difficult because several roles have to be taken into account for each variable, which requires a lot of calculations, even for the reallocation of only one variable. Poon et al. [10] also took into account the multi-partition setting, called as facet determination in their article. The model considered is similar to that of Galimberti and Soffritti [8], but it also allows tree dependency between latent class variables, resulting in the Pouch Latent Tree Models (PLTM). Model selection is performed by a greed search to maximize the BIC criterion. The resulting model allows a broad understanding of the data, but the tree structure search makes estimation even more difficult as the number of variables increases. More recently, Marbac and Vandewalle [11] proposed a tractable muti-partition clustering algorithm not limited to continuous data; in the Gaussian setting it can be seen as particular case of Galimberti and Soffritti [8] where they assume a diagonal covariance matrix allowing a particularly efficient search of the partition of the variables in sub-vectors.
In this article we suppose that the data can contain several clustering latent variables, that is we are in the multiple partition setting. But contrary to Marbac and Vandewalle [11] where it is assumed that variables are divided into blocks each one related to some clustering of the data, we are looking for clustering subspaces, i.e., linear projections of the data each one related to some particular clustering latent variable thus replacing the combinatorial question of finding the partition of the variables in independent sub-vectors by the question of finding the coefficients of the linear combinations. The proposed approach can be related to the independent factor analysis [12] where the author deals with source separation, in our framework a source can be interpreted as some specific clustering subspace; however, their approach becomes intractable as the numbers of sources increases and does not allow to consider multivariate subspaces. Moreover, it is not invariant up to a rotation and rescaling of the data, where our proposed methodology is.
The organisation of the paper is the following, in Section 2 we present a reinterpretation of the factorial discriminant analysis as a search of discriminant components and of independent non-discriminant components. In Section 3, the multi-partitions subspace clustering model and the EM algorithm to estimate the parameters of the model will be presented. In Section 4, results on simulated and real data will show the interest of the method in practice. In Section 5, a conclusion and discussion of future extension of the paper will be made.

2. Probabilistic Interpretation of the Factorial Discriminant Analysis

2.1. Linear Discriminant Analysis (LDA)

It is supposed that n quantitative data in dimension d are available, the data number i will be denoted by x i = ( x i 1 , , x i d ) T , where x i j is the value of variable j of data i. The whole dataset will be denoted by x = ( x 1 , , x n ) T . Let assume that the data is clustered in K clusters, the class label of data i will be denoted by z i = ( z i 1 , , z i K ) T , with z i k equals to 1 if data i belongs to cluster k and 0 otherwise. Let also denote by z = ( z 1 , , z n ) the partition of x . In this section z is supposed to be known. For sake of simplicity the random variables and their realisations will be denoted in lower case, and p will be used as a generic notation to denote a probability distribution function (p.d.f.) which will be interpreted according to its arguments.
In the context of linear discriminant analysis [13], it is supposed that the distribution x i given the cluster follows a d-variate Gaussian distribution with common covariance matrices:
k { 1 , , K } , x i | z i k = 1 N d ( μ k , Σ ) ,
with μ k the vector of means in cluster k and Σ the common class conditional covariance matrix. Let also denote by π k = p ( z i k = 1 ) the prior weights of each cluster.
The posterior cluster membership probabilities can be computed using the Bayes formula:
p ( z i k = 1 | x i ) = π k ϕ d ( x i ; μ k , Σ ) k = 1 K π k ϕ d ( x i ; μ k , Σ ) ,
where ϕ d ( · ; μ k , Σ ) stands for the p.d.f. of the d-variable Gaussian distribution with expectation μ k and covariance matrix Σ .
Let x ¯ k denote the class conditional mean in cluster k:
x ¯ k = 1 n k i = 1 n z i k x i ,
with n k = i = 1 n z i k the number of data in cluster k, and by x ¯ the unconditional mean. Let also denote by W the empirical intra-class covariance matrix:
W = 1 n k = 1 K i = 1 n z i k ( x i x ¯ k ) ( x i x ¯ k ) T ,
and by B the empirical between class covariance matrix:
B = 1 n k = 1 K n k ( x ¯ k x ¯ ) ( x ¯ k x ¯ ) T .
If the data are supposed to be independent, the likelihood can simply be written as:
( π 1 , , π K , μ 1 , , μ K , Σ ; x , z ) = n 2 log ( det ( Σ ) ) 1 2 i = 1 n k = 1 K z i k x i μ k Σ 1 2 + i = 1 n k = 1 K z i k log ( π k ) n 2 log ( 2 π ) .
The maximum likelihood estimators of the parameters of π k , μ k and Σ are π ^ k = n k n , μ ^ k = x ¯ k and Σ ^ = W . A new data point can be then classified by plugin the estimated values of the parameters in Equation (1):
z ^ i = argmax k { 1 , , K } μ ^ k T Σ ^ 1 x i 1 2 μ k T Σ ^ 1 μ k + log ( π ^ k ) ,
the resulting classification boundary being linear in this case.

2.2. Factorial Discriminant Analysis (FDA)

Let us note that, from a descriptive viewpoint, one can be interested in dimension reduction in order to visualize the data. This could be done by using PCA, but from a classification perspective the component explaining the largest variability in the data are often not the same that the components providing the best separation between the clusters.
The goal of factorial discriminant analysis (FDA) is to find the component maximizing the variance explained by the cluster above the intra-class variance. The coefficients of the first discriminant component v 1 R d is defined by
v 1 = arg max v R d v T B v v T W v .
It is well known that v 1 is the eigen vector associated with the highest eigen value λ 1 of W 1 B [14]. The remaining discriminant components are obtained through the remaining eigen vectors of W 1 B . Let denote by λ 1 , , λ K 1 the eigen values of W 1 B sorted in decreasing order and by v 1 , , v K 1 the associated eigen vectors. Moreover, if each component is constrained to have an intra-class variance equal to one (i.e., v k T W v k = 1 , k { 1 , , K 1 } ), the classification obtained using the Mahalonobis distance can simply be obtained by using the Euclidean distance on the data projected on the discriminant components.

2.3. Equivalence between LDA and FDA

As proved in Campbell [15] and detailed in Trevor Hastie [16], the FDA can be interpreted in a probabilistic way as an LDA where the rank of { μ 1 , , μ K } is constrained to be equal to p with p K 1 under the common class covariance matrix assumption. This allows us to reparametrize the probabilistic model in the following way:
x i | z i k = 1 N d ( A ( ν k γ ) , A A T ) ,
where ν k R p , γ R d p and A M d , d ( R ) . Let notice that this new parametrization at this step is not unique but the model can easily be made identifiable by imposing some constraints on the parameters.
Let y i R p and u i R d p two random variables, the new parametrization can be reinterpreted in the following generative framework:
  • Draw z i : z i M ( 1 ; π 1 , , π K ) where M stands for the multinomial distribution
  • Draw y i | z i : y i | z i k = 1 N p ( ν k , I p )
  • Draw u i : u i N d p ( γ , I d p )
  • Compute x i based on y i and u i : x i = A ( y i u i ) .
Thus the p.d.f. of x i can be factorized in the following way:
p ( x i ) = 1 | A | p ( u i ) p ( y i ) .
From a graphical angle the model can be reinterpreted as in Figure 1, where u i and y i are latent random variables.
In practice we are interested in finding y i and u i from x i . We will denote by V M p , d ( R ) and R M d p , d ( R ) the matrices which allow this computation: y i = V x i , u i = R x i . It is obvious that ( V R ) = A 1 , for the rest of the article only the parametrization in terms of V and R will be used.
The main interest of this parametrization is that
p ( z i k = 1 | x i ) = p ( z i k = 1 | y i , u i ) = p ( z i k = 1 | y i ) = p ( z i k = 1 | V x i ) .
It means that only V x i is required to compute the posterior class membership probabilities:
p ( z i k = 1 | x i ) = π k ϕ p ( V x i ; ν k , I p ) k = 1 K π k ϕ p ( V x i ; ν k , I p ) .
Parameters are estimated by maximum likelihood, using the variable change formulae the likelihood can be written:
( π 1 , , π K , V , R , γ , ν 1 , , ν K ; x , z ) = n log det ( V R ) 1 2 i = 1 n k = 1 K z i k V x i ν k 2 + i = 1 n k = 1 K z i k log ( π k ) 1 2 i = 1 n R x i γ 2 n 2 log ( 2 π ) .
The first term is related to the variable change, the second term is related to the discriminant components and the third term is related to prior class membership probabilities and the fourth term is related to the non-discriminant components. In practice, this decomposition is the corner stone of the proposed approach since it separates the clustering part from the non-clustering part.
As stated in Campbell [15] the maximum likelihood estimator of V are the first p eigen vectors v 1 , , v p of W 1 B in rows:
V ^ = ( v 1 T v p T ) ,
renormalized such that:
V ^ W V ^ T = I p .
The maximum likelihood estimator of R is obtained such that:
R ^ W V ^ T = 0
and that
R ^ W R ^ T = I d p ,
with D d p the diagonal matrix of the d p last eigen-values of W 1 B (with the d K last eigen-values which are null). R ^ T can simply be obtained by multiplying the last d p egien-vectors of W 1 / 2 B W 1 / 2 by W 1 / 2 then renormalize them such that Equation (3) is satisfied. Such renormalization makes the parameters V et R identifiable up to a sign.
Moreover ν k and γ are estimated by
ν ^ k = 1 n k i = 1 n z i k V ^ x i ,
γ ^ = 1 n i = 1 n R ^ x i .
As stated in Trevor Hastie [16] the link between the two parametrizations is as follows:
μ ^ k = W V ^ T R ^ ( x ¯ k x ¯ ) + x ¯ ,
and
Σ ^ = W + W R ^ T R ^ B R ^ T R ^ W .
From this formula we can clearly see that the reduced rank constraint operates some regularization on the parameters estimation. Moreover, from a practical perspective the reparametrization Equation (2) is more efficient for computing the posterior class membership probabilities.

2.4. Application in the Clustering Setting

As in Trevor Hastie [16], the model can easily be used in the clustering setting using the EM algorithm to maximise the likelihood. Thus W , B , x ¯ k are recomputed at each iteration by using their version weighted by their posterior membership probabilities.
The EM algorithm is now presented. The algorithm is first initialized with some starting value of the parameters or of the partition. Then the E step and the M step are iterated until convergence. Let ( r ) denote the value of the parameters at iteration r, the E and the M steps are the followings:
  • E step: compute the posterior class membership probabilities.
    t i k ( r + 1 ) = π k ϕ d ( x i ; μ k ( r ) , Σ ( r ) ) k = 1 K π k ϕ d ( x i ; μ k ( r ) , Σ ( r ) ) = π k ϕ p ( y i ( r ) ; ν k ( r ) , I p ) k = 1 K π k ϕ p ( y i ( r ) ; ν k ( r ) , I p ) ,
    with y i ( r ) = V ( r ) x i .
  • M step: Compute
    x ¯ k ( r + 1 ) = 1 n k ( r + 1 ) i = 1 n t i k ( r + 1 ) x i ,
    W ( r + 1 ) = 1 n k = 1 K i = 1 n t i k ( r + 1 ) ( x i x ¯ k ( r + 1 ) ) ( x i x ¯ k ( r + 1 ) ) T ,
    B ( r + 1 ) = 1 n k = 1 K n k ( r + 1 ) ( x ¯ k ( r + 1 ) x ¯ ) ( x ¯ k ( r + 1 ) x ¯ ) T .
    Then deduce V ( r + 1 ) , R ( r + 1 ) , ν 1 ( r + 1 ) , , ν K ( r + 1 ) and γ ( r + 1 ) as in the previous section using the eigenvalue decomposition of W ( r + 1 ) 1 B ( r + 1 ) .
As noticed in Trevor Hastie [16], this approach is not equivalent to performing a standard EM algorithm and then performing FDA at the end of the EM algorithm. FDA must be computed at each iteration of the EM algorithm since the posterior membership probabilities are only computed based on the p first clustering projections.
Let us also notice that the M step can be interpreted in a Fisher-M step since FDA is required. In this sense it can be interpreted as a particular version of the Fisher-EM algorithm of Bouveyron and Brunet [6]. Although the homoscedasticity could be seen as particularly constraining, it is the best framework for introducing our model in the next section, since it is easily interpretable and allows for efficient computation owing to the closed form of FDA. This limitation could be easily overcome by using rigorous extensions of the FDA in the heteroscedastic setting as in Kumar and Andreou [17]; however, the computation would be much more intensive, since in this case no closed form formula is available and an iterative algorithm would be required even for finding the best projection.

3. Multi-Partition Subspace Mixture Model

3.1. Presentation of the Model

Let us now suppose that instead of having only one class variable z i for data i, we now have H class variables z i 1 , , z i H with K 1 , , K H modalities. It is assumed that z i 1 , , z i H are independent, with p ( z i k h = 1 ) denoted by π k h . Let also denote by y i h the variables related to the clustering variable z i h such that:
y i h | z i k h = 1 N p h ( ν k h , I p h )
and that we will denote by p = h = 1 H p h .
Let us still denote by u i the non clustering variables
u i N d p ( γ , I d p ) .
Let us also define x i by:
x i = V 1 V H R 1 y i 1 y i H u i .
Thus,
p ( x ) = det V 1 V H R p ( u ) h = 1 H p ( y h ) .
The Figure 2 illustrates the model in the case of H = 2 .
Let us notice that this model allows us to visualize many clustering viewpoints in a low dimensional space since x i can be summarized by y i 1 , , y i H . For instance, one can suppose that p 1 = = p H = 1 . In this case each clustering variable can be visualized on one component. We will denote by θ = ( V 1 , , V H , R , γ , ν 1 1 , , ν K H H ) the parameters of the model to be estimated.

3.2. Discussion about the Model

The Cartesian product of cluster spaces results in h = 1 H K h clusters, which can be very large without needing many parameters. Thus the proposed model can be interpreted as being a very sparse Gaussian mixture model allowing the possibility to deal with a very large number of clusters, the resulting conditional means and covariances matrices are given in the following formulas:
E ( x i | z i k 1 1 = 1 , z i k 1 2 = 1 , , z i k H H = 1 ) = V 1 V H R 1 ν k 1 1 ν k H H γ ,
and
V ( x i | z i k 1 1 = 1 , z i k 1 2 = 2 , , z i k H H = 1 ) = ( V 1 T V 1 + + V H T V H + R T R ) 1 .
Thus, the expectation of x i given in all the clusters is a linear combination of the cluster specific means which can be referred to as a multiple-way MANOVA setting. On the one hand, as a particular homoscedastic Gaussian mixture, our model is more prone to model bias than free homoscedastic Gaussian mixture, and in the case when our model would be well-specified the homoscedastic Gaussian mixture would give a similar clustering for a large sample size (i.e., the same partitions with respect to the partition resulting from the product space of our multi-partitions model). On the other hand, our approach produces a factorised version of the partition space as well as the related clustering subspaces which is not a standard output of clustering methods, and it can deal with a large number of clusters in a sparse way which can be particularly useful for a moderated sample size. In practice, the choice between our model and an other mixture model can simply be performed through the BIC criterion.
In some sense our model can be linked with the mixture of factor analyzers [18]. In mixture of factor analyzers the model is of the type:
x i = A y i + u i ,
where A is a low rank matrix. But here we have chosen a model of the type
x i = A y i u i ,
which allows us to deal with the noise in a different way. Actually, our model is invariant up to a bijective linear transformation of the data which is not the case for the mixtures of factor analyzers. On the other hand, our model can only deal with data with moderated dimension with respect to the number of statistical units; it assumes that the sources y i can be recovered from the observed data x i .

3.3. Estimation of the Parameters of the Model in the Supervised Setting

The likelihood of the model can be written:
( θ ; x , z ) = n log det V 1 V H R i = 1 n h = 1 H k = 1 K h z i k h V h T x i ν k h 2 + i = 1 n h = 1 H k = 1 K h z i k h log ( π k h ) i = 1 n R T x i γ 2 n 2 log ( 2 π ) .
The likelihood cannot be maximized directly. However, in the case of H = 1 , it reduces to the problem of Section 2. Let notice that if all the parameters are fixed except V h and R , ν k h and γ , the optimisation can be easily performed by constraining V h ( r + 1 ) and R ( r + 1 ) to be linear combinations of V h ( r ) and R ( r ) . Thus the likelihood will be optimized by using an alternate optimization algorithm. Let M M d p + p h , d p + p h ( R ) the matrix which allow to compute V h ( r + 1 ) and R ( r + 1 ) based on V h ( r ) and R ( r ) :
V h ( r + 1 ) R ( r + 1 ) = M V h ( r ) R ( r ) = M 1 M 2 V h ( r ) R ( r ) ,
where M 1 is the sub-matrix containing the p h first rows of M and M 2 the matrix containing the last d p rows of M .
Thus, the increase of the likelihood when all the parameters are fixed except V h , R , ν k h and γ becomes:
C ( M 1 , M 2 , ν k , γ ) = n log | det ( M ) | 1 2 i = 1 n k = 1 K h z i k h M 1 V h ( r ) R ( r ) x i ν k h 2 1 2 i = 1 n M 2 V h ( r ) R ( r ) T x i γ 2 .
By denoting
y i h ( r ) u i ( r ) = V h ( r ) R ( r ) x i ,
we have to maximize:
C ( M 1 , M 2 , ν k , γ ) = n log | det ( M ) | 1 2 i = 1 n k = 1 K h z i k h M 1 y i h ( r ) u i ( r ) ν k h 2 1 2 i = 1 n M 2 y i h ( r ) u i ( r ) γ 2 .
Consequently, M and the others parameters can be obtained by applying a simple FDA on the data ( y i h ( r ) T , u i ( r ) T ) . In order to optimise over all the parameters, we can loop over all the clustering dimensions.
Thus, in the case of mixed continuous and categorical data, this model can be used to visualize the clustering behavior of the categorical variables with respect to the quantitative ones.

3.4. Estimation of the Parameters of the Model in the Clustering Setting

Here our main goal is to consider the clustering setting, that is, when z i 1 , , z i H are unknown. Consequently we will use an EM algorithm to “reconstitute the missing label” in order to maximize the likelihood. Thus the algorithm stays the same as in the supervised setting, except that the data at each iteration are now weighted by t i k h ( r + 1 ) instead of z i k h .
The algorithm is the following:
  • Until convergence, for h { 1 , , H } iterates the following steps:
    -
    E step: compute
    t i k h ( r + 1 ) = π k p ( y i h ( r ) ; ν k h ( r ) , I p ) k = 1 K π k p ( y i h ( r ) ; ν k h ( r ) , I p ) .
    -
    M step: compute π 1 h ( r + 1 ) , , π K h h ( r + 1 ) , V h ( r + 1 ) , R ( r + 1 ) , γ ( r + 1 ) and ν k h ( r + 1 ) based on formulas given in the supervised setting.

3.5. Parsimonious Models and Model Choice

The proposed model needs the user to define the number of clustering subspaces H, the number of cluster in each clustering subspace K 1 , …, K H , and the dimensionality p 1 , …, p H of each subspace. The constraints are that H < d , that p h K h 1 and p = p 1 + + p H < d . It is clear that the number of possible models can become very high. To limit the combinatorial aspect, one can impose K 1 = = K H = K and/or p 1 = = p h = p . In practice the choice of p = 1 enforces to find clustering which could be visualized in one dimension, which can help the practitioner. Moreover, choosing K = 2 is the minimal requirement in order to investigate a clustering structure. However, if possible we recommend to explore the largest possible number of models and choosing the best one with the BIC. Let us define the following parsimonious models and their related number of parameters:
  • [ K h p h ] the general form of the proposed model, the index h will be removed if values are the same for each clustering subspace.
    h = 1 H ( K h 1 ) + h = 1 H p h ( 2 K h + 2 d p h + 1 ) 2 + ( d p ) ( d + p + 3 ) 2 ,
  • [ K p h ] where the number of clusters is the same for each subspace
    H ( K 1 ) + h = 1 H p h ( 2 K + 2 d p h + 1 ) 2 + ( d p ) ( d + p + 3 ) 2 ,
  • [ K h p ] where the dimensionalities are the same for each subspaces
    h = 1 H ( K h 1 ) + h = 1 H p ( 2 K h + 2 d p + 1 ) 2 + ( d H p ) ( d + H p + 3 ) 2 ,
  • [ K h 1 ] where the dimensionalities are equals to one for each subspaces
    h = 1 H ( K h 1 ) + h = 1 H ( K h + d ) + ( d H ) ( d + H + 3 ) 2 ,
  • [ K p ] where the number of clusters is the same for each subspace and the dimensionalities are the same for each subspace
    H ( K 1 ) + H p ( 2 K + 2 d p + 1 ) 2 + ( d H p ) ( d + H p + 3 ) 2 ,
  • [ K 1 ] where the number of clusters is the same for each subspace and the dimensionalities are equals to one for each subspaces
    H ( K 1 ) + H ( K + d ) + ( d H ) ( d + H + 3 ) 2 .
For a given model m the BIC is computed as:
B I C ( m ) = ( θ ^ m ; x ) ν m 2 log n ,
where ν m is the number of parameters of the model detailed above. Thus the model choice consists of choosing the model maximising the BIC. BIC enjoys good theoretical consistency properties, thus providing a guarantee to select the true model as the number of data increases. The ICL criterion [19] could also be used to enforce the choice of well separated clusters, since from a classification perspective BIC is known to over-estimate the number of clusters if model assumption are violated. Let us however notice that in practice the user could be mainly interested by a low value of H, since even H = 2 can provide him with new insights about his data, focusing on finding several clustering view points.

4. Experiments

4.1. Experiments on Simulated Data

We now present a tutorial example. Let us consider n = 100 data, with H = 2 clustering subspaces each one of dimension one ( p 1 = p 2 = 1 ) and containing each of two clusters ( K 1 = K 2 = 2 ). Let us draw y i 1 , the first clustering variable, according to a mixture of two Gaussian univariate distributions: y i 1 0.5 N ( 0 , 1 ) + 0.5 N ( 4 , 1 ) and draw independently y i 2 , the second clustering variable, according to the same mixture distribution: y i 2 0.5 N ( 0 , 1 ) + 0.5 N ( 4 , 1 ) , then draw u i the non-classifying components according to a multivariate Gaussian distribution in R 4 , u i N 4 ( 0 , I 4 ) . Finally compute x i based on the formula: x i = A y i 1 y i 2 u i , where the 36 entries of the matrix A have been drawn according to independent N ( 0 , 1 ) for sake of simplicity.
Thus the proposed model, based on the observed data x , aims at recovering the clustering variables y 1 and y 2 as well as the associated cluster variables z 1 and z 2 . The initial data x are presented on Figure 3, we see that the clustering structure of the data is not apparent from these scatter plots. The underlying clustering variables are presented on Figure 4, where we see the separation of the colors on the first clustering variable Y 1 and separation of the shapes on the second clustering variable Y 2 . Let us notice that such factorization gives a more synthetic view of the clustering than seeing these clusters as four clusters. Using standard dimension reduction techniques such as PCA does not succeed in recovering the clustering subspace see Figure 5. Performing a factorial discriminant analysis considering the four clusters in the supervised setting we get Figure 6, it finds good separation between the clusters, however we do not obtain the factorised interpretation of the clustering as the Cartesian product of two independent clusterings. Finally by performing the estimation of the model parameters in an unsupervised setting based on the data x we get Figure 7. We see that Y ^ 1 succeeds in recovering Y 1 up to a sign and that Y ^ 2 succeeds in recovering Y 2 .
Moreover, supposing p 1 = p 2 = 1 we can choose K 1 and K 2 according to the BIC criterion. Values of the BIC criterion are presented in Table 1, where we show that the selected model is the true one with K 1 = K 2 = 2 . The lower diagonal of the table is not presented for symmetry reasons, and we limited ourselves to K 1 , K 2 { 1 , , 5 } .

4.2. Experiments on Real Data

Let us consider the crabs dataset [20]. It consists of 200 crabs morphological data, each crab has two categorical (cluster) attributes—the species, orange or blue, and the sex, male or female. The dataset is composed of 50 males orange, 50 males blue, 50 females orange, 50 females blue for which 5 numerical attributes have been measured: the frontal lobe size, the rear width, the carapace length, the carapace width and the body depth. We can see the PCA of the data in Figure 8. We see that component two separates males and females well, whereas component three separates orange and blue subspecies. However, we will see that by applying our model we obtained a better separation of the clusters.
Like in the tutorial example we will take p 1 = p 2 = 1 and K 1 , K 2 { 1 , , 5 } . The resulting BIC tabular is given in Table 2, it suggests the choice of K 1 = 3 and K 2 = 4 . The resulting visualization of the clustering variables in given Figure 9. Let us notice that Y 2 is divided in four clusters however, however we only see three since two of them have the same mean. We can see that even if the numbers of clusters do not correspond, the first clustering subspace finds the subspecies, whereas the second clustering subspace finds the sex. We could also look at the solution provided by K 1 = K 2 = 2 on Figure 10, this one has a lower BIC but seems more natural for the problem at hand. We see that the obtained map is in fact quite similar the map obtained Figure 9; however, we notice that from a density approximation point of view we obtain a lower fit. In fact, if we look at the correlations between Y 1 Figure 9 and Y 2 Figure 10 we have a correlation of 0.97 , and a similar correlation is obtained between Y 2 Figure 9 and Y 1 Figure 10. Thus, the produced subspace are finally quite similar.

5. Conclusions and Perspectives

We have proposed a model which allows us to combine visualization and clustering with many clustering viewpoints. Moreover, we have shown the possibility of performing model choice by using the BIC criterion. The proposed model can provide new information on the structure present in the data by trying to reinterpret the cluster as a result of the Cartesian product of several clustering variables.
The proposed model is limited to the homoscedatic setting, which could be seen as a limitation; however, from our point of view this is more robust than the heteroscedastic setting, which is known to be jeopardized by the degeneracy issue [21]. However, the extension of our work on the heteroscedastic setting can easily be performed from the modeling point of view; the main issue in this case would be the parameters estimation where an extension of the FDA to the heteroscedastic setting would be needed, as presented in Kumar and Andreou [17]. Another difficult issue is the choice of H, K 1 , K H and p 1 , , p H , which is very combinatorial. Here we have proposed an estimation strategy for all these tuning parameters being fixed, and then performed a selection of the best tuning according to BIC. However, in future work, a model selection strategy to perform the model selection through a modified version of the EM algorithm will also be investigated as in Green [22]; it would thus limit the combinatorial aspect of the global model search through EM-wise local model searches.

Funding

This research received no external funding.

Acknowledgments

I would like to thank Rohit BHAGWAT for his preliminary work on the experimental part of the topic during his internship. I also would like to thank anonymous reviewers for their comments which contributed to improve the quality of the article.

Conflicts of Interest

The author declares no conflict of interest.

References

  1. McLachlan, G.; Peel, D. Finite Mixture Models; John Wiley & Sons: Hoboken, NJ, USA, 2004. [Google Scholar]
  2. Schwarz, G. Estimating the Dimension of a Model. Ann. Stat. 1978, 6, 461–464. [Google Scholar] [CrossRef]
  3. Tipping, M.E.; Bishop, C.M. Probabilistic principal component analysis. J. R. Stat. Soc. Ser. B Stat. Methodol. 1999, 61, 611–622. [Google Scholar] [CrossRef]
  4. Celeux, G.; Govaert, G. Gaussian parsimonious clustering models. Pattern Recognit. 1995, 28, 781–793. [Google Scholar] [CrossRef] [Green Version]
  5. Bouveyron, C.; Girard, S.; Schmid, C. High-dimensional data clustering. Comput. Stat. Data Anal. 2007, 52, 502–519. [Google Scholar] [CrossRef] [Green Version]
  6. Bouveyron, C.; Brunet, C. Simultaneous model-based clustering and visualization in the Fisher discriminative subspace. Stat. Comput. 2012, 22, 301–324. [Google Scholar] [CrossRef] [Green Version]
  7. Dempster, A.; Laird, N.; Rubin, D. Maximum likelihood from incomplete data via the EM algorithm. J. R. Stat. Society. Ser. B Methodol. 1977, 39, 1–38. [Google Scholar]
  8. Galimberti, G.; Soffritti, G. Model-based methods to identify multiple cluster structures in a data set. Comput. Stat. Data Anal. 2007, 52, 520–536. [Google Scholar] [CrossRef]
  9. Galimberti, G.; Manisi, A.; Soffritti, G. Modelling the role of variables in model-based cluster analysis. Stat. Comput. 2018, 28, 145–169. [Google Scholar] [CrossRef]
  10. Poon, L.K.; Zhang, N.L.; Liu, T.; Liu, A.H. Model-based clustering of high-dimensional data: Variable selection versus facet determination. Int. J. Approx. Reason. 2013, 54, 196–215. [Google Scholar] [CrossRef]
  11. Marbac, M.; Vandewalle, V. A tractable multi-partitions clustering. Comput. Stat. Data Anal. 2019, 132, 167–179. [Google Scholar] [CrossRef] [Green Version]
  12. Attias, H. Independent factor analysis. Neural Comput. 1999, 11, 803–851. [Google Scholar] [CrossRef] [PubMed]
  13. Fisher, R.A. The use of multiple measurements in taxonomic problems. Ann. Eugen. 1936, 7, 179–188. [Google Scholar] [CrossRef]
  14. Friedman, J.; Hastie, T.; Tibshirani, R. The Elements of Statistical Learning; Springer Series in Statistics; Springer: Berlin, Germany, 2001; Volume 1. [Google Scholar]
  15. Campbell, N.A. Canonical variate analysis—A general model formulation. Aust. J. Stat. 1984, 26, 86–96. [Google Scholar] [CrossRef]
  16. Hastie, T.; Tibshirani, R. Discriminant Analysis by Gaussian Mixtures. J. R. Stat. Society. Ser. B Methodol. 1996, 58, 155–176. [Google Scholar] [CrossRef]
  17. Kumar, N.; Andreou, A.G. Heteroscedastic discriminant analysis and reduced rank HMMs for improved speech recognition. Speech Commun. 1998, 26, 283–297. [Google Scholar] [CrossRef]
  18. Ghahramani, Z.; Hinton, G.E. The EM Algorithm for Mixtures of Factor Analyzers; Technical Report, Technical Report CRG-TR-96-1; University of Toronto: Toronto, ON, Canada, 1996. [Google Scholar]
  19. Biernacki, C.; Celeux, G.; Govaert, G. Assessing a mixture model for clustering with the integrated completed likelihood. IEEE Trans. Pattern Anal. Mach. Intell. 2000, 22, 719–725. [Google Scholar] [CrossRef]
  20. Campbell, N.; Mahon, R. A multivariate study of variation in two species of rock crab of the genus Leptograpsus. Aust. J. Zool. 1974, 22, 417–425. [Google Scholar] [CrossRef]
  21. Biernacki, C.; Chrétien, S. Degeneracy in the maximum likelihood estimation of univariate Gaussian mixtures with EM. Stat. Probab. Lett. 2003, 61, 373–382. [Google Scholar] [CrossRef]
  22. Green, P.J. On use of the EM for penalized likelihood estimation. J. R. Stat. Soc. Ser. Methodol. 1990, 52, 443–452. [Google Scholar]
Figure 1. Bayesian dependency graph for factorial discriminant analysis.
Figure 1. Bayesian dependency graph for factorial discriminant analysis.
Mathematics 08 00597 g001
Figure 2. Adapted Bayesian dependency graph to the multi-partition setting, for H = 2 clustering variables.
Figure 2. Adapted Bayesian dependency graph to the multi-partition setting, for H = 2 clustering variables.
Mathematics 08 00597 g002
Figure 3. Scatter plots of the initial data on the illustrative example. The color depends on the first cluster variable, and the shape depends on the second cluster variable.
Figure 3. Scatter plots of the initial data on the illustrative example. The color depends on the first cluster variable, and the shape depends on the second cluster variable.
Mathematics 08 00597 g003
Figure 4. Scatter plot of the illustrative data on the two original clustering subspaces.
Figure 4. Scatter plot of the illustrative data on the two original clustering subspaces.
Mathematics 08 00597 g004
Figure 5. Scatter plot of the illustrative data on components one to four of the proimcipal component analysis (PCA).
Figure 5. Scatter plot of the illustrative data on components one to four of the proimcipal component analysis (PCA).
Mathematics 08 00597 g005
Figure 6. Scatter plot of the component of the factorial discriminant analysis for the illustrative example.
Figure 6. Scatter plot of the component of the factorial discriminant analysis for the illustrative example.
Mathematics 08 00597 g006
Figure 7. Comparison of the true and of the estimated clustering subspaces on the illustrative example, points are marked according to the estimated clusters.
Figure 7. Comparison of the true and of the estimated clustering subspaces on the illustrative example, points are marked according to the estimated clusters.
Mathematics 08 00597 g007
Figure 8. Scatter plots of the crabs data after PCA. Subspecies are represented according to their color, and sex is represented according to its symbol.
Figure 8. Scatter plots of the crabs data after PCA. Subspecies are represented according to their color, and sex is represented according to its symbol.
Mathematics 08 00597 g008
Figure 9. Scatter plots of the clustering subspace on the crabs data for K 1 = 3 and K 2 = 4 , 95 % isodensity is given for each component resulting of the Cartesian product.
Figure 9. Scatter plots of the clustering subspace on the crabs data for K 1 = 3 and K 2 = 4 , 95 % isodensity is given for each component resulting of the Cartesian product.
Mathematics 08 00597 g009
Figure 10. Scatter plots of the clustering subspace on the crabs data for K 1 = K 2 = 2 , 95 % isodensity is given for each component resulting of the Cartesian product.
Figure 10. Scatter plots of the clustering subspace on the crabs data for K 1 = K 2 = 2 , 95 % isodensity is given for each component resulting of the Cartesian product.
Mathematics 08 00597 g010
Table 1. Value of the BIC criterion according to K 1 and K 2 , for the choice of the number of clusters on the illustrative example, best value in bold.
Table 1. Value of the BIC criterion according to K 1 and K 2 , for the choice of the number of clusters on the illustrative example, best value in bold.
K 1 \ K 2 12345
1−1318.40−1301.15−1305.20−1305.47−1310.08
2 −1291.80−1293.22−1292.28−1307.07
3 −1296.70−1303.68−1310.95
4 −1306.09−1320.29
5 −1319.22
Table 2. Value of the BIC criterion according to K 1 and K 2 , for the choice of the number of clusters on the crabs dataset, best value in bold.
Table 2. Value of the BIC criterion according to K 1 and K 2 , for the choice of the number of clusters on the crabs dataset, best value in bold.
K 1 \ K 2 12345
1−62.660.4110.405.110.80
2 17.8216.5718.880.49
3 3.7522.5217.44
4 −26.65−26.64
5 12.06

Share and Cite

MDPI and ACS Style

Vandewalle, V. Multi-Partitions Subspace Clustering. Mathematics 2020, 8, 597. https://doi.org/10.3390/math8040597

AMA Style

Vandewalle V. Multi-Partitions Subspace Clustering. Mathematics. 2020; 8(4):597. https://doi.org/10.3390/math8040597

Chicago/Turabian Style

Vandewalle, Vincent. 2020. "Multi-Partitions Subspace Clustering" Mathematics 8, no. 4: 597. https://doi.org/10.3390/math8040597

APA Style

Vandewalle, V. (2020). Multi-Partitions Subspace Clustering. Mathematics, 8(4), 597. https://doi.org/10.3390/math8040597

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop