Next Article in Journal
Magnetotelluric Signal-Noise Identification and Separation Based on ApEn-MSE and StOMP
Next Article in Special Issue
Informed Weighted Non-Negative Matrix Factorization Using αβ-Divergence Applied to Source Apportionment
Previous Article in Journal
Spatial Organization of the Gene Regulatory Program: An Information Theoretical Approach to Breast Cancer Transcriptomics
Previous Article in Special Issue
New Construction of Maximum Distance Separable (MDS) Self-Dual Codes over Finite Fields
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Centroid-Based Clustering with αβ-Divergences

Departamento de Teoría de la Señal y Comunicaciones, Escuela Técnica Superior de Ingeniería, Universidad de Sevilla, Camino de los descubrimientos, S/N, 41092 Sevilla, Spain
*
Authors to whom correspondence should be addressed.
Entropy 2019, 21(2), 196; https://doi.org/10.3390/e21020196
Submission received: 18 January 2019 / Revised: 6 February 2019 / Accepted: 14 February 2019 / Published: 19 February 2019
(This article belongs to the Special Issue Information Theory Applications in Signal Processing)

Abstract

:
Centroid-based clustering is a widely used technique within unsupervised learning algorithms in many research fields. The success of any centroid-based clustering relies on the choice of the similarity measure under use. In recent years, most studies focused on including several divergence measures in the traditional hard k-means algorithm. In this article, we consider the problem of centroid-based clustering using the family of α β -divergences, which is governed by two parameters, α and β . We propose a new iterative algorithm, α β -k-means, giving closed-form solutions for the computation of the sided centroids. The algorithm can be fine-tuned by means of this pair of values, yielding a wide range of the most frequently used divergences. Moreover, it is guaranteed to converge to local minima for a wide range of values of the pair ( α , β ). Our theoretical contribution has been validated by several experiments performed with synthetic and real data and exploring the ( α , β ) plane. The numerical results obtained confirm the quality of the algorithm and its suitability to be used in several practical applications.

1. Introduction

The clustering problem is related to the partition of an analyzed set of samples into a settled number of pairwise disjoint classes or clusters, where samples in the same cluster are more similar to each other than those samples of other clusters. Center-based clustering methods group the samples based on some measure of distance from cluster centers. In this context, the center of a cluster can be a medoid or a centroid. A medoid is the most representative point of a cluster, while a centroid is usually calculated as a minimizer of an optimization problem, with a measure of distortion as the objective function. The choice of a proper measure of similarity or dissimilarity (distance) is a key factor in cluster analysis, since the performance of clustering algorithms greatly relies on this choice.
Arguably, the most popular clustering algorithm is k-means with Lloyd’s heuristic, in which squared Euclidean distance is used to compute the distortion. However, there has been a recent burst of interest in extending classical k-means algorithm to a larger family of distortion measures. In particular, the use of divergence-based distance functions as a similarity measure has recently gained attention. Research on this topic makes use mainly of two families of divergences, Csiszár f-divergences and Bregman divergences. Both families include some well-known divergences. For example, α -divergences, which includes the Kullback-Leibler (KL) divergence, is a type of Bregman and Csiszár f-divergences. In fact, α -divergence is the unique class of divergence sitting at the intersection of the Csiszár f-divergence and Bregman divergence classes [1]. Other notable Csiszár f-divergences are the Hellinger distance and the χ -squared distance [2], whereas the squared Euclidean distance, the Itakura-Saito (IS) divergence and the β -divergence are special cases of Bregman divergences [3].
Such distance functions do not always satisfy certain properties, such as triangular inequality and distance symmetry, making them an improper metric. Thus, for the development of a clustering strategy, one must consider two kinds of centroids obtained by performing the minimization process either on the left argument or on the right argument of the distance function, yielding the left-sided and right-sided centroids, respectively. Closed formulas for both sided centroids computation have been proposed in the literature for different divergence families.
In [4], the classical hard k-means algorithm is generalized to the large family of Bregman divergences. The resulting Bregman k-means algorithm works for any given Bregman divergence. Since Bregman divergences are not necessarily symmetric, it is necessary to distinguish the two aforementioned sided centroids. It has been shown that the left-centroid is the generalized means of the cluster, also called cluster’s f-mean, whereas the right centroid is the center of mass of the cluster, independently of the considered Bregman divergence [5]. For the specific case of α -divergences, closed formulas for the computation of the sided centroids were derived in [6] for the right-type, and in [7,8] for the left-type. Symmetrized centroids have also been derived for clustering histograms in the Bag-of-Words modeling paradigm in [9]. Total Bregman divergences (TDB), which are invariant to particular transformations on the natural space, have also been used for estimating the center of a set of vectors in [10] in the context of the shape retrieval problem. Complete formulation of sided centroids in k-means algorithm with TDB are reported in [11]. To the best of our knowledge, there is no closed formulation for the computation of centroids for the whole of Csiszár f-divergence family. One of the works that relates Csiszár f-divergences and clustering can be found in [12], in which a generalized version of f-divergences, called ( f , l ) -divergence is used for clustering in the particular case of KL-divergence. Finally, other classes of distance functions which are not necessarily Bregman or Csiszár f-divergences have been employed for clustering. For example, a notable recent study can be found in [13], in which a k-means algorithm using the S-divergence is developed for feature coding.
We propose a new center-based clustering algorithm, namely α β -k-means algorithm, using α β -divergence family as a measure of distortion. Our motivations to explore the family of α β -divergences for centroid-based clustering are the great flexibility to obtain a rich collection on particular divergences by just tuning the parameters ( α , β ) , and the possibility to yield simple closed formulas for the centroids computation that are of interests for processing several types of data.
α β -divergences, which were introduced in [14] as a new dissimilarity measure for positive data, have been proven useful in several applications, such as for example the separation of convolved speech mixtures [15], to perform the canonical correlation analysis [16] and noise-robust speech recognition [17]. This family of divergences are governed by two parameters α and β , and cover many of the divergences previously used for clustering, such as α -divergence, β -divergence and KL-divergence.
This is not the first attempt to take into account the α β -divergences in a clustering approach. In [18], a variation of a k-medoid clustering algorithm is presented based on the α β -divergences. The resulting algorithm fixes the value of the α parameter to α = 1 , and varies the value of the parameter β in each iteration through the prominence of the cluster. However, the method is completely different to our proposal. The algorithm we propose in this article computes the centroid of the cluster by solving a minimization problem, whereas the algorithm in [18] obtains the center of the cluster by an exhaustive search optimization technique on the set of the current members of the cluster. Thus, we propose a k-means-type algorithm whereas in [18] it is presented as a k-medoid-type algorithm.
Finally, some authors have pointed out that k-means clustering can be formulated as a constrained matrix factorization problem [19,20]. For instance, in [19] the authors showed that orthogonal symmetric Non-negative Matrix Factorization (NMF) with the sum of squared error cost function is equivalent to Kernel k-means clustering. Other variants of NMF, such as Semi-NMF, Convex-NMF, Cluster-NMF and Kernel-NMF are all soft versions of k-means clustering [21]. There is also a relationship between NMF based on certain divergences and some clustering approaches. NMF with generalized KL-divergence or I-divergence is equivalent to Probabilistic Latent Semantic Indexing (PLSI) [22]. In addition, in [23] it is established that orthogonal NMF based on Bregman divergence problem is equivalent to Bregman hard clustering derived in [4].
This paper is organized as follows: we begin with the formal definition and some properties of α β -divergences in Section 2. In Section 3 we derive the closed-form formula for the sided centroids by using the α β -divergence, and generalize the k-means algorithm to the α β -k-means algorithm. In Section 4 we demonstrate that the obtained formula for centroid computation match with previous formula for some specific distances and divergences that belong to α β -divergence family. Section 5 presents some experimental clustering results for synthetic and real datasets. Finally, Section 6 summarizes our main conclusions and provides some suggestions for future research.

2. α β -Divergences

In this section, we recall the definition and some useful properties of the α β -divergences [14].
Definition 1.
Given two non-negative data matrices of same dimension P R + I × T and Q R + I × T , with entries p i t = P i t and q i t = Q i t , the α β -divergence is given by
D A B α , β P Q = i t d A B α , β p i t , q i t
where
d A B α , β p i t , q i t = 1 α β p i t α q i t β α α + β p i t α + β β α + β q i t α + β , f o r α , β , α + β 0 1 α 2 p i t α ln p i t α q i t α p i t α + q i t α , f o r α 0 , β = 0 1 α 2 ln q i t α p i t α + q i t α p i t α 1 1 , f o r α = β 0 1 β 2 q i t β ln q i t β p i t β q i t β + p i t β , f o r α = 0 , β 0 1 2 ln p i t ln q i t 2 , f o r α , β = 0
It must be pointed out that some specific choices of ( α , β ) parameters simplify the α β -divergence into some known divergences or families of divergences, allowing smooth interpolation between many known divergences. In particular, when α = β = 0 the α β -divergence takes the form of a Log-Euclidean distance
D A B 0 , 0 P Q = D E log P log Q
where D E represents the squared Euclidean distance. When α + β = 0 , with α , β 0 , the α β -divergence can also be expressed in terms of a generalized Itakura-Saito distance D I S with an α -zoom of the arguments
D A B α , α P Q = 1 α 2 D I S P . α Q . α = 1 α 2 i t log q i t α p i t α + p i t α q i t α 1
where P . α denotes the one-to-one transformation that raises each element of the vector P to the power α . When α + β = 1 the α β -divergence reduces to the α -divergence,
D A B α , 1 α P Q = D A α P Q
whereas when α = 1 , it reduces to the β -divergence
D A B 1 , β P Q = D B β P Q .
For α = β = 0.5 , it is proportional to the Hellinger distance
D A B 0.5 , 0.5 P Q = 4 D H P Q .
Also, the AB-divergence reduces to the standard KL-divergence for α = 1 and β = 0 ,
D A B 1 , 0 P Q = D K L P Q .
Although α β -divergences are not true metrics, they satisfy some interesting properties (see [14] for details and proofs), such us duality, inversion, and scaling:
  • D A B α , β P Q = D A B β , α Q P (Duality)
  • D A B α , β P Q = D A B α , β Q . 1 P . 1 (Inversion)
  • D A B w α , w β P Q = 1 w 2 D A B α , β P . w Q . w (Scaling)
α β -divergence is more flexible, powerful, and robust against errors and noise than other divergence families, such as the α -divergence and β -divergence [14]. The role of the hyperparameters α and β in the robustness property of the α β -divergence is described in [14]. Formally, it has been shown that if we assume that the right argument of the divergence, Q is a function of a vector of parameters θ , then
D A B α , β P Q θ = i t q i t θ q i t α + β 1 w e i g h t s ln 1 α p i t q i t α z o o m .
In this case, the parameter α can be used to control the influence of large or small ratios p i t q i t by the deformed logarithm of order 1 α , while the parameter β provides some control on the weighting of the ratios by scaling factors q i t α + β 1 .

3. K-means Clustering with α β -Divergences

One of the most popular and well-studied data analysis methods for clustering is k-means clustering. Let X be a random variable that take values in a finite set X = x i i = 1 n R T . k-means clustering aims to split X into k disjoint partitions c i i = 1 k , finding a set of centroids M = m h h = 1 k R T , with M = k and k < n . The standard k-means formulation finds the partition C = c i i = 1 k through the minimization of the sum of the squared Euclidean distances between each sample and its cluster centroid. Formally speaking, for a fixed integer k, one can define the following squared Euclidean loss function
L M = min c 1 , , c k h = 1 k x i c h x i m h 2 2 .
The most popular optimization method for this minimization problem is Lloyd’s algorithm [24], which converges to a stable fixed point that corresponds to a local minimum of the loss function. For a given initial partition, Lloyd’s algorithm finds the partition in a two-step iterative process. In the assignment step, each data is assigned with the cluster whose centroid is closest. In the update step, the centroids are updated as the arithmetic mean of its assigned points. It is well established that the arithmetic mean is the optimal centroid m h , D E * for the Euclidean distance
m h , D E * 1 | c h | x i c h x i = arg min m h x i c h x i m h 2 2
where | c h | refers to the cardinality of th h-th cluster.
We propose in this paper to generalize the standard k-means clustering to the α β -divergences. As in k-means standard technique, our objective is to find the set of centroids M that minimizes the AB-divergence of points in the set X to their corresponding centroids. In this context, the centroid of a cluster is defined as the optimizer of the minimum average α β -divergence. However, since the α β -divergences are not symmetrical, one must consider two kinds of centroids obtained by performing the minimization process either on the left argument or on the right argument of the divergences. We shall consider the right-sided centroid, m h , D A B R * the optimizer when the minimization process is performed with respect to the right side of the divergence, and the left-sided centroid, m h , D A B L * , the optimizer when the minimization is performed with respect to the left
m h , D A B R * ( α , β ) = arg min m h x i c h D A B α , β x i m h = a r g min m h x i c h t d A B α , β x i t , m h t
m h , D A B L * ( α , β ) = arg min m h x i c h D A B α , β m h x i = a r g min m h x i c h t d A B α , β m h t , x i t .
In [5] it is proven that sided centroids with respect to Bregman divergences coincide with the center of mass for the right-type, and with the center of mass for the gradient point set that is a f-mean, for the left-type. This implies that the formula for the right-type centroid computation does not depend on the Bregman divergence considered, whereas the formula for the computation of the left-type centroid strongly depends on it. Moreover, the sided centroids for Bregman divergences exhibit different characteristics, and therefore it is necessary to choose between the left and right centroid depending on the application. Contrary to the Bregman divergences case, we can establish a relationship between the sided centroids obtained with α β -divergences that unifies the optimization process for the sided centroids in a unique problem.
Lemma 1.
Let m h , D A B R * ( α , β ) denote the optimal right-sided centroid defined in Equation (12) for a given parametrization pair ( α , β ) . The left-sided centroid for the same parametrization is
m h , D A B L * ( α , β ) = m h , D A B R * ( β , α ) .
Proof. 
Using the duality property of α β -divergences, we observe that
m h , D A B L * ( α , β ) = arg min m h x i c h D A B α , β m h x i = arg min m h x i c h D A B β , α x i m h = m h , D A B R * ( β , α ) .
This last result allows us to formulate the following theorem.
Theorem 1.
(Sided α β -centroids) The right-sided m h , D A B R * ( α , β ) and left-sided m h , D A B L * ( α , β ) α β -centroid coordinates of a set of point x i c h are:
m h , D A B R * ( α , β ) = | c h | 1 α x i c h x i α 1 α f o r α 0 x i c h x i 1 | c h | f o r α = 0
m h , D A B L * ( α , β ) = | c h | 1 β x i c h x i β 1 β f o r β 0 x i c h x i 1 | c h | f o r β = 0 .
The proof is reported in the Appendix A. Please note that the expression obtained for α = 0 in Equation (16) corresponds to the limit when α 0 evaluated with L’Hôpital’s rule, in the same way as with β = 0 in Equation (17). Closed-form formulas presented in Theorem 1 are essential, since it allows us to develop efficient k-means algorithms using α β -divergences.
Now, we can introduce the following iterative Algorithm 1, known as the α β -k-means algorithm. As in the traditional k-means algorithm, the algorithm begins with an initial guess of the centroids (usually at random), and then alternates between the assignment and update steps. In the assignment step, each data point is assigned to the closest cluster centroid, measuring the distance through the α β -divergence. In the updated step, the centroids are computed using the results on Theorem 1. The algorithm is reiterated until convergence is met. In practice, we can control the stopping criterion by taking the difference between the cost function of two successive iterations. If it is less than a prescribed threshold ϵ the algorithm will stop. A precise definition of the aforementioned strategy using the right-type centroid is presented in Algorithm 1.
Algorithm 1 α β -divergence clustering.
Input: Set X = x i i = 1 n R + T , hyperparameters ( α , β ) , number of clusters k.
Output: M α β R * , local minimizer of L α β ( M ) = h = 1 k x i c h D A B α , β x i m h where M = m h h = 1 k , hard partitioning C α β R * = c h h = 1 k of X .
Method:
 Initialize m h h = 1 k with m h R + T
repeat
  (The Assignment Step)
  Set c h 0 , 1 h k
  for i = 1 to n do
    c h c h x i
   where h = h * ( x i ) = arg min h D A B ( α , β ) ( x i m h )
  end for
  (The Update Step)
  for h = 1 to k do
    m h | c h | 1 α x i c h x i α 1 α for α 0 x i c h x i 1 | c h | for α = 0
  end for
until convergence
 return M α β R * m h h = 1 k , C α β R * c h h = 1 k
Obviously, this algorithm can be extended to the left-type centroid in a straightforward manner by just changing the order of the arguments of the α β -divergence in the assignment step, and using the left-centroid formulas in the update step to obtain the set of centroids M α β L * and the hard partitioning of data C α β L * . However, it is easy to check that there is a relationship between the α β -k-means algorithm for the right-type centroid and for the left-type centroid. For the same initialization, due to the property the duality of the α β -divergences and the result of Lemma 1 one can get that M α β L * = M β α R * and C α β L * = C β α R * . Therefore, the behavior of the left-type α β -k-means algorithm in the α β plane is equal to behavior of the right-type reflected in the line α = β .

Conditions for the Convergence of Algorithm

Looking for the conditions ensuring the existence of an optimal set of centroids that achieves the minimum of the clustering loss function based on the α β -divergence, it is necessary to take into account that α β -divergences are not necessarily convex in the second argument. In particular, the conditions required for convexity depends on the value of α and β as follows [14]
p i t q i t exp 1 α 1 β 1 for β < min 1 , 1 α always   convex for β min 1 , 1 α , max 1 , 1 α p i t q i t exp 1 α 1 β 1 for β < max 1 , 1 α
where exp 1 α ( · ) is a 1 α deformed exponential
exp 1 α ( z ) = exp ( z ) for α = 0 1 + α z 1 α for α 0 and 1 + α z 0 0 for α 0 ; and 1 + α z < 0 .
Figure 1 shows an analysis of the convergence region of the α β -k-means algorithm. The region filled in blue represents the convex cone delimited by the lines α + β = 1 and β = 1 , in which the α β -divergence is always convex with respect to the second argument. Therefore, the proposed algorithm converges to a local minimum within this cone, independently of the values of the data to be analyzed. However, Equation (18) shows that the convexity of the divergence with respect to q i t holds outside the convex cone when the ratios p i t q i t are bounded by the function exp 1 α 1 β 1 . Therefore, theoretically, the convergence region of the proposed algorithm is greater than the convex cone, and its borders depend on the values that the arguments of the divergence could take. In particular, the maximum and the minimum values of the ratios p i t q i t determine the upper and the lower boundaries, respectively. Blue lines in Figure 1 represent the boundaries of the convergence region in the α β plane for some values of the function exp 1 α 1 β 1 . In practice, for relatively small errors between x i t and m h t , Algorithm 1 is guaranteed to converge to a local minimum in a wide range of ( α , β ) values.

4. Relations with Known Centroid-Based Algorithms

The obtained formulas for sided centroid computation match the previous expressions obtained for various specific divergences and distances. In fact, the above novel algorithm unifies many existing algorithms for k-means. For example, apart from the most popular squared Euclidean distance, the algorithm includes those k-means algorithms based on α -divergences and β -divergences, as well as the particular cases KL-divergence and IS-divergence.
We shall start with the squared Euclidean distance that can be obtained from the α β -divergence for α = β = 1 . By substituting these α and β values on the sided centroids of Equations (16) and (17), we directly obtain that both centroids are the arithmetic mean (11)
m h , E * = m h , D A B R * ( 1 , 1 ) = m h , D A B L * ( 1 , 1 ) = | c h | 1 x i c h x i .
Another interesting case of study is the α -divergence that can be obtained from the α β -divergences for the parametrization α + β = 1 . As mentioned before, there are closed formulas for the computation of sided centroids for the k-means with α -divergences [7,9]. However, to compare the formulas, it is necessary to take into account that there are two equivalent ways to define the α -divergence family. In particular, some authors employ a slightly different notation that depends on the parameter α A [3], which is related to the parameter α as follows
α A = 1 2 α .
The α -divergence parametrized by α A takes the following form
D A α A P Q = 4 1 α A 2 i t 1 α A 2 p i t + 1 + α A 2 q i t p i t 1 α A 2 q i t 1 + α A 2 , for α A ± 1 i t p i t log p i t q i t + q i t p i t , for α A = 1 i t p i t log q i t p i t + p i t q i t , for α A = 1 .
Closed formulas for sided centroids employing this notation are:
m h , D A R * ( α A ) = | c h | 2 1 α A x i c h x i 1 α A 2 2 1 α A , for α A 1 x i c h x i 1 | c h | , for α A = 1
m h , D A L * ( α A ) = | c h | 2 1 α A x i c h x i 1 + α A 2 2 1 + α A , for α A 1 x i c h x i 1 | c h | , for α A = 1 .
Substituting Equation (21) in Equation (4), it is easy to check that D A α A P Q = D A α P Q . Also, substituting Equation (21) in Equations (23) and (24) we obtain
m h , D A R * ( α A ) = m h , D A B R * ( α , 1 α )
m h , D A L * ( α A ) = m h , D A B L * ( α , 1 α ) .
One widely employed distance in clustering tasks is the KL-divergence, obtained either for ( α , β ) = ( 1 , 0 ) from the α β -divergence, or α A = 1 from the α -divergence. In this specific case, the left-sided and right-sided centroids are computed as geometric mean and arithmetic mean respectively [5]. It must be noticed that it is enough to consider ( α , β ) = ( 1 , 0 ) in Equations (23) and (24) to obtain the same formulas for centroid’s computation than those reported in [5]
m h , D K L R * = m h , D A B R * ( 1 , 0 ) = 1 | c h | x i c h x i
m h , D K L L * = m h , D A B L * ( 1 , 0 ) = x i c h x i 1 | c h | .
Finally, another divergence frequently used in the spectral analysis of speech signals [25] is the IS-divergence. This divergence was first proposed in the context of vector quantization in [26], and can be expressed by the β -divergence, the Bregman divergence and the α β -divergence for ( α , β ) = ( 1 , 1 ) . As all the Bregman divergences, the right-sided centroid is computed as the arithmetic mean, but in this case, the left-sided centroid corresponds to the harmonic mean [5]. Again, applying ( α , β ) = ( 1 , 1 ) in Equations (23) and (24), we get the same results than in [5]
m h , D I S R * = m h , D A B R * ( 1 , 1 ) = 1 | c h | x i c h x i
m h , D I S L * = m h , D A B L * ( 1 , 1 ) = | c h | x i c h 1 x i .

5. Experimental Results and Discussion

We have evaluated the proposed α β -k-means algorithm on various data types with experiments on both synthetic and real datasets. The first experiment studies the behavior of the algorithm on four different synthetic datasets, whose density is known, whereas the second experiment considers the task of audio genre classification using two different sets of descriptors. The third experiment analyzes the performance of the algorithm in two datasets from the UCI Machine Learning Repository [27]: Iris and Wine. It is expected that the behavior of the algorithm strongly depends on the choice of the tuning parameters and on the type of the data. Given the huge flexibility of the α β -divergence framework, we must restrict the experiment to some particular cases. Therefore, in our simulations we have varied α and β within the range 2 α 2 and 2 β 2 with steps of 0.1 . In each simulation we run 10 replicates from different randomly selected starting points and determined the partition with the lowest total sum of distances over all replicates. The resulting clusters have been evaluated in terms of accuracy degree (ACC) measured as the number of correctly assigned data points divided by the total number of data. Let us denote g i the ground truth label, l i the clustering assignment produced by the algorithm and m a p ( l i ) the optimal one-to-one mapping function that permutes clustering labels to match the ground truth labels by using the Hungarian algorithm [28]. The ACC is defined as
A C C = i = 1 n δ g i , m a p ( l i ) n
where δ ( x , y ) = 1 if x = y and δ ( x , y ) = 0 otherwise.

5.1. Clustering on Synthetic Datasets

In this experiment we have generated four 1-dimensional datasets of 3000 samples each, based on mixture models of Gaussian, Log-Gaussian, Poisson, and Binomial distributions, respectively. The datasets have three components of 1000 samples each with means equals to 70, 80, and 100, respectively. The standard deviation of the Gaussian and Log-Gaussian densities was set to 5, and the number of trials of the binomial distribution was set to 1000. According this, the variance of the four models are approximately the same. The density functions of the generative models are depicted in Figure 2.
The choice of these density functions is not arbitrary. Gaussian, Poisson, and Binomial distributions are class members of the exponential family of distributions, and in [4] it is well stablished that for every exponential family distribution, it exists a corresponding generalized distance measure. For instance, normal distributions are associated with Euclidean distance, the exponential distribution with IS-divergence and Binomial distribution with KL-divergence. Additionally, Tweedie distributions, which are a particular type of exponential family, are tied to β -divergences [29,30]. Tweedie distributions include Poisson distribution which has been connected to β -divergence with β = 0 , that is the KL-divergence. Thus, we expect to obtain satisfactory clustering results for some specific divergences in each dataset analyzed.
Furthermore, with the choice of these distributions we want to verify the relation between the convergence region of the algorithm and the extreme values of the different datasets. For example, Poisson and Binomial datasets have similar extreme values and therefore the theoretical convergence region of the algorithm in both cases should be similar. On the contrary, the convergence region for the Log-Gaussian dataset is theoretically more extensive than the convergence regions of the other three datasets considered. In fact, for each cluster it seems probable that the ratio x i t / m h t approaches to unit, so that the convergence region covers the whole α β plane.
We have repeated the clustering experiment based on 1000 different random datasets, but preserving the same random initialization to run the left and right-type α β -k-means algorithm in each trial. The contour plots of the average ACC results of α β -k-means for the right and left cases are shown in Figure 3 and Figure 4, respectively. In these contour plots, the areas between isolines are filled with constant colors. In Table 1, we show the mean and standard deviation of ACC over 1000 trials obtained for some representative values of the ( α , β ) pair for the right-type algorithm. We have omitted the values for the left-type since they can be easily inferred from the right-type solutions.
These results can be interpreted from different points of view. First, as we predicted in Section 3, one can check that the behavior of the left-type algorithm is equal to the behavior of the right-type reflected in the line α = β . Second, we observe that the Euclidean distance ( α = 1 , β = 1 ) is well suited to Gaussian dataset, and KL-divergence ( α = 1 , β = 0 ) to Poisson and Binomial datasets, as expected. For the Log-Gaussian dataset, the α β -k-means algorithm behaves practically in the same way in the whole ( α , β ) region studied. This can be interpreted as the algorithm converges in the whole α β plane studied, as we had foreseen. Furthermore, for the Gaussian dataset, we observe that the performance of the algorithm begins to decrease significantly for ( α , β ) values below the line α + β = 2 . For the Poisson and Binomial datasets, the region in which the algorithm achieves good results are very similar, although the optimal values of α and β within this region are different for each type of distribution. Finally, we would like to highlight that the regions with good performance for Poisson and Binomial datasets are narrower than the region obtained for the Gaussian dataset.

5.2. Musical Genre Clustering

In this experiment we have classified music audio according to the genre of the track. The goal of this experiment is to investigate the optimal divergence for different feature vectors by performing clustering experiments using different ( α , β ) pairs. Moreover, we have considered two levels of complexity building datasets with K = 3 and K = 5 genres. The tracks were extracted from the Music Information Retrieval Evaluation eXchange (MIREX) database for US Pop Music Genre Classification that considers a total of 10 genres. Each genre contains 100 tracks of 30-sec duration and the tracks are all 22,050 Hz Mono 16-bit audio files in .wav format. The first built dataset is composed of the genres classical, metal and pop from the MIREX dataset, while the second built dataset adds to the three previous genres country and disco. Therefore, the first dataset is composed of 300 tracks whereas the second dataset is composed of 500 tracks.
Relevant descriptors that could sufficiently discriminate the audio genre were extracted from the audio tracks. Two sets of positive features were analyzed: Discrete-Fourier-Transform-based (DFT-based) descriptors and acoustic spectro-temporal features. The feature extraction process for the computation of the DFT-based descriptors was composed of the following stages [18]. First, the audio segment was divided into L overlapping frames of size 2048 points, with 25 % overlap between contiguous sections. For each audio track the number of segments obtained was L = 430 . Then, we performed the DFT of each segment using the Fast-Fourier Transform (FFT) algorithm with Hamming-windowing. After that, we calculated the arithmetic average in each frequency bin considering the absolute values of the complex DFT-vectors, and finally we normalized the average DFT by the sum of the DFT average coefficients. The length of the DFT-based descriptors was 1025.
Figure 5 shows the performance of the proposed algorithm and Table 2 resumes the results for some specific distances and divergences that belong to the α β -divergences family. The best performance for K = 3 , ACC = 0.9767 , was obtained for the pair ( α , β ) = ( 1.7 , 1 ) , which is quite close to the α -divergence. For K = 5 , the best performance, ACC = 0.6860 was obtained for ( α , β ) = ( 0.6 , 0.4 ) that is also close to an α -divergence, and in particular to the Hellinger distance.
The acoustics descriptors were composed of temporal features and spectral features. The temporal features extracted were Beat-Per-Minute (BPM) [31], and mean and standard deviation of: Zero-Crossing Rate (ZCR), energy and entropy. The spectral descriptors used were mean and standard deviation of: Spectral Centroid, Spectral Entropy, Spectral Flux, Spectral Flux Centroid, and Spectral Roll-off. The MATLAB Audio Analysis Library [32] was employed for the calculation of all the features except the BPM. The total feature vector consisted on 17 elements.
Figure 6 shows simulation results for the acoustic descriptors for the right-type α β -k-means algorithm. The best performance for K = 3 , with ACC = 0.9167 , was obtained for the pair ( α , β ) = ( 0.2 , 0 ) , which is quite close to the Log-Gaussian distance, while for K = 5 , the highest value, ACC = 0.7000 was obtained for ( α , β ) = ( 0.8 , 0.7 ) , which is quite close to IS-divergence. In general, the graph reveals that the region with best performance for this feature vector is close to the generalized IS-divergence, obtained from the α β -divergence for the values α = β . Table 3 summarizes the results for some specific divergences and distances obtained with α β -divergences.
It can be clearly seen from Figure 5 and Figure 6 that the region where the performance is satisfactory is enclosed between the lines α = β and α + β = 1 for the two tested feature vectors. These two lines correspond to the generalized IS-divergence and the α -divergence, respectively. However, a theoretical explanation for this fact is not trivial. As reported in Section 2, the hyperparameters ( α , β ) can control influence of individual ratios x i t m h t in the centroid computation (see Equation 9). In particular, for α > 1 the smaller values of the ratio are down-weighted with respect to the larger ones, whereas for α < 1 , the larger ratios are down-weighted with respect to the smaller ones. Simultaneously, those ratios are weighted by scaling factors m h t α + β 1 .
In our experiments, one can observe that in the region with best performance the value of α + β seems to be constant and close to unity for the DFT-based descriptors, and to zero for the acoustic descriptors. This cause that the multiplicative weighting factor m h t α + β 1 does not really affect the estimation of the centroid coordinates for the DFT-based descriptors, whereas the large values in the centroid coordinates are slightly down-weighted compared to the smaller values for the acoustic descriptors. Additionally, we observe that for α + β < 0 , the performance of the algorithm deteriorates drastically for the two sets of descriptors analyzed, probably due to the inversion of the arguments of the α β -divergence. It is easy to see that for α + β < 0 , the inversion property of the α β -divergence causes small values of the data to be strongly enhanced.
It is worth mentioning here that the values of the hyperparameters that provide the best classification could not be generalized to other classification problems. For other classification tasks, it is preferable to carry out a search on the ( α , β ) plane to obtain the best performance.

5.3. Clustering Analysis in UCI Repository

We have conducted experiment over two popular datasets from the UCI repository: Iris and Wine. The Iris dataset consist of three species of iris, 50 specimens in each. The features are measurements in centimeters of the length and the width of the sepals and petals. Wine dataset comprises results of chemical analyses of the content of wine grown in the same region but derived from three different cultivars. The dataset has 13 features and 178 instances. In this experiment we have varied α and β within the range 2 α 2 and 2 β 2 with steps of 0.2. Figure 7 shows the contour plots of the average accuracy obtained over 50 trials with random initializations. The best performance for Iris dataset, with ACC = 0.9600 , was obtained in the ( α , β ) region delimited by the generalized IS-divergence and the α -divergence. For Wine dataset, the highest value, ACC = 0.9663 was obtained for ( α , β ) = ( 1 , 1.2 ) , which is quite close to the generalized IS-divergence. In this case, the α -divergence did not get good results. Average ACC levels for some specific distances and divergences are presented in Table 4. It is important to emphasize that the Euclidean distance is not included in the distances and divergences with better results in the two datasets analyzed.

6. Conclusions

In this article, we have derived a centroid-based hard clustering algorithm that involves the family of α β -divergences which are governed by two parameters ( α , β ) and have many desirable properties useful for clustering tasks. First, α β -divergences admit closed-form expressions for the computation of the two sided—left or right—centroids, something relevant from the point of view of the implementation of the clustering algorithm. Second, the proposed algorithm, called α β -k-means, unifies many existing implementations of k-means obtained with a rich collection of particular divergences that belong to the large family of α β -divergences. In fact, we have demonstrated that our formulas for sided centroids coincide with other formulas previously developed for some specific cases of the α β -divergences, such us α -divergences, IS-divergence, and KL-divergence. Finally, the convergence of the algorithm is theoretically assured for a wide region in the ( α , β ) plane. Although the boundaries of the region of convergence depends on the ratio between the extreme values of the data, in practice, the algorithm seems to work properly outside of that theoretical region.
One of the important and still-open problems is how to tune the parameters ( α , β ) depending on the distribution of the available dataset and noise or outliers. Our experiments with synthetic datasets have allowed us to verify that the optimal values of the parameters α and β are related to the distribution of the data to be clustered. With this relationship, we can restrict the search range of the α and β values to some α β -divergences close to other well-known divergences, such us KL-divergence or α -divergences. The derivation of a precise formula for the choice of the parameters α and β is beyond the scope of this work.
Finally, it would be very interesting to study the relationship between the proposed α β -k-means algorithm and the multiplicative NMF algorithm based on the α β -divergence developed in [14].

Author Contributions

Conceptualization, A.S. and I.F.; Formal analysis, A.S., I.F. and I.D.-D.; Software, A.S.; Supervision, S.C.; Validation, A.S., I.F. and I.D.-D.; Visualization, A.S.; Writing—original draft, A.S.; Writing—review & editing, A.S., I.F., I.D.-D. and S.C.

Funding

This research was funded by the Spanish Ministry of Economy and Competitiveness (MINECO), grant number TEC2017-82807-P.

Conflicts of Interest

The authors declare no conflict of interest. The founding sponsors had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

Abbreviations

The following abbreviations are used in this manuscript:
ACCAccuracy
ISItakura-Saito
KLKullback-Leibler
NMFNon-negative Matrix Factorization

Appendix A

In this appendix, we prove the formulated theorem. Without loss of generality, we consider the computation of the right-sided centroid as the minimizer of the following optimization task
m h , D A B R * ( α , β ) = arg min m h x i c h t d A B α , β x i t , m h t .
We obtain the solution taking the gradient with respect to the second argument m h t , removing all terms independent of m h t , and setting it to zero
m h t x i c h t d A B α , β x i t , m h t = x i c h t d A B α , β x i t , m h t m h t = x i c h d A B α , β x i t , m h t m h t = 0 .
For the case of α , β , α + β 0 , the partial derivative is given by
x i c h d A B α , β x i t , m h t m h t = x i c h m h t 1 α β x i t α m h t β α α + β x i t α + β β α + β m h t α + β = 1 α β x i c h β x i t α m h t ( β 1 ) β m h t ( α + β 1 ) = x i c h m h t ( β 1 ) m h t α x i t α α .
By setting the derivative in Equation (A3) to zero and considering the non-negative character of x i t and m h t , we obtain
x i c h m h t α x i t α = 0 m h t α = 1 | c h | x i c h x i t α m h t = | c h | 1 α x i c h x i t α 1 α , t .
Analogously, for α 0 , β = 0 , the partial derivative becomes
x i c h d A B α , β x i t , m h t m h t = x i c h m h t 1 α 2 x i t α ln x i t α m h t α x i t α + m h t α = 1 α 2 x i c h α x i t α m h t ( α 1 ) m h t α + α m h t ( α 1 ) = x i c h m h t ( α 1 ) 1 x i t α m h t α α = 0 .
In this case, the solution is achieved as
x i c h 1 x i t α m h t α = 0 m h t α = 1 | c h | x i c h x i t α m h t = | c h | 1 α x i c h x i t α 1 α , t .
Similarly, for α = β 0 , the partial derivative takes the following form
x i c h d A B α , β x i t , m h t m h t = x i c h m h t 1 α 2 ln m h t α x i t α + m h t α x i t α 1 1 = 1 α 2 x i c h α m h t ( α 1 ) m h t α α x i t ( α ) m h t ( α 1 ) = x i c h m h t ( 1 ) 1 x i t α m h t α α = 0
which solution takes de same form of Equation (A6).
For α = 0 , β 0
x i c h d A B α , β x i t , m h t m h t = x i c h m h t 1 β 2 m h t β ln m h t β x i t β m m t β + x i t β = 1 β 2 x i c h β m m t ( β 1 ) ln m h t β x i t β + β m m t ( β 1 ) β m m t ( β 1 ) = x i c h m h t ( β 1 ) ln m h t β ln x i t β β = 0 .
In this case, the solution is given by
x i c h ln m h t β ln x i t β β = x i c h ln m h t ln x i t = 0 ln m h t | c h | = x i c h ln x i t = ln x i c h x i t m h t = x i c h x i t 1 | c h | , t .
Finally, for α = β = 0 , the partial derivative becomes
x i c h d A B α , β x i t , m h t m h t = x i c h 1 2 ln x i t ln m h t 2 = 1 2 x i c h 2 ln m h t m h t 2 ln x i t m h t = x i c h m h t 1 ln m h t ln x i t = 0
and we get the same solution as in Equation (A9).
After combining the cases in a single expression, the above actualization rules for computing the right-sided centroid become
m h , D A B R * ( α , β ) = | c h | 1 α x i c h x i α 1 α for α 0 x i c h x i 1 | c h | for α = 0 .
Using Lemma 1, we can directly obtain the formula for the left-sided centroid
m h , D A B L * ( α , β ) = | c h | 1 β x i c h x i β 1 β for β 0 x i c h x i 1 | c h | for β = 0 .

References

  1. Amari, S. α-Divergence Is Unique, Belonging to Both f-Divergence and Bregman Divergence Classes. IEEE Trans. Inf. Theory 2009, 55, 4925–4931. [Google Scholar] [CrossRef]
  2. Taneja, I.J.; Kumar, P. Relative Information of Type s, CsiszáR’s F-divergence, and Information Inequalities. Inf. Sci. 2004, 166, 105–125. [Google Scholar] [CrossRef]
  3. Cichocki, A.; Amari, S.i. Families of Alpha-Beta- and Gamma-Divergences: Flexible and Robust Measures of Similarities. Entropy 2010, 12, 1532–1568. [Google Scholar] [CrossRef]
  4. Banerjee, A.; Merugu, S.; Dhillon, I.S.; Ghosh, J. Clustering with Bregman Divergences. J. Mach. Learn. Res. 2005, 6, 1705–1749. [Google Scholar]
  5. Nielsen, F.; Nock, R. Sided and Symmetrized Bregman Centroids. IEEE Trans. Inf. Theory 2009, 55, 2882–2904. [Google Scholar] [CrossRef] [Green Version]
  6. Amari, S.I. Integration of Stochastic Models by Minimizing α-Divergence. Neural Comput. 2007, 19, 2780–2796. [Google Scholar] [CrossRef] [PubMed]
  7. Nielsen, F.; Nock, R. The Dual Voronoi Diagrams with Respect to Representational Bregman Divergences. In Proceedings of the 2009 Sixth International Symposium on Voronoi Diagrams, Copenhagen, Denmark, 23–26 June 2009; pp. 71–78. [Google Scholar] [CrossRef]
  8. Schwander, O.; Nielsen, F. Non-flat clustering with alpha-divergences. In Proceedings of the 2011 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Prague, Czech Repblic, 22–27 May 2011; pp. 2100–2103. [Google Scholar] [CrossRef]
  9. Nielsen, F.; Nock, R.; Amari, S.I. On Clustering Histograms with k-Means by Using Mixed α-Divergences. Entropy 2014, 16, 3273–3301. [Google Scholar] [CrossRef] [Green Version]
  10. Liu, M.; Vemuri, B.C.; Amari, S.; Nielsen, F. Total Bregman divergence and its applications to shape retrieval. In Proceedings of the 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Francisco, CA, USA, 13–18 June 2010; pp. 3463–3468. [Google Scholar] [CrossRef]
  11. Nock, R.; Nielsen, F.; Amari, S. On conformal divergences and their population minimizers. IEEE Trans. Inf. Theory 2016, 62, 527–538. [Google Scholar] [CrossRef]
  12. García-García, D.; von Luxburg, U.; Santos-Rodríguez, R. Risk-Based Generalizations of f-divergences. In Proceedings of the 28th International Conference on Machine Learning, ICML 2011, Bellevue, WA, USA, 28 June–2 July 2011; pp. 417–424. [Google Scholar]
  13. Chakraborty, S.; Das, S. k-Means clustering with a new divergence-based distance metric: Convergence and performance analysis. Pattern Recognit. Lett. 2017, 100, 67–73. [Google Scholar] [CrossRef]
  14. Cichocki, A.; Cruces, S.; Amari, S.I. Generalized Alpha-Beta Divergences and Their Application to Robust Nonnegative Matrix Factorization. Entropy 2011, 13, 134–170. [Google Scholar] [CrossRef] [Green Version]
  15. Sarmiento, A.; Durán-Díaz, I.; Cichocki, A.; Cruces, S. A Contrast Function Based on Generalized Divergences for Solving the Permutation Problem in Convolved Speech Mixtures. IEEE/ACM Trans. Audio Speech Lang. Process. 2015, 23, 1713–1726. [Google Scholar] [CrossRef]
  16. Mandal, A.; Cichocki, A. Non-Linear Canonical Correlation Analysis Using Alpha-Beta Divergence. Entropy 2013, 15, 2788–2804. [Google Scholar] [CrossRef] [Green Version]
  17. Yilmaz, E.; Gemmeke, J.F.; hamme, H.V. Noise-robust speech recognition with exemplar-based sparse representations using Alpha-Beta divergence. In Proceedings of the 2014 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Florence, Italy, 4–9 May 2014; pp. 5502–5506. [Google Scholar] [CrossRef]
  18. Olszewski, D.; Šter, B. Asymmetric clustering using the alpha-beta divergence. Pattern Recognit. 2014, 47, 2031–2041. [Google Scholar] [CrossRef]
  19. Ding, C.; He, X.; Simon, H.D. On the equivalence of nonnegative matrix factorization and spectral clustering. In Proceedings of the SIAM International Conference on Data Mining, Newport Beach, CA, USA, 21–23 April 2005. [Google Scholar]
  20. Kim, H.; Park, H. Nonnegative Matrix Factorization Based on Alternating Nonnegativity Constrained Least Squares and Active Set Method. SIAM J. Matrix Anal. Appl. 2008, 30, 713–730. [Google Scholar] [CrossRef] [Green Version]
  21. Ding, C.H.Q.; Li, T.; Jordan, M.I. Convex and Semi-Nonnegative Matrix Factorizations. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 32, 45–55. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  22. Ding, C.; Li, T.; Peng, W. On the equivalence between Non-negative Matrix Factorization and Probabilistic Latent Semantic Indexing. Comput. Stat. Data Anal. 2008, 52, 3913–3927. [Google Scholar] [CrossRef] [Green Version]
  23. Kimura, K.; Kudo, M.; Tanaka, Y. A column-wise update algorithm for nonnegative matrix factorization in Bregman divergence with an orthogonal constraint. Mach. Learn. 2016, 103, 285–306. [Google Scholar] [CrossRef] [Green Version]
  24. Lloyd, S. Least Squares Quantization in PCM. IEEE Trans. Inf. Theor. 2006, 28, 129–137. [Google Scholar] [CrossRef]
  25. Wei, B.; Gibson, J.D. Comparison of Distance Measures in Discrete Spectral Modeling. In Proceedings of the 9th DSP Workshop/1st Signal Process. Education Workshop, Hunt, TX, USA, 15–18 October 2000. [Google Scholar]
  26. Buzo, A.; Gray, A.; Gray, R.; Markel, J. Speech coding based upon vector quantization. IEEE Trans. Acoust. Speech Signal Process. 1980, 28, 562–574. [Google Scholar] [CrossRef] [Green Version]
  27. Dheeru, D.; Karra Taniskidou, E. UCI Machine Learning Repository, 2017. Available online: http://archive.ics.uci.edu/ml (accessed on 6 February 2019).
  28. Kuhn, H.W.; Yaw, B. The Hungarian method for the assignment problem. Nav. Res. Logist. Q. 1955, 2, 83–97. [Google Scholar] [CrossRef] [Green Version]
  29. Cichocki, A.; Zdunek, R.; Phan, A.H.; Amari, S.I. Nonnegative Matrix and Tensor Factorizations: Applications to Exploratory Multi-Way Data Analysis and Blind Source Separation; Wiley Publishing: Hoboken, NJ, USA, 2009. [Google Scholar]
  30. Kenan Yilmaz, Y.; Taylan Cemgil, A. Alpha/Beta Divergences and Tweedie Models. arXiv, 2012; arXiv:1209.4280. [Google Scholar]
  31. Ellis, D. Beat Tracking by Dynamic Programming. J. New Music Res. Spec. Issue Beat Tempo Extr. 2007, 36, 51–56. [Google Scholar] [CrossRef]
  32. Giannakopoulos, T.; Pikrakis, A. The Matlab Audio Analysis Library. In Introduction to Audio Analysis, A MATLAB Approach; Academic Press: Cambridge, MA, USA, 2014; pp. 233–240. [Google Scholar] [CrossRef]
Figure 1. Analysis of the convergence region of the α β -k-means algorithm. The region in blue shows the convex cone that guarantee the convergence of the algorithm to a local minimum for any dataset. Blue lines represent the boundaries of the convergence region for some values of the function exp 1 α 1 β 1 .
Figure 1. Analysis of the convergence region of the α β -k-means algorithm. The region in blue shows the convex cone that guarantee the convergence of the algorithm to a local minimum for any dataset. Blue lines represent the boundaries of the convergence region for some values of the function exp 1 α 1 β 1 .
Entropy 21 00196 g001
Figure 2. Generative models for dataset used in experiment 1. Each of the four mixture models have three components of Gaussian, Log-Gaussian, Poisson, and Binomial distribution, respectively.
Figure 2. Generative models for dataset used in experiment 1. Each of the four mixture models have three components of Gaussian, Log-Gaussian, Poisson, and Binomial distribution, respectively.
Entropy 21 00196 g002
Figure 3. Average ACC obtained with the right centroid α β -k-means algorithm for four different datasets.
Figure 3. Average ACC obtained with the right centroid α β -k-means algorithm for four different datasets.
Entropy 21 00196 g003
Figure 4. Average ACC obtained with the left-centroid α β -k-means algorithm for four different datasets.
Figure 4. Average ACC obtained with the left-centroid α β -k-means algorithm for four different datasets.
Entropy 21 00196 g004
Figure 5. Performance of the α β -k-means algorithm in terms of accuracy for DFT-based descriptors considering K = 3 classes and K = 5 classes.
Figure 5. Performance of the α β -k-means algorithm in terms of accuracy for DFT-based descriptors considering K = 3 classes and K = 5 classes.
Entropy 21 00196 g005
Figure 6. Performance of the α β -k-means algorithm in terms of accuracy for acoustic descriptors considering K = 3 classes and K = 5 classes.
Figure 6. Performance of the α β -k-means algorithm in terms of accuracy for acoustic descriptors considering K = 3 classes and K = 5 classes.
Entropy 21 00196 g006
Figure 7. Performance of the α β -k-means algorithm in terms of average accuracy over 50 trials for two UCI datasets: Iris and Wine.
Figure 7. Performance of the α β -k-means algorithm in terms of average accuracy over 50 trials for two UCI datasets: Iris and Wine.
Entropy 21 00196 g007
Table 1. Average and standard deviation of ACC for the synthetic dataset by applying the right-type α β -k-means algorithm for some specific values of ( α , β ) .
Table 1. Average and standard deviation of ACC for the synthetic dataset by applying the right-type α β -k-means algorithm for some specific values of ( α , β ) .
( α β ) ParametrizationGenerative Model
GaussianLog-GaussianPoissonBinomial
(1,1) Euclidean distance0.8784 ± 0.00560.9915 ± 0.00160.6948 ± 0.01430.7089 ± 0.0144
(0,0) Log-Euclidean distance0.8754 ± 0.00590.9909 ± 0.00170.7085 ± 0.00880.7216 ± 0.0081
(1,0) Kullback-Leibler divergence0.8783 ± 0.00560.9912 ± 0.00170.7057 ± 0.01020.7195 ± 0.0092
(1,−1) Itakura-Saito divergence0.8755 ± 0.00590.9909 ± 0.00170.7089 ± 0.00880.7220 ± 0.0082
(0.5,0.5) proportional to Hellinger distance0.8782 ± 0.00560.9913 ± 0.00170.7062 ± 0.00990.7199 ± 0.0092
Table 2. Clustering results in terms of accuracy of the α β -k-means algorithm for DFT-based descriptors for some specific distances and divergences.
Table 2. Clustering results in terms of accuracy of the α β -k-means algorithm for DFT-based descriptors for some specific distances and divergences.
( α , β )Distance or Divergence K = 3 K = 5
(1,1)Euclidean distance0.52670.4700
(0,0)Log-Euclidean distance0.92330.4880
(1,0)Kullback-Leibler divergence0.96330.6200
(1,−1)Itakura-Saito divergence0.89670.4820
(0.5,0.5)proportional to the Hellinger distance0.95670.6560
(1.7,−1)-0.97670.6240
(0.6,0.4)-0.95670.6860
Table 3. Clustering results of the α β -k-means algorithm for acoustic descriptors for some specific distances and divergences.
Table 3. Clustering results of the α β -k-means algorithm for acoustic descriptors for some specific distances and divergences.
( α , β )Distance or Divergence K = 3 K = 5
(1,1)Euclidean distance0.84670.4980
(0,0)Log-Euclidean distance0.88330.5580
(1,0)Kullback-Leibler divergence0.88000.6420
(1,−1)Itakura-Saito divergence0.88000.5920
(0.5,0.5)proportional to the Hellinger distance0.88000.6460
(0,0.2)-0.91670.6820
(0.8,−0.7)-0.91330.7000
Table 4. Clustering results of the α β -k-means algorithm for two UCI datasets: Iris and Wine, for some specific distances and divergences.
Table 4. Clustering results of the α β -k-means algorithm for two UCI datasets: Iris and Wine, for some specific distances and divergences.
( α , β )Distance or DivergenceIris DatasetWine Dataset
(1,1)Euclidean distance0.89330.7022
(0,0)Log-Euclidean distance0.96000.9157
(1,0)Kullback-Leibler divergence0.95760.7135
(1,−1)Itakura-Saito divergence0.96000.9157
(0.5,0.5)proportional to the Hellinger distance0.95360.7135
(−1,1.2)-0.96000.9663

Share and Cite

MDPI and ACS Style

Sarmiento, A.; Fondón, I.; Durán-Díaz, I.; Cruces, S. Centroid-Based Clustering with αβ-Divergences. Entropy 2019, 21, 196. https://doi.org/10.3390/e21020196

AMA Style

Sarmiento A, Fondón I, Durán-Díaz I, Cruces S. Centroid-Based Clustering with αβ-Divergences. Entropy. 2019; 21(2):196. https://doi.org/10.3390/e21020196

Chicago/Turabian Style

Sarmiento, Auxiliadora, Irene Fondón, Iván Durán-Díaz, and Sergio Cruces. 2019. "Centroid-Based Clustering with αβ-Divergences" Entropy 21, no. 2: 196. https://doi.org/10.3390/e21020196

APA Style

Sarmiento, A., Fondón, I., Durán-Díaz, I., & Cruces, S. (2019). Centroid-Based Clustering with αβ-Divergences. Entropy, 21(2), 196. https://doi.org/10.3390/e21020196

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop