Next Article in Journal
Three-Dimensional Human Body Reconstruction Using Dual-View Normal Maps
Next Article in Special Issue
NLAPSMjSO-EDA: A Nonlinear Shrinking Population Strategy Algorithm for Elite Group Exploration with Symmetry Applications
Previous Article in Journal
Effect of a Uniform Axial Magnetic Field on the Spatiotemporal Evolution of Thermocapillary Convection in a High Prandtl Fluid Under Microgravity
Previous Article in Special Issue
A Novel Reinforcement Learning-Based Particle Swarm Optimization Algorithm for Better Symmetry between Convergence Speed and Diversity
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Weighted Multiview K-Means Clustering with L2 Regularization

1
Department of Applied Mathematics, Chung Yuan Christian University, Chung-Li, Taoyuan 32023, Taiwan
2
Department of Information Systems, Satya Wacana Christian University, Salatiga 50711, Indonesia
3
Department of Statistics, Government College Gilgit, Gilgit Baltistan 15100, Pakistan
4
School of Artificial Intelligence, Xidian University, Xi’an 710126, China
5
Department of Computer Engineering, College of Computers and Information Technology, Taif University, P.O. Box 11099, Taif 21944, Saudi Arabia
6
Department of Software Engineering, College of Computer Science and Engineering, University of Jeddah, Jeddah 23218, Saudi Arabia
*
Author to whom correspondence should be addressed.
Symmetry 2024, 16(12), 1646; https://doi.org/10.3390/sym16121646
Submission received: 28 October 2024 / Revised: 5 December 2024 / Accepted: 9 December 2024 / Published: 12 December 2024
(This article belongs to the Special Issue Symmetry in Intelligent Algorithms)

Abstract

:
In the era of big data, cloud, internet of things, virtual communities, and interconnected networks, the prominence of multiview data is undeniable. This type of data encapsulates diverse feature components across varying perspectives, each offering unique insights into the same underlying samples. Despite being sourced from diverse settings and domains, these data serve the common purpose of describing the same samples, establishing a significant interrelation among them. Thus, there arises a necessity for the development of multiview clustering methodologies capable of leveraging the wealth of information available across multiple views. This study introduces two novel weighted multiview k-means algorithms, W-MV-KM and weighted multiview k-means using L2 regularization, W-MV-KM-L2, designed specifically for clustering multiview data. These algorithms incorporate feature weights and view weights within the k-means (KM) framework. Our approach emphasizes a weighted multiview learning strategy, which assigns varying degrees of importance to individual views. We evaluate the clustering performance of our algorithms on seven diverse benchmark datasets spanning dermatology, textual, image, and digit domains. Through extensive experimentation and comparisons with existing methods, we showcase the superior effectiveness and utility of our newly introduced W-MV-KM-L2 algorithm.

1. Introduction

Clustering is one of the basic methods to uncover the underline structure within data. Clustering divides a dataset into clusters, where data points contained within each cluster display strong similarity, while those belonging to separate clusters demonstrate significant dissimilarity [1,2]. In the era of the internet of things (IoT) and big data, a substantial volume of data is generated across various areas of interest. The collected data originate from various sources or multiple views with diverse representations, including documents in different languages, mobile internet usage, social networks, multimedia data analysis, information and image retrieval, healthcare applications, and scientific research. In multivariate analysis, clustering methods can generally be categorized into two main approaches: probability model-based clustering and nonparametric clustering. This study focuses on the latter, emphasizing nonparametric clustering techniques. Several widely used nonparametric clustering methods are used in the literature, including the k-means [3,4], fuzzy c-means [5,6,7], and possibilistic c-means [8,9]. However, while these clustering techniques may work well for low-dimensional data, they are not well suited for high-dimensional datasets. Multiview learning is gaining ground as a promising path within the field of pattern recognition and machine learning. Unlike conventional single-view clustering, multiview clustering leverages valuable feature knowledge from diverse perspectives to enhance clustering performance; multiview data encompass more shared, complementary, and redundant information. Numerous methodologies for multiview clustering methods have been proposed in order to increase the efficacy.
In the multiview clustering literature, Bickel and Scheffer [10] were the first to introduce multiview clustering as a method for handling multiview data. Since then, numerous researchers in the multiview clustering literature have contributed to advancements across various domains, such as in the paradigm of unsupervised multiview clustering (Yang and Hussain [11] and Hussain et al. [12]). In the field of multiview graph learning, various researchers have made significant contributions [13,14,15,16,17]. For incomplete multiview clustering, algorithms were proposed by [18,19,20,21,22]. In the area of tensor-based multiview clustering, notable algorithms have been proposed by Liu and Song [23], Li et al. [24], Chen et al. [25] and Benjamin and Yang [26]. Regarding multiview feature learning, several schemes have been developed. For example, Xu et al. [27] proposed feature learning for contrastive clustering, Zhang et al. [28] introduced sparse feature selection for images, Tang et al. [29] introduced consensus learning for unsupervised feature selection, and Xu et al. [30] presented a weighted multiview scheme for feature selection. In weighted multiview clustering, several researchers have made valuable contributions. Jiang et al. [31] introduced collaborative fuzzy clustering, Wang and Chen [32] proposed minimax optimization for fuzzy clustering, Yang and Sinaga [33] developed a feature reduction algorithm for k-means, and Yang and Sinaga [34] also proposed a collaborative scheme for FCM clustering. Some recent weighted multiview clustering techniques have been proposed. Benjamin and Yang [35] presented an algorithm for PCM using L2 regularization, Khan et al. [36] introduced a weighted concept-based scheme for incomplete multiview clustering, Zhou et al. [37] proposed evidential c-means clustering in a multiview scenario, Liu et al. [38] introduced an adaptive scheme with feature preference, Ouadfel and Abd Elaziz [39] proposed a multi-objective gradient optimizer approach-based scheme, Houfar et al. [40] introduced a binary weighted scheme in a multiview scenario, and Liu et al. [41] proposed a weighted scheme in evidential clustering in a multiview paradigm.
The rest of this study is structured as follows: Section 2 introduces a literature review. In Section 3, we introduce the weighted multiview k-means algorithms designed for clustering multiview datasets. Experimental results and comparisons with existing methods are discussed in Section 4. Discussions are outlined in Section 5. Lastly, Section 6 concludes this study and offers some future recommendations, particularly suggesting the use of a point symmetry-based distance instead of Euclidean distance to better capture cluster symmetry behaviors [42,43].

2. Literature Review

In this section, we review some multiview clustering algorithms that we compared with our proposed algorithms. We use two hard clustering algorithms built on multiview k-means, and two soft clustering algorithms based on multiview fuzzy c-means are discussed.
Xu et al. [30] introduced a multiview clustering algorithm based on k-means, termed the multiview weighted algorithm with feature selection (WMCFS). This algorithm incorporates two approaches for data points and feature selection. The objective function of WMCFS, as described by Xu et al. [30], is as follows:
J W M C F S U , A , W = h = 1 s v h α k = 1 c i = 1 n z i k j = 1 d h w j h x i j h a k j h 2 + τ h s j = 1 d h w j h 2
The parameter α is utilized to regulate the sparsity of view weights, while τ is employed to regulate the sparsity of feature weights. In WMCFS, the balancing parameter τ is utilized to control the feature weights within each view.
Yang and Sinaga [33] introduced the feature reduction multiview k-means (FRMVKM) clustering algorithm. The objective function of FRMVKM [33] is defined as:
J F R M V K M U , A , V , W = h = 1 s ( v h ) α k = 1 c i = 1 n z i k j = 1 d h w j h δ j h x i j h a k j h 2 + n d h h = 1 s j = 1 d h w j h ln w j h δ j h
In this formulation, δ j h represents the parameter used to control the feature weights within the k-th cluster of each view, while α > 0 denotes the view weight.
Jiang et al. [31] presented another extension to the multiview clustering algorithm, termed WV-Co-FCM. In this approach, they accounted for varying weights for each view, incorporating parameters to control the distribution of these view weights. The objective function of WV-Co-FCM is expressed as:
J W V C o F C M U , A , V = h = 1 s v h i = 1 n k = 1 c z i k h m d i k h 2 + Δ h + λ h = 1 s v h ln v h
where v h is the h-th view weight with h = 1 s v h = 1 ,   λ > 0 used to regulate view weights, Δ h = i = 1 n α i k h k = 1 c z i k h 1 z i k h m 1 i = 1 n τ i k h k = 1 c z i k h 1 z i k h m 1 , and d i k h = j = 1 d h x i j h a k j h 2 .
WV-Co-FCM is designed to address multiview data, where weights play a crucial role in the final step. Specifically, without the influence of weights, the membership of the cluster center and each object are updated independently.
Wang and Chen [32] presented the Minmax-FCM clustering algorithm, which operates without a collaboration step and considers different weights for each view. Minmax-FCM is constructed based on the single-view FCM, where it evaluates the least distance between the membership matrix ( U * ) and cluster centers ( A h ) in each view. The extreme value within views ( v h ) influences the smallest separation between ( U * ) and ( A h ) . Consequently, the consistent clustering results in Minmax-FCM are achieved through Minmax optimization, aiming to minimize dissimilarities across diverse views. The formulation of Minmax-FCM [35] is as follows:
min U * , A h h = 1 s max v h h = 1 s h = 1 s v h α Q h
In the Minmax-FCM scenario, Q h = i = 1 n k = 1 c z i k * m x i h a k h 2 and α serve as user-defined parameters to regulate the distribution of views.
While previous works, such as WMCFS, FRMVKM, WVCoFCM, and Minmax-FCM, have made significant contributions to multiview clustering by focusing on feature selection, relevance filtering and different weights for views, there remains a gap in the effective integration of feature and view weights within the k-means framework. Specifically, existing algorithms focus either exclusively on feature selection (e.g., WMCFS) or weight adjustment (e.g., FRMVKM and WVCoFCM) but fail to comprehensively combine feature and view weights within a unified clustering strategy. Although the Minmax FCM takes view weights into account, it does not include a collaboration step, which further limits its ability to effectively capture the interdependencies between features and views. In response to this gap, we propose two novel algorithms, W-MV-KM and W-MV-KN-L2-Norm, both of which are based on the k-means algorithm and integrate feature weights and view weights in a synergistic manner. These algorithms provide an innovative approach by extending the k-means framework with weighted strategies to better capture the complex relationships in multiview datasets, improve clustering performance, and ensure more accurate feature selection. This contribution addresses the limitations of previous methods and presents a more comprehensive solution for multiview clustering.

3. Weighted Multiview K-Means Clustering

In this section, we introduce two weighted multiview k-means clustering algorithms tailored for multiview data. We additionally explore the clustering performance of weighted multiview k-means in comparison to existing multiview clustering algorithms. Consider a multiview dataset denoted as Χ , comprising h views and d h features, where Χ = x 1 , x 2 , , x n in the Euclidean space d h where x i = x i h h = 1 s . Let us consider the view vector weight v h 0 , 1 ,   V = v h 1 × s with h = 1 s v h = 1 . let Z = z i k n × c , where z i k = 1 assuming the data point x i belongs to the k-th cluster, z i k = 0 otherwise, i.e., z i k 0 , 1 , let W = w j 1 × d h where w j = w j h h = 1 s , let A = a 1 , a 2 , , a c be the c clusters. Therefore, we introduce our first weighted multiview k-means (W-MV-KM) algorithm, which assigns distinct weights to both views and features. Our objective is to devise a weighted scheme that aims to discern the significance of views as well as features within each view. The objective function for the proposed W-MV-FCM is
J V , Z , A , W = h = 1 s v h α i = 1 n k = 1 c z i k j = 1 d h w j h τ x i j h a k j h 2
Theorem 1.
Assume  z i k 0 , 1 let  W = w j 1 × d h , j = 1 d h w j h = 1  and  v h  is the view weight for the h-th view  h = 1 s v h = 1 .  The constraints for minimizing the objective function of W-MV-KM  J V , Z , A , W  are as follows:
a k j h = i = 1 n z i k x i j h i = 1 n z i k
z i k = 1   if   h = 1 s v h α j = 1 d h w j h τ x i j h a k j h 2 = 1 k c min h = 1 s v h α j = 1 d h w j h τ x i j h a k j h 2 0   otherwise
v h = h = 1 s i = 1 n k = 1 c z i k j = 1 d h w j h τ x i j h a k j h 2 i = 1 n k = 1 c z i k j = 1 d h w j h τ x i j h a k j h 2 1 α 1 1
w j h = j = 1 d h v h α i = 1 n k = 1 c z i k x i j h a k j h 2 v h α i = 1 n k = 1 c z i k x i j h a k j h 2 1 τ 1 1
Proof. 
For the optimization of the proposed W-MV-KM, the Lagrange multiplier technique is used. J V , Z , A , W = h = 1 s v h α i = 1 n k = 1 c z i k j = 1 d h w j h τ x i j h a k j h 2 + λ 1 h = 1 s v h 1 + λ 2 j = 1 d h w j h 1 regarding the cluster center, membership, view weight, and feature weight, the constraints for minimizing the W-MV-KM are as follows. Initially, we hold V, W, and Z fixed and solve the updated equation for the cluster center. By taking the partial derivative of the J V , Z , A , W with respect to a k j h and making them equal to zero, we have J a k j h = v h α i = 1 n z i k w j h τ 2 x i j h a k j h = 0 ; solving for a k j h , we obtain the updated Equation (6). Subsequently, with V, A, and W held constant, we solve the updated equation for the membership. By taking the partial derivative of the Lagrangian concerning z i k and equating it to 0, we obtain Equation (7). Assuming Z, A, and W are fixed, we proceed with the partial derivative of the Lagrangian J V , Z , A , W with respect to v h and equate the result to 0. J v h = α v h α 1 i = 1 n k = 1 c z i k j = 1 d h w j h τ x i j h a k j h 2 + λ 1 = 0 . We obtain the updated Equation (8) for the view weight. Given fixed Z, A, and V, we compute the partial derivative of the Lagrangian J V , Z , A , W with respect to v h and set it equal to zero, such that J w j h = v h α i = 1 n k = 1 c z i k τ w j h τ 1 x i j h a k j h 2 + λ 2 = 0 . Upon solving the equation, we can drive the updating Equation (9) for the feature weight w j h . Therefore, the proposed W-MV-KM algorithm can be summarized as Algorithm 1. □
Algorithm 1. The W-MV-KM Algorithm.
Input :   Dataset   X = x 1 , x 2 , , x n   with   x i = x i h h = 1 s   and   x i h = x i j h j = 1 d h ,
  c   number   of   cluster   and   ε > 0
Output :   a k j h ,   z i k ,   w j h   and   v h .
Initialization :   Randomly   generate   Z t . Initialize the feature weight W h 0 = w j h 1 × d h , 1 j d h   with   j = 1 d w j = 1   by   w j h = 1 / d h   , initialize view weight
V 0 = v h 1 × s   u s e r   may   define   v h = 1 / s h , and set t = 1.
Step 1:  Compute   the   cluster   centers   A h ( t )   using   Z t 1 by Equation (6).
Step 2:  Update   the   membership   matrix   Z t   using   A h ( t ) , V t 1   and   W h t 1 by
Equation (7).
Step 3:  Update   the   view   weight   V t   by   ( 8 )   using   Z t , A h ( t )   and   W h t 1 .
Step 4:  Update   the   feature   weight   W h t   using   ( 9 )   Z t , A h ( t )   and   V t .
Step 5:  If   W h t W h t 1 < ε , then stop;
Else go back to Step 1.
In W-MV-KM, α   and   τ represent two exponent parameters utilized for regulating the distribution of feature weights and view weights. When both α   and   τ are set to zero, the W-MV-KM objective function simplifies to the k-means objective function.
We aim to discover an optimal combination of α   and   τ values to minimize the intra-cluster variance of the h-th view and j-th features. The parameter α regulates the distribution of view weights. Let F 1 ( h ) = k = 1 c i = 1 n j = 1 d h z i k w j h τ x i j h a k j h 2 then Equation (8) will can be expressed as v h = 1 / h = 1 1 F 1 h / F 1 h 1 α 1 , where F 1 h is the intra-cluster variance. When values of α 1 , Equation (8) suggests that only one view is chosen for clustering, which is not suitable for handling multiview datasets. As the values of α increase, the distribution of view weights becomes more uniform, indicating that more views will contribute to the clustering process. Conversely, the parameter τ , which governs the distribution of feature weights, can be examined as follows: Let us use F 2 ( h ) = h = 1 s v h α k = 1 c i = 1 n z i k x i j h a k j h 2 to represent the intra-cluster variance of the j-th feature. With this, Equation (9) can be expressed as w j h = 1 / j = 1 d h F 2 j / F 2 j 1 τ 1 . If τ = 1 , this indicates the selection of only one feature. Then, we need to select values of τ that stabilize the algorithm and ultimately yield favorable results for W-MV-KM.
The clustering outcomes are influenced by the parameters α   and   τ in the W-MV-KM algorithm, posing challenges in both the control and estimation of results. Hence, to manage and stabilize these effects, we set the values of both α   and   τ to 2. This is how we apply regularization to the objective function for both the feature and view weights. We subsequently introduce another multiview learning approach for W-MV-KM, integrating L2 regularization, which we refer to as W-MV-KM-L2. In W-MV-KM-L2, L2 regularization is applied to both the features and view weights. The objective function of the W-MV-KM-L2 is formulated in the following:
J V , Z , A , W = h = 1 s v h 2 i = 1 n k = 1 c j = 1 d h z i k w j h 2 x i j h a k j h 2 + α h = 1 s v h 2 2 + τ j = 1 d h w j h 2 2 h = 1 s v h = 1 ,   j = 1 d h w j h = 1 , j = 1 d h w j h 2 2 = j = 1 d h w j h 2 ,   h = 1 s v j h 2 2 = h = 1 s v h 2 .
In Algorithm 2, we incorporate the exponential parameters α   and   τ with the regularization framework. This integration serves to control the distribution of both features and views. In W-MV-KM, the parameters α   and   τ are determined by the user, while in W-MV-KM-L2, we calculate the parameters within the regularization terms by means of the equations α = c / d h and τ = i = 1 n x i h x i h ¯ 2 / n . Likewise, we derive the updating equations for the W-MV-KM-L2 objective function through the utilization of the Lagrange multiplier. Consequently, we derive Theorem 2, as presented below.
Theorem 2.
Suppose  h = 1 s v h = 1 ,   v h 0 , 1   and   j = 1 d h w j h = 1 ,   w j 0 , 1 .  The necessary conditions for minimizing the objective function  J W M V K M L 2 V , Z , A , W  are
a k j h = i = 1 n z i k x i j h i = 1 n z i k
z i k = 1   if   h = 1 s v h 2 j = 1 d h w j h 2 x i j h a k j h 2 = 1 k c min h = 1 s v h 2 j = 1 d h w j h 2 x i j h a k j h 2 0   otherwise
v h = h = 1 s i = 1 n k = 1 c z i k j = 1 d h w j h 2 x i j h a k j h 2 + α i = 1 n k = 1 c z i k j = 1 d h w j h 2 x i j h a k j h 2 + α 1
w j h = j = 1 d h v h 2 i = 1 n k = 1 c z i k x i j h a k j h 2 + τ v h 2 i = 1 n k = 1 c z i k x i j h a k j h 2 + τ 1
Proof. 
To demonstrate the optimization problem described in Equation (10), we employ the Lagrange multipliers. Consequently, we construct the Lagrangian function as follows. We first set W, V and Z to solve the updated equation for cluster centers a k j h . We achieve this by computing partial derivatives with respect to a k j h and subsequently setting them equal to zero, yielding the following equation J a k j h = v h 2 i = 1 n z i k w j h 2 x i j h a k j h 2 = 0 . Upon solving for a k j h , we obtained the updated Equation (11) a k j h = i = 1 n z i k x i j h i = 1 n z i k . Subsequently, by fixing A, W, and V, we proceed to solve the updated equation for the membership degree z i k . After computing the partial derivative of J V , Z , A , W with respect to z i k and setting it equal to zero, we obtain J z i k = h = 1 s v h 2 j = 1 d h w j h 2 x i j h a k j h 2 = 0 . From this, we obtain the updated Equation (12). Consequently, with fixed Z, A, and W, we compute the partial derivative of J Z , A , W , V with respect to v h and equating it to zero, we have J v h = 2 v h i = 1 n k = 1 c z i k j = 1 d h w j h 2 x i j h a k j h 2 + 2 α v h + λ 1 = 0 . Further solving this equation leads to the derivation of the updated Equation (13). We compute the partial derivative of J V , Z , A , W with respect to w j h and set it to zero. The results obtained are as follows J w j h = v h 2 i = 1 n k = 1 c z i k 2 w j h x i j h a k j h 2 + 2 τ w j h + λ 3 = 0 ; therefore, the updated Equation (14) is obtained. □
Algorithm 2. The W-MV-KM-L2 Algorithm
Input :   Dataset   X = x 1 , x 2 , , x n   with   x i = x i h h = 1 s  and  x i h = x i j h j = 1 d h , “c” number
of cluster and ε > 0 .
Output: a k j h ,   z i k ,   w j h   and   v h .
Initialization :   Randomly   generate   Z t .   Initialize   the   feature   weight   W h 0 = w j h 1 × d h ,
1 j d h   with   j = 1 d w j = 1   by   w j h = 1 / d h   ,   initialize   view   weight   V 0 = v h 1 × s
u s e r   may   define   v h = 1 / s h , and set t = 1.
Step 1:    Calculate   α = c / d h   and   τ = i = 1 n x i h x i h ¯ 2 / n  where τ is the covariance of
the data points in the h t h   view .
Step 2:  Compute   the   cluster   centers   A h ( t )   using   Z t 1 by Equation (11).
Step 3:  Update   the   membership   matrix   Z t   using   A h ( t ) , W h t 1   and   V t 1 by
Equation (12).
Step 4:  Update   the   view   weight   V t   using   ( 13 )   Z t , W h t 1   and   A h ( t ) .
Step 5:  Update   the   feature   weight   W h t   using   Equation   ( 14 )   Z t , V t   and   A h ( t ) .
Step 6:  If   W h t W h t 1 < ε , then stop.
Otherwise, go back to Step 2.
The W-MV-KM-L2 incorporates parameters α   and   τ , which play a vital role in determining the classification of view weights and feature weights. To regulate the behavior of α   and   τ , we select appropriate values of parameters in a manner that ensures the stability of the algorithm, thus enabling the computation of final estimates. The parameters α   and   τ in the W-MV-KM-L2 objective function are computed through α = c / d h and τ = i = 1 n x i h x i h ¯ 2 / n , where c signifies the total number of clusters and d h signifies the total number of features in each view. If the value of d h is larger than the value of α , it becomes very small and consequently does not significantly contribute to the determination of view weights. The role of c becomes crucial in regulating the values of α . Let G 1 ( h ) = k = 1 c i = 1 n j = 1 d h z i k w j h 2 x i j h a k j h 2 be the intra-cluster variances in the h-th view. Then, it becomes v h = h = 1 s G 1 h + α h / G 1 h + α h . Smaller values of α h will result in smaller weights, whereas large values of α h will lead to larger view weights v h . This demonstrates that α h supervises the weights in each view. Thus, view weights are utilized to update the feature weights w j h in each view. Meanwhile, the other parameter τ is estimated using the sample variance for each view, such that τ = i = 1 n x i h x i h ¯ 2 / n and it is utilized to regulate the distribution in each view. From Equation (14), we regard G 2 j = h = 1 s v h 2 k = 1 c i = 1 n z i k x i j h a k j h 2 as the intra-cluster variance of the features in view. Thus, Equation (14) becomes w j h = j = 1 d h G 2 j + τ j G 2 j + τ j . If the values of τ j are larger, then the feature weights also become larger, whereas if the values of τ j are smaller, then the feature weights also become smaller. Hence, τ j is employed to diminish the feature weights that are irrelevant. Consequently, features with larger weights are prioritized over those with smaller weights.
Computational Complexity: We conduct an analysis of computational complexity for both the proposed W-MV-KM algorithm and the W-MV-KM-L2 algorithm. These algorithms can be segmented into four parts: (1) compute the cluster center a k h with O(nsc); (2) update the membership degree z i k with O(nsc2d); (3) update the view weight v h with O n s c d h 2 ; (4) update the feature weights v h with O n s c d h 2 . The total computational complexity for the W-MV-KM and W-MV-KM-L2 algorithm is O n s c 2 d h + n s c 2 d h . where n represents the number of data points, s denotes the number of views, c signifies the number of clusters, and d h stands for the dimensionality of data points.

4. Experimental Comparisons and Results

In this section, two synthetic and seven real-life datasets are used to demonstrate the performance of the proposed W-MV-KM-L2 algorithm. We compare the W-MV-KM-L2 algorithm with five existing algorithms, W-MV-KM, WMCFS [30], WV-Co-FCM [31], Minmax-FCM [32], and FRMVKM [33], in this section. For the experiments, we use the same initializations for the cluster center, feature weight, and view weights. For measuring the clustering performance, we use the following evaluation measures: accuracy rate (AR), Jaccard index (JI) [44], Fowlkes–Mallows index (FMI) [45], Rand Index (RI) [46], and normalized mutual information (NMI) [47]. Accuracy rate (AR) is the term used to describe the proportion of correct predictions or classifications made by a model or system, expressed as a percentage of the total number of predictions or classifications attempted. The Jaccard index (JI) [44] shows how similar two sets are by comparing the number of items they have in common to the overall count of unique items in both sets. The FMI [45] is a metric utilized to assess the similarity between two clustering results. It calculates the geometric mean of the pairwise similarity between points in the same clusters across two different clustering results. The RI [46] measures how similar two sets of clusters are. It counts pairs of data points that are either in the same cluster in both sets or different clusters in both sets, and the NMI [47] is like a friendly compass that helps us navigate through clusters of data. It measures how well two different clustering results match up, considering the size and arrangement of the clusters. The greater the magnitudes of AR, JI, FMI, RI, and NMI, the more exceptional the clustering performance becomes.
The experiments are executed in Matlab 2020a with identical initializations. They are repeated 100 times using distinct random initializations of the WMCFS, FRMVKM, WV-Co-FCM, Minmax-FCM, W-MV-KM, and W-MV-KM-L2 algorithms. For Minmax-FCM and WV-Co-FCM, the fuzziness index m is 2 for the methods.
Example 1.
We generate a dataset with two distinct views using a Gaussian mixture model:  h = 1 2 1 2 N μ i h , i h where h represents the view index. These two views are then merged to form the samples  x i 1 , x i 2 , c k . The dataset consists of two clusters across both views, with 1500 data points sampled for each view. For the first view, the cluster means are set to  μ 1 1 = 1.5 1.5  and  μ 2 1 = 7 7 . In the second view, the cluster means are  μ 1 2 = 2.5 2.5  and  μ 2 2 = 5 5 . The covariance matrices  1 i = 0.25 0 0 0.25 , i = 1,…,4, are applied to both views. Each view consists of two main features:  x 1 1   &   x 2 1  for view 1, and  x 1 2   &   x 2 2  for view 2. Additionally, we introduce one noise feature per view. For view 1, the noise feature is  x 3 1 , sampled from a uniform distribution in the range [0, 5], and for view 2, the noise feature is  x 3 2 , sampled from a uniform distribution in the range [0, 10], named Artificial 1 dataset. A 2D and 3D graph illustrating the dataset is shown in Figure 1a–d for both view 1 and view 2, respectively.
Figure 2 shows the clustering visualization of the two algorithms W-MV-KM and W-MV-KM-L2 for both view 1 and view 2 in 3D representation. Table 1 shows the clustering performance of the proposed algorithms with the existing algorithms, where the W-MV-KM-L2 algorithm performs well, as compared to other algorithms in all evaluation measures. In this example, we try different values for two parameters α   and   τ and report the values of the algorithms in Table 2.
In W-MV-KM, we try different values for the two parameters, as mentioned in Table 2. If we choose for both parameters α = τ = 1 , then the algorithm is unable to give results; if we choose α = 2 , then the algorithm performs, but the results are very poor as compared to other existing algorithms. Thus, we propose W-MV-KM-L2, where the values of α   and   τ are determined by the following formulas, as discussed in Section 3. In the W-MV-KM algorithm, if we keep values of α < 2 , then the algorithm is unable to perform. So, for this, it should be α 2 . If we increase the parameter values α = τ 2 , even then, the clustering performance is not improved. In both algorithms, W-MV-KM and W-MV-KM-L2, the view weight should be 2 because if it is < 2 , then only one view is selected, which is against the basic principal of multiview clustering.
Example 2.
In this example, we use the synthetic data used in [35], called Syn1. To implement the dataset in the proposed and existing algorithms, we keep different initializations for existing algorithms: we choose  α = 2   and   τ  = 0.001 for WMCFS; for FRMVKM, we assume  α = 2  ; and for WV-Co-FCM, τ = 1 . 3 . These serve as the parameters. With the given initializations, we apply these algorithms to the dataset, and their clustering performance is summarized in Table 3. The results indicate that W-MV-KM-L2 obtained the highest values across all evaluation metrics.
Example 3.
In this example, we use the synthetic dataset used in [48] to implement the W-MV-KM-L2, W-MV-KM, Minmax-FCM, WV-Co-FCM, FRMVKM, and WMCFS in the Syn500 dataset. For our initializations, we choose  α = 4   and   τ  = 0.01 for WMCFS; for FRMVKM, we assume  α = 2 ; and for WV-Co-FCM,  τ = 4 . For Minmax-FCM and WV-Co-FCM, the fuzziness index is m = 2. With the given initializations, we apply these algorithms to the dataset, and their clustering performance is summarized in Table 4. The results indicate that W-MV-KM-L2 obtained the highest values across all evaluation metrics.
We first normalized all the real datasets. We tested the performance of our proposed algorithms with seven real datasets. Seven real-world multiview datasets serve as the benchmark to assess the efficacy of the W-MV-KM-L2 algorithm. These are the Minist4 dataset [49], Handwritten4 (HW) [50], Caltech2 dataset [51], UCI Derm dataset [52], HumanEva 3D Motion [53,54,55] dataset, UCI 3views dataset [50], and Microsoft Research Cambridge Volume 1 (MSRC-V1) dataset [56], respectively. The characteristics of these seven real datasets are displayed in Table 5 in terms of cluster number c, data type s, the data number n, the feature dimension dh, and the view numbers.
Example 4.
In this example, we use seven real datasets for comparing W-MV-KM-L2 with all the algorithms, such as WMCFS, FRMVKM, WV-Co-FCM, Minmax-FCM and W-MV-KM. Various combinations of  α  and  τ  will result in different distributions for the view and feature weight vectors, respectively. This is the case for WMCFS and WV-Co-FCM, where  α  and  τ  are user defined. For FRMVKM, we use  α = 2  for all datasets. In W-MV-KM-L2, the regularization parameters  α  and  τ  are used to control the sparsity of the view weights and feature weights, respectively. The parameters  α  and  τ  in W-MV-KM-L2 are simply calculated by  α = c / d h  and  τ = i = 1 n x i h x i h ¯ 2 / n . We compare the results of the proposed algorithms with existing algorithms. To achieve this, we utilize real datasets and apply the algorithms using the assumed values of  τ  in Table 6.
For each given value of α , we ran 100 simulations for each algorithm using different seed generations across all clustering methods and recorded their clustering performance metrics. This process was repeated for all real datasets listed in Table 6. We present only the overall average mean and standard deviation of the performance measures for each algorithm and compare them with the results of W-MV-KM-L2. The outcomes, summarized in Table 7, show that W-MV-KM-L2 consistently achieves higher results for AR, RI, NMI, JI, and FMI across all real datasets.
Example 5.
In this example, we compare the number of iterations required for convergence, along with the running time, for the WMCFS, FRMVKM, WV-Co-FCM, Minmax-FCM, W-MV-KM, and W-MV-KM-L2 clustering algorithms. The number of iterations needed for convergence is reported in Table 8. The initial parameters are set as follows: m = 2 is used as the fuzzifier value for Minimax-FCM and WV-Co-FCM. The convergence (error) tolerance for all the algorithms is set to be  1 × 10 3 . The running time per second achieved by all the six algorithms is reported in Table 9, and the lowest running time for any dataset achieved by any algorithm is highlighted in bold face.
Example 6.
In this example, we use the multiview cluster validity indices proposed by Yang and Hussain [11], the multiview Dunn (MV-Dunn) index, and multiview generalized Dunn (MV-G-Dunn) index. We implement the multiview validity indices only based on k-means such as WMCFS, FRMVKM, and W-MV-KM-L2. We consider the WMCFS+MV-Dunn, WMCFS+MV-G-Dunn, FRMVKM+MV-Dunn, FRMVKM+MV-G-Dunn, W-MV-KM-L2+MV-Dunn, and W-MV-KM-L2+MV-G-Dunn regularization with 40 different initializations. The estimated number of clusters, along with their percentages, determined using the MV-Dunn and MV-G-Dunn indices, are presented in Table 10 for WMCFS, Table 11 for FRMVKM, and Table 12 for WMVKML2.

5. Discussion

Defining the structure of multiview data through L2 regularization for k-means and incorporating view weights and feature weights into its objective function presents a significant challenge. In this manuscript, we introduce two innovative clustering algorithms for multiview data, namely W-MV-KM and W-MV-KM-L2, building on the foundation of k-means. In W-MV-KM, users are tasked with assigning weights to both features and views. The parameters α   and   τ are utilized to regulate the distribution of weights and features in W-MV-KM, respectively. As the values of α   and   τ are user-determined for these weights, identifying an optimal combination of view and feature weights can present a challenge. To tackle this concern, we suggest an alternative method called W-MV-KM-L2, which integrates L2 regularization into the W-MV-KM framework. The parameters α   and   τ utilized for the regularization of W-MV-KM-L2 are subsequently estimated. This approach enhances the significance of results for both view and feature weights, ultimately leading to an improvement in clustering performance.

6. Conclusions

In the empirical findings for real-world datasets for both the proposed algorithm and existing algorithms, W-MV-KM-L2 outperformed in clustering performance. In our future work, we aim to explore making the W-MV-KM-L2 algorithm independent of initialization requirements. Additionally, we plan to enhance the W-MV-KM-L2 algorithm by incorporating the ability to determine an unsupervised optimal number of clusters. Since point symmetry-based distance [42] is more effective at representing cluster symmetry behavior [43] compared to Euclidean distance, we plan to extend the W-MV-KM-L2 algorithm in future work by incorporating point symmetry-based distance to enhance its ability to present cluster symmetry.

Author Contributions

Conceptualization, I.H. and Y.N.; methodology, I.H., Y.N., M.A. and A.H.; software, I.H. and Y.N.; validation, I.H., Y.N., M.A. and A.H.; formal analysis, I.H., Y.N., M.A. and A.H.; investigation, I.H., Y.N., M.A., A.H. and F.A.A.; data curation, A.H., H.M.M., F.A.A. and N.M.R.; writing—original draft preparation, I.H. and M.A.; writing—review and editing, I.H, M.A.; visualization, A.H., H.M.M., F.A.A. and N.M.R.; supervision, I.H.; funding acquisition, H.M.M., F.A.A. and N.M.R. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Taif University, Saudi Arabia, under project number (TU-DSPP2024-139).

Data Availability Statement

All data generated or analyzed during this study are included in this article.

Acknowledgments

The authors wish to acknowledge the anonymous reviewers and editors for their valuable feedback and insights.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Jain, A.K.; Dubes, R.C. Algorithms for Clustering Data; Prentice-Hall, Inc.: Hoboken, NJ, USA, 1988. [Google Scholar]
  2. Kaufman, L.; Rousseeuw, P.J. Finding Groups in Data: An Introduction to Cluster Analysis; Wiley: New York, NY, USA, 1990. [Google Scholar]
  3. MacQueen, J. Some methods for classification and analysis of multivariate observations. In Proceedings of the 5th Berkeley Symposium on Mathematical Statistics and Probability; University of California Press: Berkeley, CA, USA, 1967; Volume 1, pp. 281–297. [Google Scholar]
  4. Jain, A.K. Data clustering: 50 years beyond K-means. Pattern Recognit. Lett. 2020, 31, 651–666. [Google Scholar] [CrossRef]
  5. Bezdek, J.C. Pattern Recognition with Fuzzy Objective Function Algorithms; Kluwer: Norwell, MA, USA, 1981. [Google Scholar]
  6. Ruspini, E.H.; Bezdek, J.C.; Keller, J.M. Fuzzy clustering: A historical perspective. IEEE Comput. Intell. Mag. 2019, 14, 45–55. [Google Scholar] [CrossRef]
  7. Yang, M.S.; Nataliani, Y. Robust-learning fuzzy c-means clustering algorithm with unknown number of clusters. Pattern Recognit. 2017, 71, 45–59. [Google Scholar] [CrossRef]
  8. Krishnapuram, R.; Keller, J.M. A possibilistic approach to clustering. IEEE Trans. Fuzzy Syst. 1993, 1, 98–110. [Google Scholar] [CrossRef]
  9. Yang, M.S.; Benjamin, J.B. Sparse possibilistic c-means clustering with Lasso. Pattern Recognit. 2023, 138, 109348. [Google Scholar] [CrossRef]
  10. Bickel, S.; Scheffer, T. Multi-view clustering. In Proceedings of the 4th IEEE International Conference on Data Mining, Brighton, UK, 1–4 November 2004; pp. 19–26. [Google Scholar]
  11. Yang, M.S.; Hussain, I. Unsupervised multi-view K-means clustering algorithm. IEEE Access 2023, 11, 13574–13593. [Google Scholar] [CrossRef]
  12. Hussain, I.; Sinaga, K.P.; Yang, M.S. Unsupervised multiview fuzzy c-means clustering algorithm. Electronics 2023, 12, 4467. [Google Scholar] [CrossRef]
  13. Shi, S.; Nie, F.; Wang, R.; Li, X. Multi-view clustering via nonnegative and orthogonal graph reconstruction. IEEE Trans. Neural Netw. Learn. Syst. 2021, 34, 201–214. [Google Scholar] [CrossRef]
  14. Yu, W.; Xing, L.; Nie, F.; Li, X. Multi-View Fuzzy Clustering Based on Anchor Graph. IEEE Trans. Fuzzy Syst. 2023, 32, 755–766. [Google Scholar] [CrossRef]
  15. Wang, R.; Wang, P.; Wu, D.; Sun, Z.; Nie, F.; Li, X. Multi-view and multi-order structured graph learning. IEEE Trans. Neural Netw. Learn. Syst. 2023, 35, 14437–14448. [Google Scholar] [CrossRef]
  16. Yang, B.; Zhang, X.; Li, Z.; Nie, F.; Wang, F. Efficient multi-view K-means clustering with multiple anchor graphs. IEEE Trans. Knowl. Data Eng. 2022, 35, 6887–6900. [Google Scholar] [CrossRef]
  17. Zhao, M.; Yang, W.; Nie, F. Auto-weighted orthogonal and nonnegative graph reconstruction for multi-view clustering. Inf. Sci. 2023, 632, 324–339. [Google Scholar] [CrossRef]
  18. Yang, M.; Li, Y.; Hu, P.; Bai, J.; Lv, J.; Peng, X. Robust multi-view clustering with incomplete information. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 1055–1069. [Google Scholar] [CrossRef]
  19. Lin, Y.; Gou, Y.; Liu, Z.; Li, B.; Lv, J.; Peng, X. Completer: Incomplete multi-view clustering via contrastive prediction. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 11174–11183. [Google Scholar]
  20. Cai, Y.; Che, H.; Pan, B.; Leung, M.F.; Liu, C.; Wen, S. Projected cross-view learning for unbalanced incomplete multi-view clustering. Inf. Fusion 2024, 105, 102245. [Google Scholar] [CrossRef]
  21. Chao, G.; Jiang, Y.; Chu, D. Incomplete Contrastive Multi-View Clustering with High-Confidence Guiding. In Proceedings of the AAAI Conference on Artificial Intelligence, Vancouver, BC, Canada, 20–27 February 2024; Volume 38, pp. 11221–11229. [Google Scholar]
  22. Zhan, J.; Cai, M. A cost-minimized two-stage three-way dynamic consensus mechanism for social network-large scale group decision-making: Utilizing K-nearest neighbors for incomplete fuzzy preference relations. Expert Syst. Appl. 2024, 263, 125705. [Google Scholar] [CrossRef]
  23. Liu, Z.; Song, P. Deep low-rank tensor embedding for, multi-view subspace clustering. Expert Syst. Appl. 2024, 237, 121518. [Google Scholar] [CrossRef]
  24. Li, J.; Gao, Q.; Wang, Q.; Yang, M.; Xia, W. Orthogonal non-negative tensor factorization based multi-view clustering. Adv. Neural Inf. Process. Syst. 2024, 36, 18186–18202. [Google Scholar]
  25. Chen, M.S.; Li, Z.Y.; Lin, J.Q.; Wang, C.D.; Huang, D. Sparse Graph Tensor Learning for Multi-View Spectral Clustering. IEEE Trans. Emerg. Top. Comput. Intell. 2024, 8, 3534–3543. [Google Scholar] [CrossRef]
  26. Benjamin, J.B.; Yang, M.S. Tensor-Based Possibilistic C-Means Clustering. IEEE Trans. Fuzzy Syst. 2024, 32, 5939–5950. [Google Scholar] [CrossRef]
  27. Xu, J.; Tang, H.; Ren, Y.; Peng, L.; Zhu, X.; He, L. Multi-level feature learning for contrastive multi-view clustering. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 16051–16060. [Google Scholar]
  28. Zhang, Y.; Wu, J.; Cai, Z.; Philip, S.Y. Multi-view multi-label learning with sparse feature selection for image annotation. IEEE Trans. Multimed. 2020, 22, 2844–2857. [Google Scholar] [CrossRef]
  29. Tang, C.; Zheng, X.; Liu, X.; Zhang, W.; Zhang, J.; Xiong, J.; Wang, L. Cross-view locality preserved diversity and consensus learning for multi-view unsupervised feature selection. IEEE Trans. Knowl. Data Eng. 2021, 34, 4705–4716. [Google Scholar] [CrossRef]
  30. Xu, Y.M.; Wang, C.D.; Lai, J.H. Weighted multi-view clustering with feature selection. Pattern Recognit. 2016, 53, 25–35. [Google Scholar] [CrossRef]
  31. Jiang, Y.; Chung, F.L.; Wang, S.; Deng, Z.; Wang, J.; Qian, P. Collaborative fuzzy clustering from multiple weighted views. IEEE Trans. Cybern. 2014, 45, 688–701. [Google Scholar] [CrossRef] [PubMed]
  32. Wang, Y.; Chen, L. Multi-view fuzzy clustering with minimax optimization for effective clustering of data from multiple sources. Expert Syst. Appl. 2017, 72, 457–466. [Google Scholar] [CrossRef]
  33. Yang, M.S.; Sinaga, K.P. A feature-reduction multi-view k-means clustering algorithm. IEEE Access 2019, 7, 114472–114486. [Google Scholar] [CrossRef]
  34. Yang, M.S.; Sinaga, K.P. Collaborative feature-weighted multi-view fuzzy c-means clustering. Pattern Recognit. 2021, 119, 108064. [Google Scholar] [CrossRef]
  35. Benjamin, J.B.M.; Yang, M.S. Weighted multiview possibilistic c-means clustering with L2 regularization. IEEE Trans. Fuzzy Syst. 2021, 30, 1357–1370. [Google Scholar] [CrossRef]
  36. Khan, G.A.; Khan, J.; Anwar, T.; Ashraf, Z.; Javed, M.H.; Diallo, B. Weighted Concept Factorization based Incomplete Multi-view Clustering. IEEE Trans. Artif. Intell. 2024, 5, 5699–5708. [Google Scholar] [CrossRef]
  37. Zhou, K.; Zhu, Y.; Guo, M.; Jiang, M. MvWECM: Multi-view Weighted Evidential C-Means clustering. Pattern Recognit. 2024, 159, 111108. [Google Scholar] [CrossRef]
  38. Liu, Z.; Huang, H.; Letchmunan, S.; Deveci, M. Adaptive weighted multi-view evidential clustering with feature preference. Knowl.-Based Syst. 2024, 294, 111770. [Google Scholar] [CrossRef]
  39. Ouadfel, S.; Abd Elaziz, M. A multi-objective gradient optimizer approach-based weighted multi-view clustering. Eng. Appl. Artif. Intell. 2021, 106, 104480. [Google Scholar] [CrossRef]
  40. Houfar, K.; Samai, D.; Dornaika, F.; Benlamoudi, A.; Bensid, K.; Taleb-Ahmed, A. Automatically weighted binary multi-view clustering via deep initialization (AW-BMVC). Pattern Recognit. 2023, 137, 109281. [Google Scholar] [CrossRef]
  41. Liu, Z.; Huang, H.; Letchmunan, S. weighted multi-view evidential clustering. In International Conference on Artificial Neural Networks; Springer Nature: Cham, Switzerland, 2023; pp. 265–277. [Google Scholar]
  42. Bandyopadhyay, S.; Saha, S. A point symmetry-based clustering technique for automatic evolution of clusters. IEEE Trans. Knowl. Data Eng. 2008, 20, 1441–1457. [Google Scholar] [CrossRef]
  43. Fu, J.; Hsiao, C.A. Decoding intelligence via symmetry and asymmetry. Sci. Rep. 2024, 14, 12525. [Google Scholar] [CrossRef] [PubMed]
  44. Jaccard, P. Distribution de la flore alpine dans le bassin des Dranses et dans quelques régions voisines. Bull. Soc. Vaudoise Sci. Nat. 1901, 37, 241–272. [Google Scholar]
  45. Fowlkes, E.B.; Mallows, C.L. A method for comparing two hierarchical clusterings. J. Am. Stat. Assoc. 1983, 78, 553–569. [Google Scholar] [CrossRef]
  46. Rand, W.M. Objective criteria for the evaluation of clustering methods. J. Am. Stat. Assoc. 1971, 66, 846–850. [Google Scholar] [CrossRef]
  47. Cover, T.M. Elements of Information Theory; John Wiley & Sons: Hoboken, NJ, USA, 1999. [Google Scholar]
  48. Rong, W.; Zhuo, E.; Peng, H.; Chen, J.; Wang, H.; Han, C.; Cai, H. Learning a consensus affinity matrix for multi-view clustering via subspaces merging on Grassmann manifold. Inf. Sci. 2021, 547, 68–87. [Google Scholar] [CrossRef]
  49. Fei-Fei, L.; Fergus, R.; Perona, P. Learning generative visual models from few training examples: An incremental bayesian approach tested on 101 object categories. In Proceedings of the 2004 Conference on Computer Vision and Pattern Recognition Workshop, Washington, DC, USA, 27 June–2 July 2004; p. 178. [Google Scholar]
  50. Dua, D.; Graff, C. UCI Machine Learning Repository. 2017, Volume 7, p. 62. Available online: http://archive.ics.uci.edu/ml (accessed on 8 December 2024).
  51. Xu, P.; Deng, Z.; Choi, K.S.; Cao, L.; Wang, S. Multi-view information-theoretic co-clustering for co-occurrence data. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; Volume 33, pp. 379–386. [Google Scholar]
  52. Tao, H.; Hou, C.; Qian, Y.; Zhu, J.; Yi, D. Latent complete row space recovery for multi-view subspace clustering. IEEE Trans. Image Process. 2020, 29, 8083–8096. [Google Scholar] [CrossRef]
  53. Ma, Z.; Nie, F.; Yang, Y.; Uijlings, J.R.; Sebe, N.; Hauptmann, A.G. Discriminating joint feature analysis for multimedia data understanding. IEEE Trans. Multimed. 2012, 14, 1662–1672. [Google Scholar] [CrossRef]
  54. Ning, H.; Xu, W.; Gong, Y.; Huang, T. Discriminative learning of visual words for 3D human pose estimation. In Proceedings of the 2008 IEEE Conference on Computer Vision and Pattern Recognition, Anchorage, AK, USA, 23–28 June 2008; pp. 1–8. [Google Scholar]
  55. Yang, Y.; Xu, D.; Nie, F.; Yan, S.; Zhuang, Y. Image clustering using local discriminant models and global integration. IEEE Trans. Image Process. 2010, 19, 2761–2773. [Google Scholar] [CrossRef] [PubMed]
  56. Winn, J.; Jojic, N. Locus: Learning object classes with unsupervised segmentation. In Proceedings of the Tenth IEEE International Conference on Computer Vision, ICCV’05, Washington, DC, USA, 17–20 October 2005; Volume 1, pp. 756–763. [Google Scholar]
Figure 1. A 2D and 3D graphical representation of 2-cluster dataset with 2 views, each containing 1500 data points. (a) 2D visualization of view 1; (b) 2D visualization of view 2; (c) 3D visualization of view 1; and (d) 3D visualization of view 2.
Figure 1. A 2D and 3D graphical representation of 2-cluster dataset with 2 views, each containing 1500 data points. (a) 2D visualization of view 1; (b) 2D visualization of view 2; (c) 3D visualization of view 1; and (d) 3D visualization of view 2.
Symmetry 16 01646 g001
Figure 2. Clustering performance of WMVKM and WMVKML2 in 3D. (a) WMVKM visualization of view 1; (b) WMVKM visualization of view 2; (c) WMVKML2 visualization of view 1; and (d) WMVKML2 visualization of view 2.
Figure 2. Clustering performance of WMVKM and WMVKML2 in 3D. (a) WMVKM visualization of view 1; (b) WMVKM visualization of view 2; (c) WMVKML2 visualization of view 1; and (d) WMVKML2 visualization of view 2.
Symmetry 16 01646 g002
Table 1. Clustering performances of WMCFS, FRMVKM, WV-Co-FCM, Minmax-FCM, W-MV-KM and W-MV-KM-L2 for Artificial 1 multiview data based on average (STD).
Table 1. Clustering performances of WMCFS, FRMVKM, WV-Co-FCM, Minmax-FCM, W-MV-KM and W-MV-KM-L2 for Artificial 1 multiview data based on average (STD).
MethodsARJIFMIRINMI
WMCFS0.844 (0.212)0.764 (0.310)0.826 (0.230)0.826 (0.230)0.647 (0.464)
FRMVKM0.960 (0.116)0.926 (0.187)0.949 (0.132)0.949 (0.132)0.889 (0.280)
WV-Co-FCM0.500 (0.000)0.500 (0.000)0.707 (0.000)0.500 (0.000)0.706 (0.000)
Minmax-FCM0.500 (0.000)0.500 (0.000)0.707 (0.000)0.500 (0.000)0.706 (0.000)
W-MV-KM0.718 (0.311)0.686 (0.333)0.765 (0.764)0.768 (0.250)0.530 (0.499)
W-MV-KM-L20.960 (0.137)0.947 (0.180)0.960 (0.135)0.960 (0.135)0.920 (0.270)
Table 2. Clustering sensitivity analysis of the parameters α   and   τ .
Table 2. Clustering sensitivity analysis of the parameters α   and   τ .
Algorithms α τ
W-MV-KM1, 2, 3…1, 2, 3…
W-MV-KM-L2 α = c / d h τ = i = 1 n x i h x i h ¯ 2 / n
Table 3. Clustering performances of WMCFS, FRMVKM, WV-Co-FCM, Minmax-FCM, W-MV-KM and W-MV-KM-L2 for Syn1 multiview data based on average (STD).
Table 3. Clustering performances of WMCFS, FRMVKM, WV-Co-FCM, Minmax-FCM, W-MV-KM and W-MV-KM-L2 for Syn1 multiview data based on average (STD).
MethodsARJIFMIRINMI
WMCFS0.861 (0.218)0.807 (0.303)0.855 (0.227)0.855 (0.227)0.710 (0.453)
FRMVKM0.924 (0.154)0.870 (0.249)0.906 (0.180)0.906 (0.180)0.805 (0.371)
WV-Co-FCM0.943 (0.132)0.902 (0.184)0.938 (0.125)0.929 (0.147)0.842 (0.291)
Minmax-FCM0.500 (0.000)0.499 (0.000)0.707 (0.000)0.499 (0.000)0.000 (0.000)
W-MV-KM0.856 (0.291)0.868 (0.265)0.901 (0.198)0.901 (0.198)0.802 (0.297)
W-MV-KM-L20.972 (0.112)0.960 (0.158)0.970 (0.118)0.970 (0.118)0.940 (0.237)
Table 4. Clustering performances of WMCFS, FRMVKM, WV-Co-FCM, Minmax-FCM, W-MV-KM and W-MV-KM-L2 for Syn500 multiview data based on average (STD).
Table 4. Clustering performances of WMCFS, FRMVKM, WV-Co-FCM, Minmax-FCM, W-MV-KM and W-MV-KM-L2 for Syn500 multiview data based on average (STD).
MethodsARJIFMIRINMI
WMCFS0.671 (0.138)0.429 (0.116)0.593 (0.101)0.592 (0.102)0.152 (0.178)
FRMVKM0.873 (0.015)0.644 (0.029)0.784 (0.023)0.778 (0.021)0.532 (0.062)
WV-Co-FCM0.740 (0.000)0.446 (0.000)0.617 (0.000)0.614 (0.000)0.179 (0.000)
Minmax-FCM0.500 (0.000)0.499 (0.000)0.706 (0.000)0.499 (0.000)0.000 (0.000)
W-MV-KM0.740 (0.152)0.499 (0.118)0.658 (0.106)0.656 (0.104)0.265 (0.189)
W-MV-KM-L20.880 (0.027)0.657 (0.017)0.793 (0.013)0.784 (0.013)0.523 (0.027)
Table 5. Clustering performances of WMCFS, FRMVKM, WV-Co-FCM, Minmax-FCM, W-MV-KM and W-MV-KM-L2 for Syn500 multiview data based on average (STD).
Table 5. Clustering performances of WMCFS, FRMVKM, WV-Co-FCM, Minmax-FCM, W-MV-KM and W-MV-KM-L2 for Syn500 multiview data based on average (STD).
MethodsCnsd1d2d3d4d5
Caltech2320332300256---
HumanEva 3D Motion5500024848---
Derm636621112---
Minist444000330930--
UCI 3views1020003240766--
HW1256102000476216476-
MSRC-V17210524576512256254
Table 6. Clustering performances of WMCFS, FRMVKM, WV-Co-FCM, Minmax-FCM, W-MV-KM and W-MV-KM-L2 for Syn500 multiview data based on average (STD).
Table 6. Clustering performances of WMCFS, FRMVKM, WV-Co-FCM, Minmax-FCM, W-MV-KM and W-MV-KM-L2 for Syn500 multiview data based on average (STD).
Minist4Handwritten4Caltech2DermHuman Eva 3DUCI 3viewsMSRC-V1
WMCFS0.10.0010.0010.0010.010.010.01
WV-Co-FCM3532423
Table 7. Clustering performances of WMCFS, FRMVKM, WV-Co-FCM, Minmax-FCM, W-MV-KM and W-MV-KM-L2 for seven real multiview data based on average (STD).
Table 7. Clustering performances of WMCFS, FRMVKM, WV-Co-FCM, Minmax-FCM, W-MV-KM and W-MV-KM-L2 for seven real multiview data based on average (STD).
DatasetMethodsARJIFMIRINMI
Minist4WMCFS0.661 (0.166)0.445 (0.157)0.600 (0.145)0.795 (0.075)0.483 (0.175)
FRMVKM0.444 (0.163)0.267 (0.103)0.411 (0.123)0.698 (0.062)0.229 (0.170)
WV-Co-FCM0.250 (0.000)0.250 (0.000)0.500 (0.000)0.250 (0.000)0.000 (0.000)
Minmax-FCM0.250 (0.000)0.250 (0.000)0.500 (0.000)0.250 (0.000)0.000 (0.000)
W-MV-KM0.209 (0.016)0.171 (0.001)0.293 (0.001)0.609 (0.002)0.043 (0.000)
W-MV-KM-L20.789 (0.126)0.568 (0.097)0.720 (0.083)0.853 (0.051)0.637 (0.085)
HW1256WMCFS0.470 (0.055)0.284 (0.026)0.444 (0.032)0.877 (0.010)0.522 (0.028)
FRMVKM0.445 (0.078)0.260 (0.056)0.412 (0.072)0.869 (0.056)0.459 (0.072)
WV-Co-FCM0.100 (0.000)0.100 (0.000)0.000 (0.000)0.100 (0.000)0.316 (0.000)
Minmax-FCM0.200 (0.000)0.142 (0.000)0.340 (0.000)0.529 (0.000)0.264 (0.000)
W-MV-KM0.322 (0.185)0.186 (0.092)0.338 (0.110)0.676 (0.325)0.300 (0.226)
W-MV-KM-L20.470 (0.055)0.284 (0.026)0.444 (0.032)0.877 (0.010)0.522 (0.028)
Caltech2WMCFS0.748 (0.150)0.558 (0.141)0.707 (0.114)0.792 (0.082)0.543 (0.144)
FRMVKM0.404 (0.076)0.343 (0.046)0.532 (0.054)0.513 (0.058)0.144 (0.120)
WV-Co-FCM0.394 (0.000)0.354 (0.000)0.595 (0.000)0.354 (0.000)0.000 (0.000)
Minmax-FCM0.390 (0.006)0.351 (0.006)0.587 (0.015)0.362 (0.019)0.013 (0.012)
W-MV-KM0.437 (0.068)0.337 (0.018)0.517 (0.019)0.546 (0.051)0.132 (0.050)
W-MV-KM-L20.758 (0.161)0.591 (0.160)0.731 (0.126)0.808 (0.092)0.600 (0.161)
DermWMCFS0.673 (0.093)0.536 (0.099)0.695 (0.083)0.868 (0.040)0.734 (0.067)
FRMVKM0.677 (0.085)0.476 (0.070)0.643 (0.062)0.850 (0.034)0.680 (0.054)
WV-Co-FCM0.475 (0.038)0.290 (0.023)0.488 (0.023)0.636 (0.040)0.322 (0.038)
Minmax-FCM0.306 (0.000)0.199 (0.000)0.446 (0.000)0.199 (0.000)0.000 (0.000)
W-MV-KM0.633 (0.135)0.501 (0.128)0.668 (0.105)0.815 (0.170)0.669 (0.192)
W-MV-KM-L20.734 (0.117)0.629 (0.131)0.769 (0.099)0.881 (0.131)0.797 (0.168)
HumanEva 3D motionWMCFS0.448 (0.057)0.304 (0.027)0.480 (0.031)0.722 (0.038)0.412 (0.038)
FRMVKM0.200 (0.000)0.200 (0.000)0.447 (0.000)0.304 (0.000)0.000 (0.000)
WV-Co-FCM0.200 (0.000)0.200 (0.000)0.447 (0.000)0.200 (0.000)0.000 (0.000)
Minmax-FCM0.339 (0.000)0.227 (0.001)0.473 (0.001)0.621 (0.007)0.324 (0.005)
W-MV-KM0.200 (0.000)0.200 (0.000)0.447 (0.000)0.200 (0.000)0.000 (0.000)
W-MV-KM-L20.691 (0.114)0.533 (0.093)0.693 (0.077)0.865 (0.044)0.691 (0.072)
UCI 3viewsWMCFS0.471 (0.067)0.285 (0.033)0.446 (0.035)0.87 (0.079)0.526 (0.060)
FRMVKM0.316 (0.044)0.202 (0.021)0.351 (0.027)0.813 (0.029)0.375 (0.037)
WV-Co-FCM0.100 (0.000)0.100 (0.000)0.316 (0.000)0.100 (0.000)0.000 (0.000)
Minmax-FCM0.200 (0.000)0.142 (0.000)0.340 (0.000)0.529 (0.000)0.264 (0.000)
W-MV-KM0.228 (0.043)0.154 (0.016)0.277 (0.012)0.789 (0.203)0.336 (0.099)
W-MV-KM-L20.654 (0.070)0.450 (0.051)0.620 (0.048)0.920 (0.011)0.697 (0.032)
MSRC-VIWMCFS0.543 (0.080)0.314 (0.049)0.477 (0.055)0.844 (0.020)0.513 (0.054)
FRMVKM0.297 (0.047)0.137 (0.017)0.240 (0.240)0.782 (0.007)0.225 (0.033)
WV-Co-FCM0.143 (0.000)0.139 (0.000)0.373 (0.000)0.139 (0.000)0.000 (0.000)
Minmax-FCM0.286 (0.005)0.181 (0.000)0.371 (0.002)0.559 (0.007)0.234 (0.002)
W-MV-KM0.358 (0.036)0.163 (0.010)0.281 (0.015)0.786 (0.009)0.298 (0.024)
W-MV-KM-L20.590 (0.078)0.343 (0.044)0.511 (0.049)0.856 (0.016)0.547 (0.045)
Table 8. Number of iterations to converge the algorithms.
Table 8. Number of iterations to converge the algorithms.
DatasetsWMCFSFRMVKMWV-Co-FCMMinmax-FCMWMVKMW-MV-KM-L2
Minist485>=20>=502419
Handwritten436>=20>=501629
Caltech234>=20>=5073
Derm35>=20>=50611
HumanEva 3D Motion192>=20>=5035
UCI3views34>=20>=502617
MSRC-V185>=20>=502419
Table 9. The running time (RT) achieved through the proposed and other existing algorithms: WMCFS, FR-MV-KM, WV-Co-FCM, Minmax-FCM, W-MV-KM and W-MV-KM-L2 (the best is marked in boldface).
Table 9. The running time (RT) achieved through the proposed and other existing algorithms: WMCFS, FR-MV-KM, WV-Co-FCM, Minmax-FCM, W-MV-KM and W-MV-KM-L2 (the best is marked in boldface).
DatasetsWMCFSFRMVKMWV-Co-FCMMinmax-FCMWMVKMW-MV-KM-L2
Minist40.57900.4200204.36608.76101.93900.9830
Handwritten40.79800.7420424.691012.70804.96506.0090
Caltech20.62100.541030.02907.78701.69200.4800
Derm0.21000.260015.44100.85100.48400.3130
HumanEva 3D Motion2.57800.3790172.980015.76000.92500.6910
UCI3views0.74300.6440296.311013.17008.32403.9570
MSRC-V10.32800.328047.37102.64500.98300.7260
Table 10. Results of WMCFS with MV-Dunn and MV-G-Dunn.
Table 10. Results of WMCFS with MV-Dunn and MV-G-Dunn.
Datasets\IndicesMV-DunnMV-G-Dunn
Artificial 12 (65%)2 (70%)
Syn12 (50%)2 (72.5%)
Syn5002 (42.5%)2 (17.5%)
Caltech23 (10%)3 (0%)
HumanEva 3D motion5 (2.5%)5 (0%)
Derm6 (10%)6 (0%)
Minist44 (15%)4 (0%)
UCI3views10 (2.5%)10 (0%)
HW125610 (2.5%)10 (0%)
MSRC-IV7 (5%)7 (0%)
Table 11. Results of FRMVKM with MV-Dunn and MV-G-Dunn.
Table 11. Results of FRMVKM with MV-Dunn and MV-G-Dunn.
Datasets\IndicesMV-DunnMV-G-Dunn
Artificial 12 (100%)2 (2.5%)
Syn12 (100%)2 (0%)
Syn5002 (100%)2 (0%)
Caltech23 (0%)3 (0%)
HumanEva 3D motion5 (0%)5 (0%)
Derm6 (0%)6 (0%)
Minist44 (0%)4 (0%)
UCI3views10 (0%)10 (0%)
HW125610 (0%)10 (0%)
MSRC-IV7 (0%)7 (0%)
Table 12. Results of WMVKML2 with MV-Dunn and MV-G-Dunn.
Table 12. Results of WMVKML2 with MV-Dunn and MV-G-Dunn.
Datasets\IndicesMV-DunnMV-G-Dunn
Artificial 12 (100%)2 (2.5%)
Syn12 (100%)2 (0%)
Syn5002 (100%)2 (0%)
Caltech23 (2.5%)3 (0%)
HumanEva 3D motion5 (0%)5 (0%)
Derm6 (2.5%)6 (0%)
Minist44 (0%)4 (0%)
UCI3views10 (0%)10 (0%)
HW125610 (0%)10 (0%)
MSRC-IV7 (0%)7 (0%)
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Hussain, I.; Nataliani, Y.; Ali, M.; Hussain, A.; Mujlid, H.M.; Almaliki, F.A.; Rahimi, N.M. Weighted Multiview K-Means Clustering with L2 Regularization. Symmetry 2024, 16, 1646. https://doi.org/10.3390/sym16121646

AMA Style

Hussain I, Nataliani Y, Ali M, Hussain A, Mujlid HM, Almaliki FA, Rahimi NM. Weighted Multiview K-Means Clustering with L2 Regularization. Symmetry. 2024; 16(12):1646. https://doi.org/10.3390/sym16121646

Chicago/Turabian Style

Hussain, Ishtiaq, Yessica Nataliani, Mehboob Ali, Atif Hussain, Hana M. Mujlid, Faris A. Almaliki, and Nouf M. Rahimi. 2024. "Weighted Multiview K-Means Clustering with L2 Regularization" Symmetry 16, no. 12: 1646. https://doi.org/10.3390/sym16121646

APA Style

Hussain, I., Nataliani, Y., Ali, M., Hussain, A., Mujlid, H. M., Almaliki, F. A., & Rahimi, N. M. (2024). Weighted Multiview K-Means Clustering with L2 Regularization. Symmetry, 16(12), 1646. https://doi.org/10.3390/sym16121646

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop