Next Article in Journal
k-Path-Connectivity of Completely Balanced Tripartite Graphs
Next Article in Special Issue
Application of Deep Learning and Neural Network to Speeding Ticket and Insurance Claim Count Data
Previous Article in Journal
On r-Ideals and m-k-Ideals in BN-Algebras
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Robust Spectral Clustering Incorporating Statistical Sub-Graph Affinity Model

School of Communication and Information Engineering, Xi’an University of Posts and Telecommunications, Xi’an 710121, China
*
Author to whom correspondence should be addressed.
Axioms 2022, 11(6), 269; https://doi.org/10.3390/axioms11060269
Submission received: 16 May 2022 / Revised: 29 May 2022 / Accepted: 29 May 2022 / Published: 5 June 2022
(This article belongs to the Special Issue Machine Learning: Theory, Algorithms and Applications)

Abstract

:
Hyperspectral image (HSI) clustering is a challenging work due to its high complexity. Subspace clustering has been proven to successfully excavate the intrinsic relationships between data points, while traditional subspace clustering methods ignore the inherent structural information between data points. This study uses graph convolutional subspace clustering (GCSC) for robust HSI clustering. The model remaps the self-expression of the data to non-Euclidean domains, which can generate a robust graph embedding dictionary. The EKGCSC model can achieve a globally optimal closed-form solution by using a subspace clustering model with the Frobenius norm and a Gaussian kernel function, making it easier to implement, train, and apply. However, the presence of noise can have a noteworthy negative impact on the segmentation performance. To diminish the impact of image noise, the concept of sub-graph affinity is introduced, where each node in the primary graph is modeled as a sub-graph describing the neighborhood around the node. A statistical sub-graph affinity matrix is then constructed based on the statistical relationships between sub-graphs of connected nodes in the primary graph, thus counteracting the uncertainty image noise by using more information. The model used in this work was named statistical sub-graph affinity kernel graph convolutional subspace clustering (SSAKGCSC). Experiment results on Salinas, Indian Pines, Pavia Center, and Pavia University data sets showed that the SSAKGCSC model can achieve improved segmentation performance and better noise resistance ability.

1. Introduction

Hyperspectral images (HSI) contain rich spectral and spatial information, which have been widely used in various fields, such as geological exploration, ocean monitoring, medical imaging, and forensics [1]. HSI classification, which aims to sort each pixel with a specific label, is the basis of HSI applications [2]. Supervised classification methods were the most commonly used for HSI classification [3,4]. In recent years deep learning models [5,6] and convolutional neural networks (CNNs) [7] have made significant progress in HSI classification. Due to the high cost of labeling training data, supervised classification methods are not available in HSI scenarios. Moreover, the supervised classification methods are challenging to handle unknown objects, since they are molded with existing classes.
To avoid the costs of manual data annotation, many works are devoted to expanding unsupervised HSI classification methods, i.e., HSI clustering. The HSI clustering is dedicated to seeking the intrinsic relationship between data points and determining labels in unsupervised rules automatically [8]. The critical point of HSI clustering is to measure the similarity between data points [9]. Traditional clustering methods often use Euclidean distance as a similarity measurement, such as K-means [10]. Due to unreliable measures, measures, mixed pixels, and the redundant banding problems [11], making the HSI clustering is extremely challenging [12]. Recently, subspace clustering [13] has become increasingly focused on HSI clustering [14] due to its ability to process high-dimensional data and its reliable performance. Technically, subspace clustering attempts to describe the data points as a linear combination of a self-expression dictionary in the same subspace [15]. The subspace clustering models typically include two parts, namely self-representation [16] and spectral clustering (SC) [17].
To expand the capability of subspace clustering, many works have been devoted to generating a robust affinity matrix. For example, sparse subspace clustering (SSC) [18] uses the l 1 - norm to produce a sparse affinity matrix; low-rank subspace clustering (LRSC) [19] uses a nuclear norm to enforce the affinity matrix to be low-rank. By taking the spectral and spatial information into consideration, Zhang et al. proposed a spectral-spatial subspace clustering (S4C) [14]. By mapping the data into higher kernel spaces, kernel subspace clustering [20] became a nonlinear extension of the subspace clustering model. In [12], a modified kernel subspace clustering was utilized in HSI clustering.
The subspace clustering models based on Euclidean data often ignore the inherent graph structure information contained in HSI data points. HSI data points are easily affected by noise due to the interference of imaging spectrometers and the atmosphere [21]. Therefore, the traditional subspace model is sensitive to noisy images and abnormal data in the Euclidean domain. By transforming the HSI clustering into a clustering problem of non-Euclidean domain, the additional coefficients can be alleviated or even avoided by structural data.
The graph convolutional subspace clustering (GCSC) model [22] was used to recast the traditional subspace clustering into the non-Euclidean domain. By utilizing a graph convolutional self-representation model combining both graph and character information, the GCSC framework can circumvent noise data and tends to produce a more robust affinity than the traditional subspace clustering model.
For each pixel, a similarity matrix, or an affinity matrix, W is constructed. In the matrix, W u v represents the degree of similarity between pixel u and pixel v [23]. The eigenvectors of the similarity matrix provide a measurement to identify the most significant features in the image. Ideally, each prominent image region is reformed into a set of data points in the eigenvector domain. Using clustering algorithms [24,25,26], such as k-means or hierarchically clustering [27], the resulting clusters are identified and labeled. As long as there are a sufficient number of sound measurements, spectral clustering can be valid for segmenting the image.
Two crucial challenges faced when applying spectral clustering are the noise-corrupted measurements and the texture variations within image regions [28]. In such situations, the pixels may be misclassified using traditional spectral clustering models, given their heavy reliance on the affinity between connected nodes to construct the affinity weighted graph; Pardo [29] extended pixel-to-pixel node affinity and proposed the idea of using region statistics to better classify the edges of regions in noisy images. Assuming that every region has a uniform intensity, for each part, the standard deviation and the mean are calculated. Areas with similar statistics are sorted into a single region, which does not consider the texture or region structure detected by the sub-graph method. Therefore, it is necessary to provide more robust image segmentation methods in the presence of heavy noise contamination.
The main contribution of this paper is the introduction of a robust spectral clustering method for image segmentation under heavy noise. Instead of relying on the node affinity between connected nodes, the proposed method utilizes the statistical sub-graph affinity between connected nodes to construct the affinity-weighted graphs [30]. By considering the spatial-intensity relationships within sub-graphs, it is assumed that the statistical affinity between sub-graphs is insensitive to noise, which is random and usually does not show such structural relations. By applying clustering graphs, which relate to the structure of objects, Luo [31] improves the robustness of spectral clustering. Instead, in the current work, the sub-graph was used to extend the affinity matrix including neighboring regions; the sub-graph affinity utilized region statistics before the clustering step so that similar regions, not similar nodes, will be clustered more robustly.
The rest of this article is structured as follows. We first briefly review the subspace clustering, graph convolutional subspace clustering, and HSI clustering in Section 2. Then, we present a detailed introduction of our proposed method in Section 3. In Section 4, the experimental data sets and experimental results are introduced. Finally, we conclude with a summary and future work in Section 5.

2. Related Work

2.1. Notations

Throughout this article, boldface lowercase italics symbols (e.g., x ), boldface uppercase roman symbols (e.g., X ), and regular italics symbols (e.g., x i j ) denote vectors, matrices, and scalars, respectively. A graph is represented as G = ( V , ε , A ) , where V denotes the node set of the graph with v i V and | V | = N , ε indicates the edge set with ( v i , v i ) ε , and A R N × N stands for an adjacency matrix. We define the diagonal degree matrix of the graph as D R N × N , where D i j j A i j . The graph Laplacian is defined as L = D A , and its normalized version is given by L s y m = D 1 / 2 L D 1 / 2 . In this article, X T denotes the transposition of matrix X and I N denotes an identity matrix with the size of N . The Frobenius norm of a matrix is represented as X F = ( i j | x i j | 2 ) 1 / 2 and the trace of a matrix is defined as t r ( X ) .

2.2. Subspace Clustering Models

Let X = [ x 1 , x 2 , , x N ] R m × N be a collection of N data points { x i R m } i = 1 N and draw from a union of linear or affinity subspace S 1 S 2 S n , where N , m and n denote the number of data points, features, and subspace, respectively. The subspace clustering model for the given data set is defined as the following self-representation problem [18]:
min | | W | | p   s . t .   X W = X , s . t .   d i a g ( W ) = 0
where W R N × N denotes the self-expressive coefficient matrix and d i a g ( W ) = 0 enforces the diagonal elements of W to be zero so that the trivial solutions are avoided. | | W | | P denotes a p - n o r m of matrix W , e.g., | | W | | 1 (SSC) [18].
In the SSC model, the self-expressive coefficient matrix is conceived to be sparse so the self-representation problem is formulated as:
arg min W | | X W X | | F 2 + λ | | W | | 1 ,   s . t .   d i a g ( W ) = 0
Here, the l 1 - n o r m tends to cause a sparse coefficient matrix. By using a nuclear norm, LRSC [19] rewrites the self-expression problem as:
arg min W | | X W X | | 2 , 1 2 + λ | | W | | ,   s . t .   d i a g ( W ) = 0
where | | | | and | | | | 2 , 1 denote the nuclear norm and l 2 , 1 - n o r m of a matrix. LRSC has been proven to incorporate the structure of data effectively. The above problem can be solved by using the alternating direction method of multipliers (ADMM [32]). Once the coefficient matrix W is found, the subspace clustering pursues the partition of affinity matrix A = ( 1 / 2 ) ( | W | + | W | T ) using the SC method [24].

2.3. GCSC Framework

By incorporating graph embedding into subspace clustering, we refer to the framework as GCSC. The aim of the GCSC framework is to use graph convolution to learn a robust affinity. For this target, we modify the traditional self-representation as follows:
X = X A ¯ Z , s . t .   d i a g ( Z ) = 0 .
Here, Z R N × N is the self-representing coefficient matrix and A ¯ = D ˜ 1 / 2 A ˜ D ˜ 1 / 2 denotes the normalized adjacency matrix with self-loops [28]. The X A ¯ Z operation will preserve the structure and inherent attributes of the graph by embedding the graph into vector space. We call (4) a graph convolutional self-representation. Similar to the traditional subspace clustering model, the GCSC framework can be rewritten as:
arg min Z 1 2 | | X A ¯ Z X | | q + λ 2 | | Z | | p ,   s . t .   d i a g ( Z ) = 0
where q and p denote any suitable matrix norm, such as p , q = 0 , ( 1 / 2 ) , 1 , 2 and nuclear norm, and λ is a trade-off coefficient. It is easy to prove that the traditional subspace models are a special case of this framework, i.e., the traditional subspace clustering models depend only on data features. For example, when the Frobenius norm and p = 1 (5) becomes an extension of the classical SSC [18], while q is l 2 , 1 n o r m and p is nuclear norm, (5) degenerates to LRSC [19]. Equation (5) can be efficiently solved by the same methods adopted in traditional subspace clustering methods. Once the self-represented coefficient matrix is obtained, the spectral clustering (SC) can be used to generate clustering results.

2.4. HSI Clustering Using the GCSC Models

Two essential matters need to be solved before we use the GCSC model. First, HSI data usually contain many spectral bands with redundancy, so it is difficult to obtain acceptable performance using only spectral features. Second, the HSI uses Euclidean data, whereas GCSC models are based on graph-structured data [29].
To solve the first problem, we use principal component analysis (PCA) to reduce the spectral dimension by retaining the top d PCs [30]. This not only reduces the redundant information contained in the HSI data but also improves the computational efficiency during model training. To take spectral and spatial information into consideration, we represent each data point by extracting three-dimensional data blocks [31]. Specifically, every data point is represented by the central and neighboring pixels. This method has been widely used in different HSI spectral-spatial classification models [32]. For the second problem, a k-nearest neighbor (k-NN) graph was constructed to substitute the graph structure of the data points. Specifically, each data point is viewed as a node in the graph, and the k-NNs of x i consist of edge relations. The adjacency matrix A of a k-NN graph is defined as:
A i j = { 1 , x j N k ( x i ) 0 , o t h e r w i s e
where N k ( x i ) indicates the k-NNs of x i , which is determined by computing the pairwise Euclidean distance between the PCA reduced data points.

3. Methodology

SSAKGCSC

Based on the GCSC framework, statistical sub-graph affinity model, and Gaussian kernel function mentioned above, the statistical sub-graph affinity kernel GCSC (SSAKGCSC) is proposed. In this section, we introduce the establishment and the derivation process of the SSAKGCSC model.
We set both q and p to be the Frobenius norm in the GCSC model. Formally, we denote it as:
arg min Z 1 2 | | X A ¯ Z X | | F 2 + λ 2 | | Z | | F 2 .
Ji et al. [33] has proven that the Frobenius norm will not obtain trivial results even without the constraint d i a g ( Z ) = 0 . Solving the above equation leads to a dense matrix of self-representing coefficients and a valid closed-form solution. We can rewrite (7) as:
L ( Z ) = 1 2 | | X A ¯ Z X | | F 2 + λ 2 Z F 2 = 1 2 t r [ ( X A ¯ Z X ) T ( X A ¯ Z X ) + λ Z T Z ] = 1 2 t r ( Z T A ¯ T X T X A ¯ Z + X T X 2 X T X A ¯ Z + λ Z T Z )
According to matrix trace and matrix derivative, the partial derivative of L for Z can be expressed as:
L Z = A ¯ T X T X A ¯ Z A ¯ T X T X + λ Z = ( A ¯ T X T X A ¯ + λ I N ) Z A ¯ T X T X .
Let L Z = 0 ; we get:
( A ¯ T X T X A ¯ + λ I N ) Z = A ¯ T X T X .
Finally, matrix Z can be expressed as:
Z = ( A ¯ T X T X A ¯ + λ I N ) 1 A ¯ T X T X .
Due to A ¯ T X T X A ¯ + λ I N being positive semi-definite, its inverse matrix always exist.
Unlike traditional subspace clustering models that usually require iterative optimization, this algorithm ensures high computational efficiency. Once matrix Z is obtained, we can use it to build an affinity matrix for spectral clustering. Most works compute the affinity matrix by C = | Z | T + | Z | or | Z | [34]. In this work, we use the heuristic method employed by efficient dense subspace clustering (EDSC) [35] to enhance the block structure, which has been shown to benefit the accuracy of clustering.
Due to the complexity and nonlinearity of HSI, a large body of works has demonstrated that nonlinear models will achieve better performance than their linear counterparts [36]. Therefore, we use the kernel trick for its nonlinear extension.
Let ϕ : R m H be a mapping from the input space to the regeneration kernel Hilbert space H ; we define a positive semi-definite kernel Gram matrix K X X R N × N as:
[ K X X ] i j = [ ϕ ( X ) , ϕ ( X ) H ] = ϕ ( x i ) T ϕ ( x j )
where k : R m × R m R denotes the kernel function. The Gaussian kernel has been widely used in other subspace clustering models, i.e., k ( x i , x j ) = exp ( γ | | x i x j | | 2 ) , where γ is the parameter of the Gaussian kernel function [37]. Then the EKGCSC model can be expressed as:
arg min Z 1 2 | | ϕ ( X ) A ¯ Z ϕ ( X ) | | F 2 + λ 2 | | Z | | F 2 .
By using the kernel trick, (13) can be rewritten as:
arg min Z 1 2 t r ( Z T A ¯ T K X X A ¯ Z 2 K X X A ¯ Z + K X X + λ Z T Z ) .
The above equation can be solved by taking the partial derivative of Z and setting the result to zero, so (14) can be reformulated as:
L ( Z ) = 1 2 t r ( Z T A ¯ T K X X A ¯ Z 2 K X X A ¯ Z + K X X + λ Z T Z ) .
The partial derivative of L for Z can be expressed as:
L Z = A ¯ T K X X A ¯ Z A ¯ T K X X + λ Z = ( A ¯ T K X X A ¯ Z + λ I N ) Z A ¯ T K X X .
Let L Z = 0 ; we obtain the closed form solution as:
Z = ( A ¯ T K X X A ¯ + λ I N ) 1 A ¯ T K X X .
This model maps the data points onto a higher-dimensional space, thus making a linearly inseparable issue a detachable one.
The model improves the robustness of spectral clustering by replacing the node affinity K X X used in (17) with the concept of statistical sub-graph affinity.
W u v = exp { ( x u x v ) 2 σ 2 }
To create an affinity matrix, the connectivity of each node must be expressed by a connectivity graph G . G is implicitly constructed from graph kernels K . The kernel weights, elements of K , manifest the connectivity between a node and its neighbor. The method introduces a statistical sub-graph affinity model to substitute the node affinity model [38]. Instead of comparing x u to x v for v in the neighbor of u , the sub-graph P u is compared against each sub-graph P v , which is illustrated in Figure 1.
A mathematical description is used to understand how sub-graphs can improve classification accuracy. Assuming the node u belonging to a class C k , the node features, x u can be regarded as the random processes, X u , subject to independent identical distribution (i.i.d.) noise, N :
X u = μ u + N ,
where μ u is the mean of all node features belonging to class C k .
The sub-graph contains the connectivity of node u to its neighbors. When using node affinity, the probability that two nodes, u and v , belong to the same class, P ( u C k   and   v C k ) , denoted as P u , v , n o d e , is the probability that the difference between X u and X v is equal to zero:
P u , v , n o d e = P ( X u X v = 0 | C k ) .
Since X u and X v are i.i.d., then P u , v , n o d e can be confirmed using the Laplace distribution [39]:
P u , v , n o d e = 1 ( 2 b ) n exp ( | x u x v | b )
where b is a coefficient related to N . The sub-graph affinity restrains the noise compared to the node affinity. The obtained sub-graph affinity probability P s u b is equal to:
P u , v , s u b = i P ( X u , i X v , i = 0 | C k ) .
P u , v , s u b = i 1 ( 2 b ) n exp ( | x u , i x v , i | b )
where i is the index of each factor in P u and the corresponding element in P v . The more similar the distribution in each sub-graph, the smaller x u , i x v , i becomes, and the larger the probability of P ( u C k   and   v C k ) . So, the local neighborhood information can help suppress the noise.
Given the proposed statistical sub-graph affinity model, we incorporated this model into the spectral clustering framework by introducing a weighting matrix into (17) [40]. The density of W mainly depends on the complexity of G and is independent of P u . The value of W , which depends on the kernel weights and the statistical sub-graph affinity, can be defined as:
w u v = K u v i exp { ( ( x u + i x v + i ) P u , i ) 2 2 σ 2 }
For node u and node v , The weights related with K u and P u are also imported to allow spatial weighting.
The following shows how K u and P u influence the size of the neighborhood and the computational complexity of (23). Increasing the size of K increases the size of the neighborhood; therefore, the number of x u and x v increases, which leads to decreasing the sparsity of W . Increasing the size of P u does not affect the neighborhood size and, consequently, it does not affect the sparsity of W . The increased size of P u only increases the computational complexity of calculating w u v moderately. The difference of x u and x v can be saved after first calculation. Hence, the size of K makes a more noteworthy effect on the overall computational complexity than the size of P u .
By choosing a radially symmetric kernel, the algorithm can be rotationally invariant. The kernel can also empower distance information so that the weights or connectivity of the sub-graph can be a function of spatial distance, i.e., the intensity difference of the relevant pixel can be weighted according to the distance from the center of the sub-graph [41]. By integrating the spatial intensity relationships within the sub-graph into the construction of the affinity weighted graph, the constructed graph is less sensitive to random noise. Consequently, the statistical sub-graph affinity model is stronger against noise.
We use (24) as a newly introduced ‘Gaussian kernel’ to implement the combination of the sub-graph affinity model with the GCSC model, which we denote as K0; the result of multiplying (24) with the Gaussian kernel is denoted as K0 × K1.

4. Experiments Results

4.1. Setup

4.1.1. Data Sets and Preprocessing

We implemented experiments on four HSI data sets: Salinas, Indian Pines corrected, Pavia Center, and Pavia University. For computational efficiency, we take a sub-scene of each of these data sets for evaluation [35,36], Specifically, these sub-scenes are located at [591–676, 158–240], [30–115, 24–94], [100–250, 200–350], and [150–350, 100–200], respectively. The sub-scenes obtained from the Salinas data set is referred to as the SalinasA data set. The detailed information of these four data sets were demonstrated in Table 1. In the data preprocessing phase, we reduced the spectral bands into four by retaining at least 96% of the cumulative percent variance (CPV) by PCA [42]. We constructed the spectral-spatial blocks by setting the neighborhood size to be nine for data sets. All data points were normalized by scaling into [0, 1] before clustering.
Since the HSI image with considerable noise cannot be obtained, we decided to add noise into these images manually. In the process of adding noise to HSI images, we noticed that, due to the unique characteristics of HSI images, the maximum value in each channel was no longer 255 as in RGB images, but varied from dozens to 8000. If all dimensions of a hyperspectral image are normalized uniformly, it is easy to produce smaller numbers and affect the algorithm’s stability. For these reasons, each dimension of the HSI image was normalized separately, and then Gaussian noise and salt and pepper noise was added to it.

4.1.2. Evaluation Metrics

Three noted metrics [12,25,35], are used to estimate the clustering performance of the clustering model, i.e., overall accuracy (OA), normalized mutual information (NMI), and Kappa coefficient (Kappa). These metrics range from 0 to 1, where the larger the value are, the more exact the clustering results are obtained. To assess the computational complexity of the models, we compared the running time of different models in our experiments.

4.1.3. Compared Methods

We compared the SSAKGCSC method with some HSI clustering methods, including traditional clustering methods and superior methods. To be specific, the traditional clustering methods include sparse subspace clustering (SSC) [6], efficient dense subspace clustering (EDSC) [24], and low-rank subspace clustering (LRSC) [27]. The relatively superior HSI clustering methods are spectral-spatial subspace clustering (S4C) [15], efficient graph convolutional subspace clustering (EGCSC) [31], and efficient kernel graph convolutional subspace clustering (EKGCSC) [31].
We follow their settings published in the corresponding paper for the HSI clustering methods mentioned above. The hyperparameters of SSAKGCSC are given in Table 2. All the compared methods are implemented with Python 3.7 running on an Intel(R) Core (TM) i5-5200U CPU @ 2.20GHz CPU with 8GB RAM.

4.2. Test Results

To observe the clustering results, below we illustrate the clustering maps and partial enlarged detail of the different clustering methods mentioned above. It should be noted that the color of the same class may vary in different class maps. This is because label characters can be arranged by different clustering methods.

4.2.1. Images without Noise

The performance of different clustering methods on the SalinasA, Indian Pines, Pavia Center, and Pavia University data sets are shown in Figure 2, Figure 3, Figure 4 and Figure 5. From those figures, it can be seen that the SSAKGCSC algorithm achieved extremely high classification accuracy on the four data sets. Compared to other models, SSAKGCSC shows better class maps in all three HSI data sets. In comparison, the class maps obtained by the other methods (e.g., SSC, LRSC, and EDSC) contain more noisy points caused by misclassification. The results indicate the effectiveness and advantages of the GCSC framework.
Table 3 shows the clustering performance comparison of different methods evaluated on SalinasA, Indian Pines, Pavia Center, and Pavia University data sets. From the results, it can be seen that the proposed methods achieve pretty adequate segmentation results on the HSI data sets without additional noise. It even outperforms the EKGCSC model in some aspects of OA, NMI, and Kappa coefficient. We can find a conspicuous progress in the performance of the traditional subspace clustering model with graph convolution from the SSAKGCSC significantly outperforming the EDSC results. Secondly, the SSAKGCSC algorithm that analyzes the nonlinear relationship between data points outperforms the EGCSC algorithm on all three data sets. That is, the GCSC algorithm with a nonlinear model can significantly improve the clustering performance by learning a more robust affinity matrix.

4.2.2. Images with Salt and Pepper Noise

Figure 6, Figure 7, Figure 8 and Figure 9 show the capability comparison of different clustering methods on the SalinasA, Indian Pines, Pavia Center, and Pavia University data sets with the addition of 0.1 probability of salt and pepper noise. As can be seen from Figure 5, the SSAKGCSC algorithm used in this work had a better performance compared to the EKGCSC algorithm in resisting salt and pepper noise. Since the EKGCSC algorithm cannot run on the noise-added PaviaU data set, the SSAKGCSC algorithm has better segmentation results compared to the rest of the comparison algorithms on the other two data sets.
Table 4 shows the performance comparison of different methods evaluated on SalinasA, Indian Pines, Pavia Center, and Pavia University data sets with the addition of a 0.1 probability of salt and pepper noise. From the results, it can be seen that the SSAKGCSC method used in this work achieves the best results in terms of both OA, NMI, and Kappa coefficients. In particular, the evaluation metrics on the SalinasA data set are almost double those of the EKGCSC model; on the other two data sets, our algorithm also shows a considerable improvement in three evaluation metrics. When the author runs the EKGCSC model for the PaviaU data set with noise, no matter whether it is pepper and salt noise or Gaussian noise, the program will be input the code segment of spectral clustering, so that its clustering result cannot be obtained no matter how long it runs.

4.2.3. Images with Gaussian Noise

Figure 10, Figure 11, Figure 12 and Figure 13 show the clustering results of different clustering methods on the SalinasA, Indian Pines, Pavia Center, and Pavia University data sets under Gaussian noise with a mean of 0 and variance of 0.2. It can be seen from the results that the SSAKGCSC algorithm has an excellent performance in resisting Gaussian noise. This is especially evident in Figure 12, where the segmentation results of all the compared algorithms have a large area with noisy disturb, except for the result of the SSAKGCSC algorithm.
Table 5 shows the performance comparison of different clustering methods on the SalinasA, Indian Pines, Pavia Center, and Pavia University data sets under Gaussian noise with a mean of 0 and variance of 0.2. From the results, it can be seen that the SSAKGCSC method improves the OA, NMI, and Kappa coefficients by 26%, 24%, and 36%, respectively, compared to the standard method in the SalinasA data set. The method also improves the OA, NMI, and Kappa coefficients by 16%, 15%, and 16%, respectively, compared to the standard method in the Pavia University data set.

4.2.4. Images with Gaussian Noise and Salt and Pepper Noise

Figure 14, Figure 15, Figure 16 and Figure 17 show the performance comparison of different clustering methods on the SalinasA, Indian Pines, Pavia Center, and Pavia University data sets under both Gaussian noise with a mean of 0 and variance of 0.1 and salt and pepper noise with a probability of 0.1. It can be seen that the SSAKGCSC method still has the best classification results and remains to be the strongest among all algorithms for noise suppression. From the above, it can be concluded that the combined application of the GCSC framework, statistical sub-graph affinity model, and kernel function used in this research has adequate effectiveness and superiority in noise resistance of HSI segmentation.
Table 6 shows the performance comparison of different clustering methods on the SalinasA, Indian Pines, Pavia Center, and Pavia University data sets under Gaussian noise with a mean of 0 and variance of 0.1 and salt and pepper noise with a probability of 0.1. As can be seen from the table, the SSAKGCSC method improves the OA, NMI, and Kappa coefficients by 28%, 24%, and 36%, respectively, compared to the standard method in the SalinasA data set. The method also improves the OA, NMI, and Kappa coefficients by 19%, 16%, and 20%, respectively, compared to the standard method in the Pavia University data set.

4.3. Impact of the Number of PCs

Dimension reduction usually keeps the balance between computational efficiency and model performance. When more channels are taken into account, more specific information is included. However, possessing more channels does not always result in better clustering performance. To scientifically discuss the influence of the number of PCs, we implement the proposed methods with varying PCs from one to eight, then display the obtained OA and CPV in Figure 18. As shown in Figure 18, the clustering OA increased in the first four steps and then decreased. This appearance is known as the Hughes phenomenon [43], which is primarily due to increased correlation information and a lack of enough data points. As seen in Figure 18, the first four PCs of the data set can provide more than 96% of the CPV. This means that the remaining channels only possess restricted information and increase the risk of causing redundancy. Additionally, to identify subspaces correctly, the model must use more data points to model the statistical property of the data. To balance model accuracy with operational efficiency, we chose to use the first four PCs for the model.

4.4. Comparison of Running Times

Table 7 shows the running times of our methods compared to the other competing methods. Since EGCSC, EKGCSC, and SSAKGCSC will achieve a closed-form solution without needing iterations, they are significantly quicker than LRSC, SSC, and S4C. Our method requires relatively more runtime compared to EDSC, which is because the GCSC framework requires constructing the graph from the original data points. Furthermore, SSAKGCSC needs to compute the sub-graph affinity matrix, so its running time will be increased compared to EGCSC and EKGCSC. In summary, our proposed SSAKGCSC model achieves a adequate balance between algorithm efficiency, clustering accuracy, and algorithm noise immunity.

5. Discussion

In this paper, we introduced the HSI clustering framework based on graph convolution into subspace clustering (GCSC). The critical point of this framework is to use a graph convolution self-representation to integrate the intrinsic structural information of data points, followed by an explicit dictionary to learn robust affinity matrices. We also introduced statistical sub-graph affinity, a robust spectral clustering strategy to master the heavy noise, which constructs the affinity weighted graph based on the spatial intensity relationships between connected nodes. The statistical sub-graph affinity model includes a statistical noise model when treating the sub-graphs as random processes, thus better handling the uncertainty ingredient of noise. The statistical sub-graph affinity model contains neighborhood connectivity when representing the sub-graphs as random processes, thus making use of neighborhood information to counteract uncertainties and properties of noise. The experimental results on three HSI data sets show that the SSAKGCSC model can achieve the best performance compared to many existing clustering models. Moreover, the algorithm also has excellent performance over other algorithms in terms of noise immunity; in particular, the segmentation result on the SalinasA data set under salt and pepper noise is double than standard model EKGCSC, reaching 83.47%, 85.27%, and 79.05% of the clustering OA, NMI, Kappa coefficient, respectively.

6. Conclusions

The success of the SSAKGCSC model manifests that utilizing the intrinsic graph structure among data sets is crucial for clustering. The GCSC framework used in this study encourages us to implement traditional clustering models in non-Euclidean domains; this means that any improved tactics used in the classical subspace clustering model are suitable for GCSC. However, the GCSC framework suffers from quadratic complexity in the size of data points, which makes it difficult to be used for large-scale HSI data. In addition to the independent identically distributed noise model, other statistical models can be developed similarly. This paper illustrates statistical sub-graph affinity in image segmentation, which has an adequate ability to resist the additive noise.
First, future research will focus on how to effectively process the hyperspectral image with speckle noise, which is a multiplicative noise. Second, future work will focus on how to create the similarity matrix W to more truly reflect the approximate relationship between data points. The Gaussian function used in this study has obvious limitations due to the selection of the scale parameter σ. A third future research aspect will explore how to automatically determine the number of clusters; although the number can be determined manually, it is a difficult problem to determine the exact number that has a direct effect on the efficiency and the final quality of clustering. Fourth, future work will look at how to select eigenvectors for the construction of new vector spaces, in most cases, the spectral clustering algorithm directly selects the first k maximum eigenvalues corresponding to the eigenvectors for the construction of new vector space. A final future research focus will investigate how to improve the running speed of spectral clustering and solve matrix eigenvalues and characteristic vectors. In dealing with large-scale data sets, the calculation of matrix space is very large; a solving process not only can be very time consuming but the required memory space is very large. Therefore, improving the running speed of the algorithm, reducing the memory space required to run, and reducing the time and space cost of the algorithm is another key problem in the process of expanding the application field of the algorithm.

Author Contributions

Conceptualization, C.W.; methodology, C.W.; software, J.W.; validation, J.W.; formal analysis, J.W.; investigation, J.W.; resources, J.W.; data curation, J.W.; writing—original draft preparation, J.W.; writing—review and editing, C.W.; visualization, J.W.; supervision, Z.L. and C.W.; project administration, C.W.; funding acquisition, Z.L. All authors have read and agreed to the published version of the manuscript.

Funding

The research was supported in part by the National Natural Science Foundation of China under Grant 12102341, in part by the National Natural Science Foundation of China under Grant 62071378, in part by the Natural Science Basic Research Program of Shaanxi in 2020 under Program 2020JM-580, in part by the Natural Science Basic Research Program of Shaanxi under Program 2022JQ-020, and in part by the Scientific Research Program Funded by Shaanxi Provincial Education Department under Program 21JK0904.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data sets used in this article can be downloaded at: https://www.ehu.eus/ccwintco/index.php/Hyperspectral_Remote_Sensing_Scenes, SalinasA, Indian Pines, Pavia Center and Pavia University data sets (accessed on 20 May 2022).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ahmad, M.; Shabbir, S.; Roy, S.K.; Hong, D.; Wu, X.; Yao, J.; Khan, A.M.; Mazzara, M.; Distefano, S.; Chanussot, J. Hyperspectral Image Classification—Traditional to Deep Models: A Survey for Future Prospects. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 968–999. [Google Scholar] [CrossRef]
  2. Cai, Y.; Dong, Z.; Cai, Z.; Liu, X.; Wang, G. Discriminative spectral-spatial attention-aware residual network for hyperspectral image classification. In Proceedings of the 10th Workshop on Hyperspectral Imaging and Signal Processing: Evolution in Remote Sensing (WHISPERS), Amsterdam, The Netherlands, 24–26 September 2019; pp. 1–5. [Google Scholar] [CrossRef]
  3. Xu, Z.; Yu, H.; Zheng, K.; Gao, L.; Song, M. A Novel Classification Framework for Hyperspectral Image Classification Based on Multiscale Spectral-Spatial Convolutional Network. In Proceedings of the 11th Workshop on Hyperspectral Imaging and Signal Processing: Evolution in Remote Sensing (WHISPERS), Amsterdam, The Netherlands, 24–26 March 2021; pp. 1–5. [Google Scholar] [CrossRef]
  4. Qin, A.; Shang, Z.; Tian, J.; Wang, Y.; Zhang, T.; Tang, Y.Y. Spectral–Spatial Graph Convolutional Networks for Semisupervised Hyperspectral Image Classification. IEEE Geosci. Remote Sens. Lett. 2018, 16, 241–245. [Google Scholar] [CrossRef]
  5. Li, S.; Song, W.; Fang, L.; Chen, Y.; Ghamisi, P.; Benediktsson, J.A. Deep Learning for Hyperspectral Image Classification: An Overview. IEEE Trans. Geosci. Remote Sens. 2019, 57, 6690–6709. [Google Scholar] [CrossRef] [Green Version]
  6. He, L.; Li, J.; Liu, C.; Li, S. Recent Advances on Spectral–Spatial Hyperspectral Image Classification: An Overview and New Guidelines. IEEE Trans. Geosci. Remote Sens. 2017, 56, 1579–1597. [Google Scholar] [CrossRef]
  7. Chairet, R.; Ben Salem, Y.; Aoun, M. Land cover classification of GeoEye image based on Convolutional Neural Networks. In Proceedings of the 2020 17th International Multi-Conference on Systems, Signals & Devices (SSD), Monastir, Tunisia, 20–23 July 2020; pp. 458–461. [Google Scholar] [CrossRef]
  8. Hu, P.; Liu, X.; Cai, Y.; Cai, Z. Band Selection of Hyperspectral Images Using Multiobjective Optimization-Based Sparse Self-Representation. IEEE Geosci. Remote Sens. Lett. 2018, 16, 452–456. [Google Scholar] [CrossRef]
  9. Chang, J.; Meng, G.; Wang, L.; Xiang, S.; Pan, C. Deep Self-Evolution Clustering. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 42, 809–823. [Google Scholar] [CrossRef]
  10. Zhu, J.; Jiang, Z.; Evangelidis, G.D.; Zhang, C.; Pang, S.; Li, Z. Efficient registration of multi-view point sets by K-means clustering. Inf. Sci. 2019, 488, 205–218. [Google Scholar] [CrossRef] [Green Version]
  11. Zeng, M.; Cai, Y.; Cai, Z.; Liu, X.; Hu, P.; Ku, J. Unsupervised Hyperspectral Image Band Selection Based on Deep Subspace Clustering. IEEE Geosci. Remote Sens. Lett. 2019, 16, 1889–1893. [Google Scholar] [CrossRef]
  12. Zhang, L.; Zhang, L.; Du, B.; You, J.; Tao, D. Hyperspectral image unsupervised classification by robust manifold matrix factorization. Inf. Sci. 2019, 485, 154–169. [Google Scholar] [CrossRef]
  13. Yang, Y.; Wang, T. Kernel Subspace Clustering with Block Diagonal Prior. In Proceedings of the 2020 2nd International Conference on Machine Learning, Big Data and Business Intelligence (MLBDBI), Taiyuan, China, 23–25 October 2020; pp. 367–370. [Google Scholar] [CrossRef]
  14. Zhai, H.; Zhang, H.; Zhang, L.; Li, P.; Plaza, A. A New Sparse Subspace Clustering Algorithm for Hyperspectral Remote Sensing Imagery. IEEE Geosci. Remote Sens. Lett. 2017, 14, 43–47. [Google Scholar] [CrossRef]
  15. Wei, Y.; Niu, C.; Wang, Y.; Wang, H.; Liu, D. The Fast Spectral Clustering Based on Spatial Information for Large Scale Hyperspectral Image. IEEE Access 2019, 7, 141045–141054. [Google Scholar] [CrossRef]
  16. Tang, C.; Zhu, X.; Liu, X.; Li, M.; Wang, P.; Zhang, C.; Wang, L. Learning a Joint Affinity Graph for Multiview Subspace Clustering. IEEE Trans. Multimedia 2019, 21, 1724–1736. [Google Scholar] [CrossRef]
  17. Hinojosa, C.; Vera, E.; Arguello, H. A Fast and Accurate Similarity-Constrained Subspace Clustering Algorithm for Hyperspectral Image. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 10773–10783. [Google Scholar] [CrossRef]
  18. Elhamifar, E.; Vidal, R. Sparse Subspace Clustering: Algorithm, Theory, and Applications. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 2765–2781. [Google Scholar] [CrossRef] [Green Version]
  19. Vidal, R.; Favaro, P. Low rank subspace clustering (LRSC). Pattern Recognit. Lett. 2013, 43, 47–61. [Google Scholar] [CrossRef] [Green Version]
  20. Patel, V.M.; Vidal, R. Kernel sparse subspace clustering. In Proceedings of the IEEE International Conference on Image Processing (ICIP), Paris, France, 27–30 October 2014; pp. 2849–2853. [Google Scholar] [CrossRef]
  21. Zhai, H.; Zhang, H.; Zhang, L.; Li, P. Laplacian-Regularized Low-Rank Subspace Clustering for Hyperspectral Image Band Selection. IEEE Trans. Geosci. Remote Sens. 2019, 57, 1723–1740. [Google Scholar] [CrossRef]
  22. Ji, P.; Salzmann, M.; Li, H. Efficient dense subspace clustering. In Proceedings of the IEEE Winter Conference on Applications of Computer Vision, Steamboat Springs, CO, USA, 24–26 March 2014; pp. 461–468. [Google Scholar] [CrossRef]
  23. Yang, X.; Xu, W.-D.; Liu, H.; Zhu, L.-Y. Research on Dimensionality Reduction of Hyperspectral Image under Close Range. In Proceedings of the 2019 International Conference on Communications, Information System and Computer Engineering (CISCE), Haikou, China, 5–7 July 2019; pp. 171–174. [Google Scholar] [CrossRef]
  24. Tian, L.; Du, Q.; Kopriva, I. L0-Motivated Low Rank Sparse Subspace Clustering for Hyperspectral Imagery. In Proceedings of the IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Waikoloa, HI, USA, 26 September–2 October 2020; pp. 1038–1041. [Google Scholar] [CrossRef]
  25. Lu, C.; Feng, J.; Lin, Z.; Mei, T.; Yan, S. Subspace Clustering by Block Diagonal Representation. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 41, 487–501. [Google Scholar] [CrossRef] [Green Version]
  26. Kipf, T.; Welling, M. Semi-supervised classification with graph convolutional networks. In Proceedings of the International Conference on Learning Representations, 9 September 2016; pp. 1–14. [Google Scholar] [CrossRef]
  27. Xu, H.; Yao, S.; Li, Q.; Ye, Z. An Improved K-means Clustering Algorithm. In Proceedings of the 2020 IEEE 5th International Symposium on Smart and Wireless Systems within the Conferences on Intelligent Data Acquisition and Advanced Computing Systems (IDAACS-SWS), Dortmund, Germany, 17–18 September 2020; pp. 1–5. [Google Scholar] [CrossRef]
  28. Zhu, Q.; Wang, Y.; Wang, F.; Song, M.; Chang, C.-I. Hyperspectral Band Selection Based on Improved Affinity Propagation. In Proceedings of the 11th Workshop on Hyperspectral Imaging and Signal Processing: Evolution in Remote Sensing (WHISPERS), Amsterdam, The Netherlands, 24–26 March 2021; pp. 1–4. [Google Scholar] [CrossRef]
  29. Pardo, A. Analysis of non-local image denoising methods. Pattern Recognit. Lett. 2011, 32, 2145–2149. [Google Scholar] [CrossRef]
  30. Fang, X.; Cai, Y.; Cai, Z.; Jiang, X.; Chen, Z. Sparse Feature Learning of Hyperspectral Imagery via Multiobjective-Based Extreme Learning Machine. Sensors 2020, 20, 1262. [Google Scholar] [CrossRef] [Green Version]
  31. Luo, B.; Wilson, R.C.; Hancock, E. Spectral embedding of graphs. Pattern Recognit. 2003, 36, 2213–2230. [Google Scholar] [CrossRef] [Green Version]
  32. Lei, T.; Jia, X.; Zhang, Y.; Liu, S.; Meng, H.; Nandi, A.K. Superpixel-Based Fast Fuzzy C-Means Clustering for Color Image Segmentation. IEEE Trans. Fuzzy Syst. 2019, 27, 1753–1766. [Google Scholar] [CrossRef] [Green Version]
  33. Yang, Y.; Gu, H.; Han, Y.; Li, H. An End-to-End Deep Learning Change Detection Framework for Remote Sensing Images. In Proceedings of the IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Waikoloa, HI, USA, 26 September–2 October 2020; pp. 652–655. [Google Scholar] [CrossRef]
  34. Chang, J.; Oh, M.-H.; Park, B.-G. A systematic model parameter extraction using differential evolution searching. In Proceedings of the 2019 Silicon Nanoelectronics Workshop (SNW), Kyoto, Japan, 9–10 June 2019; pp. 1–2. [Google Scholar] [CrossRef]
  35. Kong, Y.; Cheng, Y.; Chen, C.L.P.; Wang, X. Hyperspectral Image Clustering Based on Unsupervised Broad Learning. IEEE Geosci. Remote Sens. Lett. 2019, 16, 1741–1745. [Google Scholar] [CrossRef]
  36. Eichel, J.A.; Wong, A.; Fieguth, P.; Clausi, D.A. Robust Spectral Clustering Using Statistical Sub-Graph Affinity Model. PLoS ONE 2013, 8, e82722. [Google Scholar] [CrossRef] [Green Version]
  37. JayaSree, M.; Rao, L.K. A Deep Insight into Deep Learning Architectures, Algorithms and Applications. In Proceedings of the 2022 International Conference on Electronics and Renewable Systems (ICEARS), Tuticorin, India, 16–18 March 2022; pp. 1134–1142. [Google Scholar] [CrossRef]
  38. Pan, Y.; Jiao, Y.; Li, T.; Gu, Y. An efficient algorithm for hyperspectral image clustering. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Brighton, UK, 12–17 May 2019; pp. 2167–2171. [Google Scholar] [CrossRef]
  39. Cai, Y.; Zhang, Z.; Cai, Z.; Liu, X.; Jiang, X.; Yan, Q. Graph Convolutional Subspace Clustering: A Robust Subspace Clustering Framework for Hyperspectral Image. IEEE Trans. Geosci. Remote Sens. 2021, 59, 4191–4202. [Google Scholar] [CrossRef]
  40. Zeng, M.; Cai, Y.; Liu, X.; Cai, Z.; Li, X. Spectral–spatial clustering of hyperspectral image based on Laplacian regularized deep subspace clustering. In Proceedings of the IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Yokohama, Japan, 28 July–2 August 2019; pp. 2694–2697. [Google Scholar] [CrossRef]
  41. Zhu, X.; Zhang, S.; Li, Y.; Zhang, J.; Yang, L.; Fang, Y. Low-Rank Sparse Subspace for Spectral Clustering. IEEE Trans. Knowl. Data Eng. 2019, 31, 1532–1543. [Google Scholar] [CrossRef]
  42. Wan, Y.; Zhong, Y.; Ma, A.; Zhang, L. Multi-Objective Sparse Subspace Clustering for Hyperspectral Imagery. IEEE Trans. Geosci. Remote Sens. 2020, 58, 2290–2307. [Google Scholar] [CrossRef]
  43. Gong, W.; Wang, Y.; Cai, Z.; Wang, L. Finding Multiple Roots of Nonlinear Equation Systems via a Repulsion-Based Adaptive Differential Evolution. IEEE Trans. Syst. Man, Cybern. Syst. 2020, 50, 1499–1513. [Google Scholar] [CrossRef]
Figure 1. Comparison of node affinity to sub-graph affinity. x u (red), x v (green), P u (light red), P v (light green). (a) Node affinity. (b) Node affinity with neighborhood statistics. (c) Sub-graph affinity.
Figure 1. Comparison of node affinity to sub-graph affinity. x u (red), x v (green), P u (light red), P v (light green). (a) Node affinity. (b) Node affinity with neighborhood statistics. (c) Sub-graph affinity.
Axioms 11 00269 g001
Figure 2. Clustering results obtained using different methods on the SalinasA data set. (a) Ground truth. (b) LRSC 50.09%. (c) SSC 75.90%. (d) S4C 80.70%. (e) EDSC 88.99%. (f) EGCSC 99.93%. (g) EKGCSC 100.00%. (h) SSAKGCSC 99.89%.
Figure 2. Clustering results obtained using different methods on the SalinasA data set. (a) Ground truth. (b) LRSC 50.09%. (c) SSC 75.90%. (d) S4C 80.70%. (e) EDSC 88.99%. (f) EGCSC 99.93%. (g) EKGCSC 100.00%. (h) SSAKGCSC 99.89%.
Axioms 11 00269 g002
Figure 3. Clustering results obtained using different methods on the Indian Pines corrected data set. (a) Ground truth. (b) LRSC 57.52%. (c) SSC 56.09%. (d) S4C 64.97%. (e) EDSC 70.26%. (f) EGCSC 88.27%. (g) EKGCSC 97.31%. (h) SSAKGCSC 97.43%.
Figure 3. Clustering results obtained using different methods on the Indian Pines corrected data set. (a) Ground truth. (b) LRSC 57.52%. (c) SSC 56.09%. (d) S4C 64.97%. (e) EDSC 70.26%. (f) EGCSC 88.27%. (g) EKGCSC 97.31%. (h) SSAKGCSC 97.43%.
Axioms 11 00269 g003
Figure 4. Clustering results obtained using different methods on the Pavia University data set. (a) Ground truth. (b) LRSC 46.16%. (c) SSC 64.27%. (d) S4C 66.72%. (e) EDSC 65.94%. (f) EGCSC 84.42%. (g) EKGCSC 97.32%. (h) SSAKGCSC 97.08%.
Figure 4. Clustering results obtained using different methods on the Pavia University data set. (a) Ground truth. (b) LRSC 46.16%. (c) SSC 64.27%. (d) S4C 66.72%. (e) EDSC 65.94%. (f) EGCSC 84.42%. (g) EKGCSC 97.32%. (h) SSAKGCSC 97.08%.
Axioms 11 00269 g004
Figure 5. Clustering results obtained using different methods on the Pavia Center data set. (a) Ground truth. (b) LRSC 53.06%. (c) SSC 60.26%. (d) S4C 72.80%. (e) EDSC 75.39%. (f) EGCSC 80.30%. (g) EKGCSC 94.48%. (h) MEKGCSC 93.89%.
Figure 5. Clustering results obtained using different methods on the Pavia Center data set. (a) Ground truth. (b) LRSC 53.06%. (c) SSC 60.26%. (d) S4C 72.80%. (e) EDSC 75.39%. (f) EGCSC 80.30%. (g) EKGCSC 94.48%. (h) MEKGCSC 93.89%.
Axioms 11 00269 g005
Figure 6. Clustering results obtained using different methods on the SalinasA data set with salt and pepper noise. (a) Ground truth. (b) LRSC 43.49%. (c) SSC 48.60%. (d) S4C 51.08%. (e) EDSC 61.03%. (f) EGCSC 68.03%. (g) EKGCSC 45.74%. (h) SSAKGCSC 83.47%.
Figure 6. Clustering results obtained using different methods on the SalinasA data set with salt and pepper noise. (a) Ground truth. (b) LRSC 43.49%. (c) SSC 48.60%. (d) S4C 51.08%. (e) EDSC 61.03%. (f) EGCSC 68.03%. (g) EKGCSC 45.74%. (h) SSAKGCSC 83.47%.
Axioms 11 00269 g006
Figure 7. Clustering results obtained using different methods on the Indian Pines corrected data set with salt and pepper noise. (a) Ground truth. (b) LRSC 40.29%. (c) SSC 44.75%. (d) S4C 50.63%. (e) EDSC 52.88%. (f) EGCSC 57.07%. (g) EKGCSC 56.09%. (h) SSAKGCSC 66.34%.
Figure 7. Clustering results obtained using different methods on the Indian Pines corrected data set with salt and pepper noise. (a) Ground truth. (b) LRSC 40.29%. (c) SSC 44.75%. (d) S4C 50.63%. (e) EDSC 52.88%. (f) EGCSC 57.07%. (g) EKGCSC 56.09%. (h) SSAKGCSC 66.34%.
Axioms 11 00269 g007
Figure 8. Clustering results obtained using different methods on the Pavia University data set with salt and pepper noise. (a) Ground truth. (b) LRSC 37.69%. (c) SSC 51.22%. (d) S4C 53.31%. (e) EDSC 53.56%. (f) EGCSC 60.12%. (g) EKGCSC. (h) SSAKGCSC 79.83%.
Figure 8. Clustering results obtained using different methods on the Pavia University data set with salt and pepper noise. (a) Ground truth. (b) LRSC 37.69%. (c) SSC 51.22%. (d) S4C 53.31%. (e) EDSC 53.56%. (f) EGCSC 60.12%. (g) EKGCSC. (h) SSAKGCSC 79.83%.
Axioms 11 00269 g008
Figure 9. Clustering results obtained using different methods on the Pavia Center data set with salt and pepper noise. (a) Ground truth. (b) LRSC 38.21%. (c) SSC 50.39%. (d) S4C 52.38%. (e) EDSC 60.33%. (f) EGCSC 64.22%. (g) EKGCSC 33.68%. (h) MEKGCSC 85.01%.
Figure 9. Clustering results obtained using different methods on the Pavia Center data set with salt and pepper noise. (a) Ground truth. (b) LRSC 38.21%. (c) SSC 50.39%. (d) S4C 52.38%. (e) EDSC 60.33%. (f) EGCSC 64.22%. (g) EKGCSC 33.68%. (h) MEKGCSC 85.01%.
Axioms 11 00269 g009
Figure 10. Clustering results obtained using different methods on the SalinasA data set with Gaussian noise. (a) Ground truth. (b) LRSC 35.96%. (c) SSC 47.61%. (d) S4C 54.77%. (e) EDSC 61.97%. (f) EGCSC 54.92%. (g) EKGCSC 60.73%. (h) SSAKGCSC 86.78%.
Figure 10. Clustering results obtained using different methods on the SalinasA data set with Gaussian noise. (a) Ground truth. (b) LRSC 35.96%. (c) SSC 47.61%. (d) S4C 54.77%. (e) EDSC 61.97%. (f) EGCSC 54.92%. (g) EKGCSC 60.73%. (h) SSAKGCSC 86.78%.
Axioms 11 00269 g010
Figure 11. Clustering results obtained using different methods on the Indian Pines corrected data set with Gaussian noise. (a) Ground truth. (b) LRSC 43.20%. (c) SSC 42.20%. (d) S4C 53.29%. (e) EDSC 56.78%. (f) EGCSC 59.30%. (g) EKGCSC 54.70%. (h) SSAKGCSC 64.95%.
Figure 11. Clustering results obtained using different methods on the Indian Pines corrected data set with Gaussian noise. (a) Ground truth. (b) LRSC 43.20%. (c) SSC 42.20%. (d) S4C 53.29%. (e) EDSC 56.78%. (f) EGCSC 59.30%. (g) EKGCSC 54.70%. (h) SSAKGCSC 64.95%.
Axioms 11 00269 g011
Figure 12. Clustering results obtained using different methods on the Pavia University data set with Gaussian noise. (a) Ground truth. (b) LRSC 36.90%. (c) SSC 36.79%. (d) S4C 42.51%. (e) EDSC 46.97%. (f) EGCSC 51.19%. (g) EKGCSC. (h) SSAKGCSC 67.34%.
Figure 12. Clustering results obtained using different methods on the Pavia University data set with Gaussian noise. (a) Ground truth. (b) LRSC 36.90%. (c) SSC 36.79%. (d) S4C 42.51%. (e) EDSC 46.97%. (f) EGCSC 51.19%. (g) EKGCSC. (h) SSAKGCSC 67.34%.
Axioms 11 00269 g012
Figure 13. Clustering results obtained using different methods on the Pavia Center data set with Gaussian noise. (a) Ground truth. (b) LRSC 31.63%. (c) SSC 37.81%. (d) S4C 49.51%. (e) EDSC 47.57%. (f) EGCSC 48.44%. (g) EKGCSC 33.15%. (h) MEKGCSC 88.55%.
Figure 13. Clustering results obtained using different methods on the Pavia Center data set with Gaussian noise. (a) Ground truth. (b) LRSC 31.63%. (c) SSC 37.81%. (d) S4C 49.51%. (e) EDSC 47.57%. (f) EGCSC 48.44%. (g) EKGCSC 33.15%. (h) MEKGCSC 88.55%.
Axioms 11 00269 g013
Figure 14. Clustering results obtained using different methods on the SalinasA data set with Gaussian noise and salt and pepper noise. (a) Ground truth. (b) LRSC 44.52%. (c) SSC 47.06%. (d) S4C 50.95%. (e) EDSC 60.79%. (f) EGCSC 62.23%. (g) EKGCSC 59.97%. (h) SSAKGCSC 88.28%.
Figure 14. Clustering results obtained using different methods on the SalinasA data set with Gaussian noise and salt and pepper noise. (a) Ground truth. (b) LRSC 44.52%. (c) SSC 47.06%. (d) S4C 50.95%. (e) EDSC 60.79%. (f) EGCSC 62.23%. (g) EKGCSC 59.97%. (h) SSAKGCSC 88.28%.
Axioms 11 00269 g014
Figure 15. Clustering results obtained using different methods on the Indian Pines corrected data set with Gaussian noise and salt and pepper noise. (a) Ground truth. (b) LRSC 41.08%. (c) SSC 43.86%. (d) S4C 51.33%. (e) EDSC 52.24%. (f) EGCSC 54.68%. (g) EKGCSC 52.49%. (h) SSAKGCSC 64.50%.
Figure 15. Clustering results obtained using different methods on the Indian Pines corrected data set with Gaussian noise and salt and pepper noise. (a) Ground truth. (b) LRSC 41.08%. (c) SSC 43.86%. (d) S4C 51.33%. (e) EDSC 52.24%. (f) EGCSC 54.68%. (g) EKGCSC 52.49%. (h) SSAKGCSC 64.50%.
Axioms 11 00269 g015
Figure 16. Clustering results obtained using different methods on the Pavia University data set with Gaussian noise and salt and pepper noise. (a) Ground truth. (b) LRSC 37.22%. (c) SSC 38.79%. (d) S4C 43.30%. (e) EDSC 47.91%. (f) EGCSC 50.92%. (g) EKGCSC. (h) SSAKGCSC 69.67%.
Figure 16. Clustering results obtained using different methods on the Pavia University data set with Gaussian noise and salt and pepper noise. (a) Ground truth. (b) LRSC 37.22%. (c) SSC 38.79%. (d) S4C 43.30%. (e) EDSC 47.91%. (f) EGCSC 50.92%. (g) EKGCSC. (h) SSAKGCSC 69.67%.
Axioms 11 00269 g016
Figure 17. Clustering results obtained using different methods on the Pavia Center data set with Gaussian noise and salt and pepper noise. (a) Ground truth. (b) LRSC 34.17%. (c) SSC 42.89%. (d) S4C 47.94%. (e) EDSC 48.46%. (f) EGCSC 52.03%. (g) EKGCSC 31.82%. (h) MEKGCSC 88.72%.
Figure 17. Clustering results obtained using different methods on the Pavia Center data set with Gaussian noise and salt and pepper noise. (a) Ground truth. (b) LRSC 34.17%. (c) SSC 42.89%. (d) S4C 47.94%. (e) EDSC 48.46%. (f) EGCSC 52.03%. (g) EKGCSC 31.82%. (h) MEKGCSC 88.72%.
Axioms 11 00269 g017
Figure 18. Clustering OA and CPV under a varying number of PCs on SalinasA, Indian Pines, Pavia Center and Pavia University data sets.
Figure 18. Clustering OA and CPV under a varying number of PCs on SalinasA, Indian Pines, Pavia Center and Pavia University data sets.
Axioms 11 00269 g018
Table 1. Summary of SalinasA, Indian Pines, Pavia University, and Pavia Center data sets.
Table 1. Summary of SalinasA, Indian Pines, Pavia University, and Pavia Center data sets.
Data SetsSalinasAIndian Pines CorrectedPavia UniversityPavia Center
Pixels83 × 8685 × 70140 × 150150 × 150
Channels204200103103
Clusters6487
Samples5348439164455710
SensorAVIRISAVIRISROSISROSIS
Table 2. Settings of the important hyperparameters in SSAKGCSC.
Table 2. Settings of the important hyperparameters in SSAKGCSC.
Data SetsNoiseREG_CoefGAMMANEIGHBORROpost_proCKernel
- 11000.2
SalinasA.+sp 20.010.2300.88, 18K0
+gs 31000.2
+sg 410.2
- 110,00010
Indian Pines.+sp 2100,0000.1300.817, 15K0 × K1
+gs 3100,0000.01
+sg 41,000,0000.001
- 160,00065
Pavia University+sp 2150100300.88, 18K0 × K1
+gs 310100
+sg 4500100
- 10.01100
Pavia Center+sp 21100100300.88, 18K0
+gs 30.6100
+sg 41000100
1 ‘-’ means the original image; 2 ‘sp’ means salt and pepper noise; 3 ‘gs’ means Gaussian noise; 4 ‘sg’ means salt and pepper noise and Gaussian noise.
Table 3. Clustering performance of the compared methods on SalinasA, Indian Pines, Pavia University, and Pavia Center data sets.
Table 3. Clustering performance of the compared methods on SalinasA, Indian Pines, Pavia University, and Pavia Center data sets.
DataMetricLRSC [19]SSC [18]S4C [14]EDSC [24]EGCSC [31]EKGCSC [31]SSAKGCSC
OA0.50090.75900.80700.88990.99931.00000.9989
SaA.NMI0.46570.72100.75630.85320.99711.00000.9960
Kappa0.48610.69530.76940.83270.99911.00000.9986
OA0.57520.56090.64970.70260.88270.97310.9743
InP.NMI0.51280.50810.60280.65440.69760.91280.9139
Kappa0.45660.49970.63250.62340.83080.96150.9632
OA0.46160.64270.66720.65940.84420.97320.9708
PaU.NMI0.43270.59940.61350.62520.84010.94820.9453
Kappa0.44250.58460.60470.60190.79680.96460.9538
OA0.53060.60260.72800.75390.80300.94480.9443
PaC.NMI0.44610.68010.75670.78730.81140.92520.9206
Kappa0.37730.47320.64010.67660.74470.92960.9289
The best results are highlighted in bold.
Table 4. Clustering performance of the compared methods on SalinasA, Indian Pines, Pavia University, and Pavia Center data sets with salt and pepper noise (the probability is 0.1).
Table 4. Clustering performance of the compared methods on SalinasA, Indian Pines, Pavia University, and Pavia Center data sets with salt and pepper noise (the probability is 0.1).
DataMetricLRSC [19]SSC [18]S4C [14]EDSC [24]EGCSC [31]EKGCSC [31]SSAKGCSC
OA0.43490.48600.51080.61030.68030.45740.8347
SaA.NMI0.44110.44130.53630.58270.75200.43310.8527
Kappa0.41570.42530.46340.53290.57810.26410.7905
OA0.40290.44750.50630.52880.57070.56090.6634
InP.NMI0.35640.38690.40160.43200.43300.46510.4665
Kappa0.26580.31860.33560.35340.37290.41510.5149
OA0.37690.51220.53310.53560.6012-0.7983
PaU.NMI0.43270.59940.61350.62520.8401-0.9453
Kappa0.44250.58460.60470.60190.7968-0.9538
OA0.38210.50390.52380.60330.64220.33680.8501
PaC.NMI0.42500.52810.50310.58330.61030.03820.8145
Kappa0.23180.35720.38170.50070.55370.03020.8103
The EKGCSC model will input the code segment of spectral clustering for the PaviaU data set with salt and pepper noise. The best results are highlighted in bold.
Table 5. Clustering performance of the compared methods on SalinasA, Indian Pines, Pavia University, and Pavia Center data sets with Gaussian noise (the mean is 0 and the variance is 0.2).
Table 5. Clustering performance of the compared methods on SalinasA, Indian Pines, Pavia University, and Pavia Center data sets with Gaussian noise (the mean is 0 and the variance is 0.2).
DataMetricLRSC [19]SSC [18]S4C [14]EDSC [24]EGCSC [31]EKGCSC [31]SSAKGCSC
OA0.35960.47610.54770.61970.54920.60730.8678
SaA.NMI0.32180.45360.55210.57470.63200.66390.9029
Kappa0.27640.40050.43530.48350.41800.47510.8310
OA0.43200.42200.53290.56780.59300.54700.6495
InP.NMI0.31690.30590.35630.51460.41480.39510.5348
Kappa0.28990.27440.31020.48370.40090.37740.4697
OA0.36900.36790.42510.46970.5119-0.6734
PaU.NMI0.31280.28430.37650.41670.4435-0.5902
Kappa0.25680.23630.27470.28580.3268-0.4807
OA0.31630.37810.49510.47570.48440.33150.8855
PaC.NMI0.01970.28830.53820.34040.42310.16120.8606
Kappa0.00280.19980.36290.27550.34590.10460.8537
The EKGCSC model will input the code segment of spectral clustering for the PaviaU data set with salt and pepper noise. The best results are highlighted in bold.
Table 6. Clustering performance of the compared methods on SalinasA, Indian Pines, Pavia University, and Pavia Center data sets with Gaussian noise (the mean is 0 and the variance is 0.2) and salt and pepper noise (the probability is 0.1).
Table 6. Clustering performance of the compared methods on SalinasA, Indian Pines, Pavia University, and Pavia Center data sets with Gaussian noise (the mean is 0 and the variance is 0.2) and salt and pepper noise (the probability is 0.1).
DataMetricLRSC [19]SSC [18]S4C [14]EDSC [24]EGCSC [31]EKGCSC [31]SSAKGCSC
OA0.44520.47060.50950.60790.62230.59970.8828
SaA.NMI0.32180.45360.55210.57470.63200.66390.9029
Kappa0.27640.40050.43530.48350.41800.47510.8310
OA0.41800.43860.51330.52240.54680.52490.6450
InP.NMI0.31690.30590.35630.44460.41480.39510.4958
Kappa0.28990.27440.31020.48370.40090.37740.4697
OA0.37220.38790.43300.47910.5092-0.6967
PaU.NMI0.32830.32470.38820.40820.4535-0.6102
Kappa0.26430.26950.33550.35640.3359-0.5338
OA0.34170.42890.47970.48460.52030.31820.8872
PaC.NMI0.17920.22700.37040.34340.50230.04510.8697
Kappa0.19990.22330.26680.30220.38950.1300.8564
The EKGCSC model will input the code segment of spectral clustering for the PaviaU data set with salt and pepper noise. The best results are highlighted in bold.
Table 7. Running time of different methods (in seconds).
Table 7. Running time of different methods (in seconds).
DataLRSC [19]SSC [18]S4C [14]EDSC [24]EGCSC [31]EKGCSC [31]SSAKGCSC
SaA.14,564183219,86585200340588
InP.8635134131255295160513
PaU.32,764213015,632202296514789
PaC.20,354195313,492121235398652
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Lin, Z.; Wang, J.; Wu, C. Robust Spectral Clustering Incorporating Statistical Sub-Graph Affinity Model. Axioms 2022, 11, 269. https://doi.org/10.3390/axioms11060269

AMA Style

Lin Z, Wang J, Wu C. Robust Spectral Clustering Incorporating Statistical Sub-Graph Affinity Model. Axioms. 2022; 11(6):269. https://doi.org/10.3390/axioms11060269

Chicago/Turabian Style

Lin, Zhenxian, Jiagang Wang, and Chengmao Wu. 2022. "Robust Spectral Clustering Incorporating Statistical Sub-Graph Affinity Model" Axioms 11, no. 6: 269. https://doi.org/10.3390/axioms11060269

APA Style

Lin, Z., Wang, J., & Wu, C. (2022). Robust Spectral Clustering Incorporating Statistical Sub-Graph Affinity Model. Axioms, 11(6), 269. https://doi.org/10.3390/axioms11060269

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop