Next Article in Journal
Hyperspectral Imaging for Phenotyping Plant Drought Stress and Nitrogen Interactions Using Multivariate Modeling and Machine Learning Techniques in Wheat
Previous Article in Journal
Complex Permittivity of Adobe Verses Frequency and Water Content
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

SMALE: Hyperspectral Image Classification via Superpixels and Manifold Learning

1
Institute of Intelligent Control and Image Engineering, Xidian University, Xi’an 710071, China
2
China Academy of Space Technology, Beijing 100094, China
3
Xi’an Institute of Optics and Precision Mechanics, Chinese Academy of Sciences, Xi’an 710119, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2024, 16(18), 3442; https://doi.org/10.3390/rs16183442
Submission received: 23 July 2024 / Revised: 23 August 2024 / Accepted: 12 September 2024 / Published: 17 September 2024

Abstract

:
As an extremely efficient preprocessing tool, superpixels have become more and more popular in various computer vision tasks. Nevertheless, there are still several drawbacks in the application of hyperspectral image (HSl) processing. Firstly, it is difficult to directly apply superpixels because of the high dimension of HSl information. Secondly, existing superpixel algorithms cannot accurately classify the HSl objects due to multi-scale feature categorization. For the processing of high-dimensional problems, we use the principle of PCA to extract three principal components from numerous bands to form three-channel images. In this paper, a novel superpixel algorithm called Seed Extend by Entropy Density (SEED) is proposed to alleviate the seed point redundancy caused by the diversified content of HSl. It also focuses on breaking the dilemma of manually setting the number of superpixels to overcome the difficulty of classification imprecision caused by multi-scale targets. Next, a space–spectrum constraint model, termed Hyperspectral Image Classification via superpixels and manifold learning (SMALE), is designed, which integrates the proposed SEED to generate a dimensionality reduction framework. By making full use of spatial context information in the process of unsupervised dimension reduction, it could effectively improve the performance of HSl classification. Experimental results show that the proposed SEED could effectively promote the classification accuracy of HSI. Meanwhile, the integrated SMALE model outperforms existing algorithms on public datasets in terms of several quantitative metrics.

1. Introduction

Remote sensing (RS) is a non-contact technique for acquiring the surface information of earth. It employs different sensors to capture electromagnetic radiation data across various wavelengths, followed by extraction and analysis [1,2,3]. Currently, RS technology primarily encompasses visible and infrared light, as well as microwave and hyperspectral categories. Visible light-based RS records the energy wave field of object reflections by leveraging their reflective properties. Infrared light utilizes the emission characteristics of electromagnetic waves to depict temperature information through color tones of different objects in the earth surface. Microwaves incorporate both reflection and scattering characteristics of electromagnetic waves to reflect the intensity of reflected echoes from objects. Hyperspectral imaging (HSI) refers to higher spectral resolution compared to multispectral imaging. Typically, multispectral imaging maintains spectral resolutions within an order of magnitude of around 10 1 λ , while hyperspectral RS achieves 10 2 λ .
HSI-related visual tasks are one of the most important applications of RS technology [4]. HSIs can capture images from satellites and determine what structures and terrain are on the ground based on spectral information reflected from the ground. The fundamental theory behind this groundbreaking technology is that different objects reflect different wavelengths of light [5,6,7,8].
In general, a HSI is visually simplified as a synthetic false-color image wherein a gray scale component corresponds to one band and the visual representation is a composite of three arbitrary bands. Actually, the data amount of a HSI is determined by the length, width, and number of bands, which is stored as a three-dimensional volume. Therefore, it could not only reflect the spatial position information but also records the spectral band information reflected by ground objects.
In practice, the combination of the two kinds of image information enables the subtle differences between different ground objects in spectral dimensions to be shown more comprehensively. As a result, the accuracy of ground object recognition and classification can be significantly boosted.
On the other hand, the HSI usually encompasses a multitude of spectral bands with a high-dimensional feature space, which potentially results in data redundancy. Furthermore, the strong correlation between adjacent spectral bands may lead to dimensional disaster and pose significant challenges for subsequent ground object recognition and classification [9,10,11,12]. Therefore, it is necessary for dimensionality reduction to address this issue by effectively utilizing the abundant data volume while minimizing the computational burden. Meanwhile, it is crucial to retain essential ground object information during the process of reducing feature space dimensionality. This is the effective method that can guarantee accuracy and efficiency in ground object recognition and classification.
Dimension reduction techniques of HSIs mainly include band selection and band extraction, which can be divided into linear mapping [13,14,15] and nonlinear mapping according to different standards. The dimensionality reduction method based on linear transformation is called linear mapping dimensionality reduction and mainly includes principal component analysis (PCA) [13], linear discriminant analysis (LDA) [14], independent component analysis (ICA) [15], etc. These methods are simple in principle and easy to implement. However, hyperspectral data are nonlinear in nature, and the nonlinear structure of hyperspectral data cannot be utilized perfectly by linear-mapping-dimension reduction methods. Therefore, many scholars have proposed different dimensionality reduction methods for nonlinear mapping, including kernel-based methods and eigenvalue-based methods. Kernel PCA (KPCA) [16], kernel ICA (KICA) [17], and kernel linear discriminant analysis (KLDA) [18] are all nonlinear dimension reduction methods based on kernel function. Isometric Mapping (ISOMAP) [19], locally linear embedding (LLE) [20], Laplacian eigenmaps (LEs) [21], and local tangent space alignment (LTSA) [22] belong to nonlinear dimensionality reduction methods based on eigenvalues (manifold learning). The nonlinear dimension reduction method based on manifold learning compensates for the defect of the linear-mapping-dimension reduction methods being unable find the nonlinear structure of HSI data. Advanced results have been achieved in computer vision tasks such as feature extraction [23,24], object classification [25,26,27,28], object recognition [29], and detection [30]. It is worth mentioning that the local manifold method has been proven to have excellent performance in [31].
The above dimension reduction problem is to simplify and refine the image dimension. Focusing HSIs can break through the traditional pixel-level processing method. According to the spatial dimension of HSI, pixel set (superpixel) is used as the minimum processing unit for image processing. This improvement reduces the amount of data to be dealt with and the computational complexity. Thus, the algorithm performance is improved. Superpixel segmentation divides images into non-overlapping homogeneous regions according to certain homogeneity criteria. These regions are sets of spatially adjacent and spectroscopically similar pixels whose shape and size can be adapted to the local structure of the image. Focusing on the field of RS, most HSI studies based on superpixels take full advantage of the homogeneity of superpixels to obtain more concise and rapid classification results. SuperPCA [32] was proposed by Jiang et al. in 2018, and SuperKPCA [33], of the same type, was presented by Zhang et al. in 2019. All of the above are classified by SVM [34]. However, these approaches are restricted only in the space range, which still essentially belongs to the pixel-level classification method.
The existing superpixel generation algorithm is difficult to apply ideally in HSIs. Firstly, the processing object of the superpixel algorithm is a two-dimensional image, while that of a HSI is a three-dimensional image. Secondly, the existing superpixel algorithm is not sensitive to the content in the initialization stage. When processing images with a large amount of data, the area of interest is difficult to focus on, which often leads to redundancy. Furthermore, the classification results depend heavily on the segmentation scale of superpixels.
Based on these considerations, a superpixel algorithm for HSI is proposed called Seed Extend by Entropy Density (SEED). In this work, we first focus on designing a superpixel seed redistribution framework that accurately aggregates regions of interest. The initializing strategy can directly and effectively replace the initialization strategy of grid sampling. Secondly, the five-dimensional color space information in the superpixel is combined with the spectral information of the hyperspectral data. By improving and optimizing the HSI classification framework globally, a novel spatial–spectral feature constraint model is constructed. It solves the problem of data redundancy caused by high-dimensional characteristics and the dimension problem caused by strong correlation. For the processing of high-dimensional problems, we use the PCA to extract three principal components from numerous bands to form three-channel images.
In the context of previous work [35,36,37], our improvements and contributions can be listed as follows:
(1) A superpixel algorithm, SEED, is proposed for HSI. This algorithm mainly solves the problem that it is difficult to make a reasonable initial seed setting according to the image content due to the large amount of data in HSI. It also focuses on breaking the dilemma of manually setting the number of superpixels to overcome the difficulty of classification imprecision caused by multi-scale targets in subsequent classification tasks. Objectively compared with the current excellent six algorithms [38,39,40,41,42,43], the experimental results show that the algorithm is effective and feasible. In addition, SEED can also solve the problem of uneven information granularity caused by multi-scale surface objects in hyperspectral images. (2) SEED is integrated into the dimensionality reduction framework. In the construction of the neighborhood, the constraint range can be more accurate, and the local manifold structure of the data in the low-dimensional space and the structural reconstruction error in the high-dimensional space can be reduced. This provides a solution for data redundancy caused by the high-dimensional feature and strong correlations in hyperspectral images. (3) The proposed space–spectrum model-SMALE considers the diversity of different homogenous regions in HSIs. The spatial context information is included in unsupervised dimension reduction, which improves the performance of HSI classification. Experiments on three public datasets show that the SMALE model significantly outperforms the traditional linear map-based HSI classification dimensionality reduction algorithm and some advanced feature extraction methods.
The remainder of this paper is organized as follows. Section 2 reviews a Simple Non-Iterative Clustering (SNIC) superpixel segmentation algorithm and Robust Local Manifold Representation (RLMR) [44] dimension reduction framework. In Section 3, the hyperspectral dimension reduction framework based on superpixel and manifold learning is introduced in detail. Additionally, we provide an analysis of the proposed SMALE algorithm. Section 4 presents the experimental results and analysis. Finally, the concluding remarks are stated in Section 5.

2. Related Works

2.1. Simple Non-Iterative Clustering

Simple Linear Iterative Clustering (SLIC) [45] was proposed in 2010, which can overcome the shortcomings of traditional superpixel segmentation algorithms. The algorithm has excellent robustness and significant improvement in regard to segmentation accuracy. After that, advanced superpixels emerge one after another.
Some of these algorithms tend to pursue high boundary fit, such as Linear Spectral Clustering (LSC) [46] and Entropy-Rate Superpixels (ERS) [38]. Some focus on high efficiency, such as Iterative Boundaries Implicit Identification (IBIS) [42]. By selecting the superpixel on the boundary to move, the superpixel is continuously refined, so as to update the superpixel. Rooted Spanning Superpixels (RSS) [47] have the advantage in computing time, using the concept of minimum spanning trees to guide pixels through path-cost functions. The superpixel growth rate of this algorithm is very fast. Compromising the above two situations, quite a few researchers prefer to seek a balanced performance of speed and precision. They believe that SLIC has a good framework foundation, so a class of SLIC-like superpixel algorithms led by SLIC appears. The most representative one is the Simple Non-Iterative Clustering (SNIC) [39] algorithm developed by the creator of SLIC, which utilized a priority queue to design a non-iterative clustering framework, replacing SLIC’s K-means clustering. By changing the clustering algorithm that needs multiple iterations into a non-iterative clustering algorithm, the SNIC algorithm inherits the advantages of the SLIC algorithm and reduces the time complexity and computing memory consumption.
The problem of segmentation boundary fitting caused by the multi-scale granularity of HSIs and the computational cost result of the large data volume of HSIs are considered. We are committed to studying a superpixel algorithm for high-dimensional RS images, which takes into account boundary fitting and computational cost to some extent. It is worth mentioning that this paper will be based on SNIC in order to lay a good foundation for subsequent work. SNIC is divided into three parts and described below. The segmentation demonstration effect is shown in Figure 1a.
(1) 
Initialization stage (seed initialization based on grid sampling). In the initial phase, SNIC follows the grid initialization strategy of SLIC. Equidistant sampling in horizontal and vertical directions is carried out with a fixed step size on two-dimensional images. We take the sampling point as the initial clustering center and use it as the starting point to complete the generation and updating of superpixels.
(2) 
Correlation Measurement (color space five-dimensional joint metric). It is assumed that the two-dimensional coordinate of pixel I i of image I in position space is P ( I i ) = [ x ( I i ) , y ( I i ) ] and the three-channel color feature in CIELAB color space is C ( I i ) = [ l ( I i ) , a ( I i ) , b ( I i ) ] , respectively. Based on the color space joint feature, F ( I i ) = [ C ( I i ) , P ( I i ) ] = [ l ( I i ) , a ( I i ) , b ( I i ) , x ( I i ) , y ( I i ) ] is used for five-dimensional characterization. Accordingly, the correlation measurement between the cluster center I m and the neighborhood I n is derived from the weighted Euclidean distance of the color difference and the location difference:
D ( I m , I n ) = D c ( I m , I n ) + ω D p ( I m , I n )
where ω = m S , m is a variable that we introduced to control the compactness of superpixels and the value range is [ 1 , 20 ] . For all the results in this article, we chose m = 10 . This experience value is derived from SNIC [39]. S is the side length of the superpixel cluster. S = N / K . N is the total number of image pixels, and K is the number of superpixels preset by the user.
D c ( I m , I n ) = C ( I m ) C ( I n ) 2 = ( ( l ( I m ) l ( I n ) ) 2 + ( a ( I m ) a ( I n ) ) 2 + ( b ( I m ) b ( I n ) ) 2 )
D p ( I m , I n ) = P ( I m ) P ( I n ) 2 = ( ( x ( I m ) x ( I n ) ) 2 + ( y ( I m ) y ( I n ) ) 2 )
Due to the data distribution of color space and location space being different, the value of ω reflects the importance of color and spatial location features to the similarity measurement. The different values of ω represent the variations in the characteristic tendency of superpixel boundary fit and visual comfort.
(3) 
Label allocation (allocation strategy based on online mean update). The iterative k-means algorithm is replaced by an online averaging updating system. The method of region growth is used to substitute the local candidate region traversal mode, which limits the search scope. Thus, more efficient global clustering can be achieved. In essence, this region growing is a greedy algorithm implemented using a priority queue. It converges all superpixel clusters globally into local aggregation of each cluster during the sequential generation of superpixels.

2.2. Robust Local Manifold Representation

Robust Local Manifold Representation (RLMR) [44] is an excellent hyperspectral dimension reduction method which inherits the advantages of LLE [20] and LTSA [22] in reasonably linear representation in local manifold space and can robustly solve the non-uniformity caused by data distribution. It improves the performance of classification and recognition. The RLMR comprises mainly four steps. The first two steps correspond to hierarchical neighbor selection (HNS), the third is the computation of affinity weights, and the last is the calculation of embedding.
Step 1. Joint Normalization (JN) is an effective preprocessing method for analyzing HSI. It aims to eliminate the effects of scale differences and ensure that the data are relatively uniform. This step generally begins with global data normalization (GDN) followed by local data normalization (LDN), which is a process that ranges from coarse to fine.
Step 2. Refined Neighbor Selection (RNS) mitigates the multicollinearity between spectral data to avoid affecting the subsequent model structure.
Step 3. Computation of Affinity Weights with spatial contextual information. This method adopts a robust optimization method, which has good robustness to outliers and noise.
Step 4. Low-dimensional Feature Representation is obtained by a local manifold structure in low-dimensional space.

3. Methods

In this section, we detail the hyperspectral image classification via superpixels and manifold learning, which is denoted as SMALE. Although many methods with good performance are pixel-by-pixel processing, for HSIs, pixel-by-pixel processing will bring a high probability of failure or difficult to deploy in the practical project. This is due to the large amount of HSI data, as well as information redundancy and clutter, resulting in inaccurate clustering results. SMALE maps spatial information into a dimensionality reduction framework in two key steps. This method can quickly focus on the region of interest quickly and reduce the calculation cost. First, SMALE uses SEED to obtain the HSI superpixel segmentation label with excellent performance, as described in Section 3.1. Secondly, Section 3.2 will elaborate on replacing the processing primitives of classification tasks with homogeneous superpixel blocks, mapping the original data through manifold learning into a lower-dimensional space to achieve nonlinear dimensionality reduction.

3.1. Seed Extend by Entropy Density (SEED)

In superpixel algorithms, the selection of seed points is crucial. There are methods based on uniform grids and gradient-based approaches. The grid-based method places seed points uniformly or according to a regular pattern on the image, which is used to label different regions or objects. The advantages of a grid-based seed point method include simplicity of implementation and ease of parallel processing, as shown in Figure 2b. Figure 2c uses gradient information from the image (such as Sobel or Prewitt operators) to identify pixels with high gradient magnitudes as seed points. This approach effectively captures edges and prominent transition areas in the image, making it suitable for applications that emphasize shapes and boundaries.
However, for images with uneven complexity, the grid-based seed point method does not consider intrinsic image features such as texture, color, and gradient information. This omission can lead to subsequent visual tasks lacking sufficient accuracy or adaptability. It is worth noting that gradient information effectively captures edges and transition areas in the image, but it may perform poorly in image regions with subtle gradient changes or high noise levels.
The paper improves image segmentation quality using the SEED, approached from an information-theoretic perspective. SEED can select the seed points adaptively according to the local characteristics of the image to better reflect the real structure of the image. The schematic diagram of the proposed method is shown in Figure 3 below. Figure 3b shows the distribution diagram of image content based on entropy density. According to the diagram, the seed distribution is guided, and it can be seen that simple and complex areas of the image can be easily distinguished by entropy density. Figure 3c shows the seed distribution method based on a uniform grid. The seed distribution is evenly scattered on the image without distinguishing the content.
Specifically, SEED first converts the image to grayscale, then calculates the global information entropy of the image H g ( X ) and the entropy of each sub-region H l ( X ) , as shown below:
H g ( X ) = i = 1 n p ( x i ) log p ( x i )
H l ( X ) = j = 1 k p ( x j ) log p ( x j )
where p ( x i ) represents the proportion of pixels in the image with a grayscale value of x i . n is the total number of pixels in the image. p ( x j ) represents the proportion of pixels in the pixel block with a grayscale value of x j . k is the number of pixels in the current local pixel block.
The complexity of the current superpixel block is assessed by considering the entropy distribution of its surrounding pixels. Seed pixels are chosen based on their relatively higher local entropy. The redistribution process involves adjusting the positions of the original seed pixels to better correspond with variations in entropy across the image. This approach allows for further optimization of segmentation results while maintaining the original segmentation accuracy, which is particularly effective when dealing with images containing complex textures or transition regions. An advantage of this method lies in its adaptive adjustment of seed pixel positions, thereby enhancing both the accuracy and stability of image segmentation. A schematic diagram illustrating this process is depicted in Figure 4.
The strategy involves seed redistribution in the image. Specifically, after seeds are initially selected from pixels with a high local entropy, refined seed selection occurs based on the global entropy H g ( X ) and the average entropy H a ( X ) of sub-regions.
H a ( X ) = 1 n i = 1 n H g ( X i )
  • H l ( X j ) < H g ( X ) or H a ( X ) , if the current area has more than one seed point, retain only one seed point. If there is a seed point in the current area, remove all;
  • H a ( X ) H l ( X j ) H g ( X ) , maintain the status quo;
  • H l ( X j ) > H g ( X ) or H a ( X ) , If there are no seed points in the current region, add a seed point.
The pseudocode summary of SEED is presented in Algorithm 1.
Algorithm 1: SEED superpixel segmentation framework
Input: the RGB image I , the expected number K
Output: coordinates of seeds
1/*Initialization*/
2 divided the whole image into grids.
3   calculate   the   global   entropy   H g ( X ) of the image I by Equation (3).
4 for each cluster region n  do
5           calculate   the   entropy   H l ( X ) of each sub-region by Equation (4).
6 end for
7           calculate   he   average   entropy   H a ( X ) of all n
8         for each cluster region n  do
9                    if   H l ( X j ) < H g ( X ) or H a ( X )  then
10                 retain only one seed point. (if the current area has more than one seed point); otherwise remove all.
11           else   if   H a ( X ) H l ( X j ) H g ( X )  then
12                  maintain the status quo.
13           else   H l ( X j ) > H g ( X ) or H a ( X )
14                  add a seed point. (if there are no seed points in the current region).
15         end if
16        end for
17        return coordinates of seeds

3.2. Space–Spectrum Model

HSI records continuous spectral information for each pixel, typically spanning tens to hundreds of spectral bands. It is stored as a multi-dimensional array, often conceptualized as a 3-D cube represented by M × N × D , where M and N denote the image’s height and width and D represents the number of spectral bands. Superpixel segmentation usually requires a RGB image with input m × n × 3 . When processing hyperspectral data, a SEED algorithm is not directly applicable, because D 3 in HIS.
First, we need to project the hyperspectral data into the plane space for SEED segmentation. After the completion of the superpixel segmentation, the result will be mapped back to the space of the original hyperspectral data. In this process, PCA is used for processing. PCA is able to find a new set of linearly independent variables in the original data with the largest square difference. Specifically, if the data X M × N × D has P ( P = M × N ) samples and D bands, a matrix X of size P × D can be formed:
X = [ x 11 x 12 x 1 D x 21 x 22 x 2 D x P 1 x P 1 x P D ] = ( x 1 , x 2 , , x D )
The linearly independent variable z 1 , z 2 , , z d ( d D ) with the maximum variance, and they satisfy:
{ z 1 = l 11 x 1 + l 12 x 2 + + l 1 D x D z 2 = l 21 x 1 + l 22 x 2 + + l 2 D x D z d = l d 1 x 1 + l d 2 x 2 + + l d D x D
The principles for determining coefficients l i j are as follows:
  • z 1 is unrelated to z 2 ( i j ; i , j = 1 , 2 , , d );
  • z 1 is the linear combination with the maximum variance among x 1 , x 2 , , x D ;
  • z 2 is the linear combination with the maximum variance among all linear combinations of x 1 , x 2 , , x D that are uncorrelated with z 1 ;
  • Continuing in this manner, z d is the linear combination with the maximum variance among all linear combinations of x 1 , x 2 , , x D that are uncorrelated with z 1 , z 2 , , z d 1 ;
  • The new variables z 1 , z 2 , , z d are, respectively, referred to as the first, second, …, d th principal components of the original variables x 1 , x 2 , , x D .
Following the above operations, we use superpixel blocks instead of traditional pixels. The superpixel block is treated as the new basic processing unit, and the hyperspectral data are denoted as X sp K × D , while K is the number of superpixels. These blocks represent contiguous regions in the image, where each superpixel block can be viewed as a small block that merges similar spectral and spatial information. These features better capture the spatial and spectral information integrity in HSIs.
Inspired by LLE [20], we capture the underlying local manifold structure of hyperspectral blocks processed through superpixels and preserve it in a low-dimensional space. Without loss of generality, the spatial–spectral dimensionality reduction model proposed in this paper mainly includes the steps shown in Figure 5. As explained in Section 2.2, the dimensionality reduction framework based on manifold is generally divided into the following three steps: (1) selection of neighbor points, as shown in Figure 5a; (2) weight calculation, as shown in Figure 5d; (3) embedded calculation, as shown in Figure 5e. On this basis, SMALE adds the SEED superpixel segmentation part, as shown in Figure 5c. The spatial information provided by SEED and the spectral information of the HSI make the dimensionality reduction based on the manifold more robust.
Let W K × K denote a sparse affinity matrix, where each entry ( i , j ) represents the affinity weight between the i th and j th superpixel blocks. The spectral data after dimensionality reduction is denoted as Y = [ y 1 , y 2 , , y K ] K × d , where d D . The calculation of embedding coordinates is generally formulated as follows:
Y ^ = arg min Y { i = 1 K j ϕ i y i y j 2 2 W i j } , s . t .   Y B Y T = I
where ϕ i is a set of neighbors of the i th superpixel block. B is a constant matrix defined according to different manifold learning methods.
This paper adopts the optimization framework of local linear embedding. It considers the reconstruction symmetry of each data point and its neighborhood points to design its method.
Accordingly, we apply it between the superpixel blocks, then calculate the reconstruction coefficient between the superpixel blocks and embed the calculation. The reconstruction coefficient is obtained by minimization, denoted as A ^ K × K .
A ^ = arg min A { i = 1 K x i j ϕ i A i j x i j 2 2 } , s . t .   j ϕ i A i j = 1
The formula for calculating its low-dimensional coordinates is as follows:
Y ^ = arg min Y { i = 1 K y i j ϕ i A i j y j 2 2 }   s . t .   1 K i = 1 K y i y i T = I ,   i = 1 K y i = 0 .
where it should be noted that A i j represents the reconstruction weight between x i and x j ; if j th data are not in the i th data’s neighborhood, then A i j = 0 .
Combining Equations (8) and (10), we obtain the following:
W i j = A i j + A j i A i j A j i
The optimized LLE model can not only obtain the local manifold structure but also capture the spatial characteristics of the data after superpixel processing. In the stage of neighborhood selection, the constraint range can be more accurate and the reconstruction error of local manifold structures in low dimensional space and structure reconstruction in high-dimensional space can be reduced.

4. Experiment and Discussion

This section aims to demonstrate the effectiveness of the SEED algorithm specifically designed for HSIs and its rational application within the Locally Linear Embedding (LLE) framework. While there are no readily available hyperspectral datasets tailored specifically for validating superpixel segmentation algorithms, the validity of the SEED algorithm can be verified through appropriate preprocessing and evaluation methods. Firstly, the superpixel dataset and hyperspectral dataset used in this paper are introduced. Secondly, SEED is comprehensively evaluated from both qualitative and quantitative perspectives compared against classical and state-of-the-art superpixel algorithms. Finally, an ablation experiment is designed to prove the excellent performance of the SMALE dimensionality reduction framework in classification tasks.

4.1. Experiment Setup

One superpixel segmentation dataset and two hyperspectral datasets are used in this experiment, which are described in detail in Section 4.1.1 and Section 4.1.2, respectively. All methods are executed on an Intel Core i7 4.2 GHz with 16 GB RAM without any parallelization or GPU processing.

4.1.1. Superpixel Segmentation Dataset

A Berkeley Segmentation Dataset (BSDS) [49] is a classic image segmentation dataset created by the Computer Vision Lab at the University of California, Berkeley. The BSDS contains hundreds of real-world images of natural scenes that cover a variety of complex visual situations and scenes. Each image is manually marked by an expert or highly trained marker to mark the boundaries and divisions of the individual objects. Therefore, it is considered to be a high-quality and reliable segmentation result.

4.1.2. Hyperspectral Datasets

The Indian Pines dataset [50] is a classic HSI classification dataset that is often used to test and evaluate the performance of HSI-processing and analysis algorithms. The Indian Pines dataset consists of a 145 × 145 pixels HSI covering 16 categories. Each pixel has 220 bands of information on the spectrum, which makes the dataset ideal for research and development of image processing algorithms involving spectral information.
The Salinas A dataset [32] is a widely used standard dataset for hyperspectral image analysis, sourced from an aerial RS mission over Salinas County, California. It includes HSIs with multiple bands, covering a range from visible to near-infrared wavelengths, and features high-spatial and-spectral resolutions. The image resolution is 83 × 86 pixels, with each pixel containing over 200 spectral bands. The primary uses of the Salinas A dataset include land cover classification, target detection, and anomaly detection. It provides researchers with a rich hyperspectral data source for algorithm development and performance evaluation. Due to its detailed spectral characteristics and comprehensive land cover information, the Salinas A dataset holds significant importance in regard to HSI processing and RS technology research.

4.2. Results of BSDS Data

In this section, the performance of SEED is qualitative and quantitative to verify its advancement. To demonstrate objectivity, we randomly selected 200 images in the data as a test set. We will list some of the representative segmentation effects and local details. In addition, several common evaluation indexes [51] are provided to quantify the accuracy of the image boundary and the segmentation results, such as Boundary Recall (BR), Achievable Segmentation Accuracy (ASA), Under-segmentation Error (UE), and Compactness (CO).

4.2.1. Visual Assessment

The qualitative analysis of the superpixel algorithm is demonstrated by a visual segmentation effect. Figure 6 shows seven superpixel algorithms, including SEED. SNIC is recognized as a classical superpixel generation algorithm. The ERS algorithm is the choice for most HSIs that use superpixels. Through visual analysis, we can intuitively see whether the segmentation result fits the target edge and whether there is over-segmentation. From the overall segmentation effect, the result of WSGL and ERS are messy and have poor anti-texture abilities compared with other algorithms, but these two algorithms perform well in capturing details. It can be seen that the pursuit of precision is to sacrifice the regularity of the shape of the superpixel block. It is worth mentioning that SEED seeks exactly this balance. As the partial zoom shows, the superpixels generated by SEED are uniform, compact, and, most importantly, excellent at fitting boundaries. The appearance of MBS is very neat, and the segmentation accuracy is also high, but its accuracy is not high in areas of similar colors, and it is also easily affected by similar colors. SEED also successfully overcomes this problem. Overall, SEED’s competitiveness is clear.

4.2.2. Metric Evaluation

In this section, SEED and other state-of-the-art algorithms will be comprehensively evaluated based on the four indexes commonly used to verify superpixel algorithms.
  • Boundary Recall (BR): BR is an important index to measure the ability of the algorithm to detect the real target boundary. Specifically, BR stands for the ability to correctly locate and cover the boundaries of real targets. The value of BR ranges from 0 to 1. The higher the value, the greater the proportion of the detected bounding box covering the real target boundary, i.e., the better the algorithm performance. Its calculation formula is as follows:
    BR = i Φ B ( min j Ω B P ( I i ) P ( I j ) < r ) Φ B
    where Φ B and Ω B represent the boundary pixels in the set. (   ) represents the logic value of 0 or 1. P ( I i ) = [ x ( I i ) , y ( I i ) ] and r is set to two pixels.
  • Under-segmentation Error (UE): UE evaluates the difference between the segmentation boundary generated by the algorithm and the real segmentation boundary. It focuses on areas or parts of the segmentation result that do not segment the target correctly. Specifically, it can be defined by the following formula:
    UE = m = 1 M ( Ω k | Ω k Φ m ϕ | Ω k | ) N N
    where N is the number of pixels in the image. Ω = { Ω k } k = 1 K and Φ = { Φ m } m = 1 M represent the calculated superpixels and the ground truth of the same image { I i } i = 1 N , respectively.
  • Achievable Segmentation Accuracy (ASA): ASA is an index used to evaluate the performance of segmentation algorithms in image segmentation tasks. It is designed to measure the highest level of segmentation accuracy that an algorithm can achieve. It is usually deduced or calculated by some theoretical analysis or idealized algorithm.
    ASA = k = 1 K arg max | Ω k Φ m | m = 1 M | Φ m |
  • Compactness (CO): CO is an index used to measure the compactness of segmentation results in an image segmentation evaluation. It is mainly concerned with the shape compactness of the segmented area or object.
    CO = i = 1 K 4 π ξ S P i | S P i | ψ S P i 2 | I |
    where ξ S P i and ψ S P i are expressed as the area and perimeter of the i th superpixel, respectively.
Figure 7 shows the test results of the above metrics on all algorithms. The red curve is the performance of the algorithm SEED proposed in this paper. As illustrated in Figure 7, SEED performs well in terms of the highest level of segmentation accuracy that can be achieved. At the same time, the number of pixels wrongly divided in the segmentation result is the fewest. Figure 7e is the compactness index corresponding to the qualitative analysis in Figure 7. The MBS visual effect is both compact and regular, and its CO indicator performs best. A larger CO value indicates that the shape of the divided area is not too scattered or complex.

4.3. Results of Indian Pines Dataset

In this section, we compare the proposed space–spectrum model-SMALE with traditional and state-of-the-art methods on the India Pines dataset [50]. It is worth noting that, in order to prove that SEED plays a catalytic role, we will replace the superpixel part of SMALE with MBS and ERS, designed as a cross test for benchmarking. Due to the large number of categories in the India Pines dataset, the dataset is considered challenging. As shown in Figure 8, SMALE demonstrated competitive precision in its ability to identify various crops. Classification accuracy is almost perfect in the Woods and Hay-windrowed regions. RLMR can be comparable, but its overall classification accuracy needs to be improved.
In this paper, the quantitative analysis is reflected by OA, AA, and Kappa [54]. The specific quantitative indicators are shown in Table 1.
  • Overall Accuracy (OA): OA is a measure of the proportion of the classifier’s predictions that are correct across the entire dataset. It is the most simple and intuitive classification performance evaluation indicator, and it is calculated as follows:
    OA = ( T P + T N ) ( T P + F N + F P + T N )
  • Average Accuracy (AA): AA refers to the average accuracy of each class. In a multi-class classification problem, different classes may have different sample sizes and levels of importance. AA provides a more detailed assessment by calculating the classification accuracy of each category and averaging it.
    AA = 1 N i = 1 r x i i
    where N is the total number of pixels. x i i is the value on the diagonal of its confusion matrix.
  • Kappa is a measure of consistency between the classifier’s predictions and ground truth. It takes into account the adjustment of the correctness of model predictions and the factors of random predictions, so it is particularly useful for working with categorically unbalanced datasets.
    Kappa = N i = 1 r x i i i = 1 r ( x i + × x + i ) N 2 i = 1 r ( x i + × x + i )
    where x i + is the number of actual reference pixels in each category and x + i is the total number of classified pixels in the class.
It is worth mentioning that this paper explores the influence of superpixels on the dimensionality reduction framework. In addition to the above experiments, we conducted more detailed experiments on the design of the superpixel module, conducted multi-scale experiments with different displays involving the numbers in the superpixels set, and conducted statistics for each category of the India Pines dataset. The data analysis is shown in Table 2.
Based on the above experiments, we can conclude that the optimal number of superpixels in Indian Pines is 200. The trend in the table data shows that the classification accuracy increases first and then decreases with the increase in the number of superpixels. This is because the size of the dataset is 145 × 145 × 200, and the two-dimensional size itself is not large. Setting too many superpixels will lead to over-segmentation, and the classification accuracy cannot be reached.
In order to more obviously prove that the appropriate superpixel algorithm can improve the classification accuracy, we deliberately selected all the classification algorithms based on the superpixels for further analysis (ERS-LLE, MBS-LLE, SuperPCA, S2DL, SMALE). As shown in Figure 9, there is no doubt that the best performance is SMALE, and the OA of RLMR also shows considerable advantages. Because both ERS-DR and SuperPCA employ ERS for superpixel preprocessing, their OA trends are similar. In general, when the number of superpixels is 200, the classification accuracy of each algorithm reaches its peak, which confirms the conclusion drawn from Table 2.

4.4. Results of Salinas a Dataset

To demonstrate the versatility of SMALE, we continue our experiments on the hyperspectral dataset Salinas A, as shown in Figure 10. SMALE achieved near-perfect classification accuracy on the Salinas A, outperforming its competitors in performance despite the unsupervised setup of our algorithm. The classification framework firstly applies superpixel technology to segment HSI and decompose them into regions with high spectral similarities. This step effectively reduces the computational complexity and improves the robustness of classification. Then, we use the manifold learning method to reduce the dimensionality of the superpixels and extract the structural features of the low-dimensional manifold to facilitate the subsequent classification task. Compared with Indian Pines, the performance of the ten algorithms on Salinas A has improved, but PCA still cannot fully explore the potential structure of the data, resulting in a limited classification performance. KPCA extends PCA to capture the nonlinear structure of the data through the kernel function, but, in the course of the experiment, we found that the performance of KPCA is highly dependent on the choice of kernel function. If the kernel function selected cannot effectively capture the nonlinear structure of the data, it may lead to a poor dimensionality reduction. In terms of computational complexity, when processing large-scale data, computational complexity and memory consumption may be very high, resulting in low algorithm efficiency. SuperPCA combines superpixel segmentation and traditional PCA to reduce dimensionality through superpixel blocks. This method improves the structured processing and dimensionality reduction in data. The S2DL is second only to SMALE, with an overall accuracy of 99.69%, which is also near perfect. RLMR further optimizes manifold learning by regularization technology, which makes the dimensionality reduction process more stable and efficient. The classification effect of RLMR on the Salinas A dataset is usually better than that of traditional manifold learning methods, which can provide high classification accuracy in complex environments.
In summary, the dimensionality reduction classification framework based on superpixels and manifold learning performs better with the Salinas A dataset than the simple dimensionality reduction method. By capturing the nonlinear and complex structure of the data, the classification effect is significantly improved.
Table 3 shows the quantitative classification results of SMALE and ten comparison algorithms in the Salinas A dataset, which are also explained through OA, AA, and Kappa.
It can be analyzed from Table 3 that the performances of eleven classification frameworks, including SMALE, in the Salinas A dataset have been greatly improved. Among them, the classification accuracy of eight frameworks is above 90%, and the performance of their Kappa index is also above 0.83, among which that of SMALE is as high as 0.9915. This index mainly measures the degree of consistency between classification results and random classification, and its value range is [ 1 , 1 ] . The higher the classification result is, the more consistent the classification result is with the actual category. The result of the classification is in good agreement with the actual category, which indicates that the model can predict the target category well.

5. Conclusions

In this paper, a superpixel generation framework called Seed Extend by Entropy Density (SEED) is proposed for practical application. In this work, we focus on designing a superpixel seed redistribution framework that can directly and effectively replace the initialization strategy of grid sampling. This algorithm can solve the problem that the region of interest is difficult to focus on due to the diversity of hyperspectral image content. By combining the five-dimensional color–space information in superpixels with the spectral information of hyperspectral data, a novel space–spectrum feature constraint model, SMALE, is constructed which can improve and optimize the classification of hyperspectral images globally and solve the problem that image processing takes too much time and is difficult to be practical. The experiment showed that SEED had excellent performance, reaching the highest ASA (0.9535%) and the lowest UE (0.0369%) while maintaining the leading level of BR and CO. At the same time, the SMALE model had a classification accuracy of 97.8700% in the Indian Pines dataset and 99.7453% on the Salinas A dataset. It has been proven that SEED has a positive effect on classification accuracy.
Future work will focus more on the application of superpixels in a variety of RS tasks, such as change detection, ground object recognition, ROI extraction, and other visual tasks. Future works should also make better use of superpixels as an efficient tool. We will look at designing superpixel algorithms with personalizations for different tasks.

Author Contributions

Conceptualization and methodology, N.L.; software and validation, J.G.; formal analysis and investigation, W.L.; resources, J.G.; data curation, W.L. and B.G.; writing—original draft preparation, N.L.; writing—review and editing, C.L. and N.L.; visualization, C.Z. supervision, B.G. and C.L.; project administration and funding acquisition, B.G. and C.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research is supported financially by National Natural Science Foundation of China (Grant No. 62171341).

Data Availability Statement

The BSDS500 dataset and the reference codes in this work are available at https://github.com/davidstutz/superpixel-benchmark (accessed on 29 January 2021). The Indian Pines dataset and the reference codes in this work are available at Hyperspectral Remote Sensing Scenes—Grupo de Inteligencia Computacional (GIC) (ehu.eus).

Acknowledgments

We thank the editor and anonymous reviewers for their suggestions and comments, which helped us to improve the quality of our work.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Bioucas-Dias, J.M.; Plaza, A.; Camps-Valls, G.; Scheunders, P.; Nasrabadi, N.; Chanussot, J. Hyperspectral Remote Sensing Data Analysis and Future Challenges. IEEE Geosci. Remote Sens. Mag. 2013, 1, 6–36. [Google Scholar] [CrossRef]
  2. He, L.; Li, J.; Liu, C.; Li, S. Recent Advances on Spectral-Spatial Hyperspectral Image Classification: An Overview and New Guidelines. IEEE Trans. Geosci. Remote Sens. 2018, 56, 1579–1597. [Google Scholar] [CrossRef]
  3. Han, Z.; Zhang, H.; Li, P.; Zhang, L. Hyperspectral image clustering: Current achievements and future lines. IEEE Geosci. Remote Sens. Mag. 2021, 9, 35–67. [Google Scholar]
  4. Lu, B.; Dao, P.D.; Liu, J.; He, Y.; Shang, J. Recent Advances of Hyperspectral Imaging Technology and Applications in Agriculture. Remote Sens. 2020, 12, 2659. [Google Scholar] [CrossRef]
  5. Fauvel, M.; Tarabalka, Y.; Benediktsson, J.A.; Chanussot, J.; Tilton, J.C. Advances in spectral-spatial classification of hyperspectral images. Proc. IEEE. 2013, 101, 652–675. [Google Scholar] [CrossRef]
  6. Zhang, L.; Zhang, L.; Du, B. Deep learning for remote sensing data: A technical tutorial on the state of the Art. IEEE Geosci. Remote Sens. Mag. 2016, 4, 22–40. [Google Scholar] [CrossRef]
  7. Ma, J.; Jiang, J.; Zhou, H.; Zhao, J.; Guo, X. Guided locality pre-serving feature matching for remote sensing image registration. IEEE Trans. Geosci. Remote Sens. 2018, 56, 4435–4447. [Google Scholar] [CrossRef]
  8. Ma, J.; Chen, C.; Li, C.; Huang, J. Infrared and visible image fusion via gradient transfer and total variation minimization. Inf. Fusion 2016, 31, 100–109. [Google Scholar] [CrossRef]
  9. Melgani, F.; Bruzzone, L. Classification of hyperspectral remote sensing images with support vector machines. IEEE Trans. Geosci. Remote Sens. 2004, 42, 1778–1790. [Google Scholar] [CrossRef]
  10. Liu, G.; Wang, L.; Liu, D.; Fei, L.; Yang, J. Hyperspectral Image Classification Based on Non-Parallel Support Vector Machine. Remote Sens. 2022, 14, 2447. [Google Scholar] [CrossRef]
  11. Kang, J.; Zhang, Y.; Liu, X.; Cheng, Z. Hyperspectral Image Classification Using Spectral–Spatial Double-Branch Attention Mechanism. Remote Sens. 2024, 16, 193. [Google Scholar] [CrossRef]
  12. Liu, G.; Wang, L.; Liu, D. Hyperspectral Image Classification Based on a Least Square Bias Constraint Additional Empirical Risk Minimization Nonparallel Support Vector Machine. Remote Sens. 2022, 14, 4263. [Google Scholar] [CrossRef]
  13. Schölkopf, B.; Smola, A.; Müller, K.-R. Nonlinear component analysis as a kernel eigenvalue problem. Neural Comput. 1998, 10, 1299–1319. [Google Scholar] [CrossRef]
  14. Prasad, S.; Bruce, L.M. Limitations of principal components analysisfor hyperspectral target recognition. IEEE Geosci. Remote Sens. Lett. 2008, 5, 625–629. [Google Scholar] [CrossRef]
  15. Wang, J.; Chang, C.-I. Independent component analysis-based dimensionality reduction with applications in hyperspectral image analysis. IEEE Trans. Geosci. Remote Sens. 2006, 44, 1586–1600. [Google Scholar] [CrossRef]
  16. Schölkopf, B.; Smola, A.; Müller, K.R. Kernel principal component analysis. In Artificial Neural Networks—ICANN’97; Gerstner, W., Germond, A., Hasler, M., Nicoud, J.D., Eds.; Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 1997; Volume 1327, pp. 583–588. [Google Scholar]
  17. Bach, F.R.; Jordan, M.I. Kernel independent component analysis. J. Mach. Learn. Res. 2002, 3, 1–48. [Google Scholar]
  18. Mika, S.; Ratsch, G.; Weston, J.; Scholkopf, B.; Mullers, K.R. Fisher discriminant analysis with kernels. In Neural Networks for Signal Processing IX: Proceedings of the 1999 IEEE Signal Processing Society Workshop (Cat. No.98TH8468); IEEE: Piscataway, NJ, USA, 1999; pp. 41–48. [Google Scholar]
  19. Tenenbaum, J.B.; De Silva, V.; Langford, J.C. A global geometric framework for nonlinear dimensionality reduction. Science 2000, 290, 2319–2323. [Google Scholar] [CrossRef]
  20. Roweis, S.T.; Saul, L.K. Nonlinear dimensionality reduction by locally linear embedding. Science 2000, 290, 2323–2326. [Google Scholar] [CrossRef]
  21. Belkin, M.; Niyogi, P. Laplacian eigenmaps for dimensionality reduction and data representation. Neural Comput. 2003, 15, 1373–1396. [Google Scholar] [CrossRef]
  22. Zhang, Z.; Zha, H. Principal manifolds and nonlinear dimension reduction via local tangent space alignment. SIAM J. Sci. Comput. 2004, 26, 313–338. [Google Scholar] [CrossRef]
  23. Bachmann, C.M.; Ainsworth, T.L.; Fusina, R.A. Exploiting manifold geometry in hyperspectral imagery. IEEE Trans. Geosci. Remote Sens. 2005, 43, 441–454. [Google Scholar] [CrossRef]
  24. He, J.; Zhang, L.; Wang, Q.; Li, Z. Using diffusion geometric coordinates for hyperspectral imagery representation. IEEE Trans. Geosci. Remote Sens. 2009, 6, 767–771. [Google Scholar]
  25. Ma, L.; Crawford, M.M.; Tian, J.W. Local manifold learning-based k-nearest-neighbor for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2010, 48, 4099–4109. [Google Scholar] [CrossRef]
  26. Ma, L.; Crawford, M.M.; Yang, X.; Guo, Y. Local-manifold-learning-based graph construction for semisupervised hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2015, 53, 2832–2844. [Google Scholar] [CrossRef]
  27. Huang, H.; Huo, H.; Fang, T. Hierarchical manifold learning with application to supervised classification for high-resolution remotely sensed images. IEEE Trans. Geosci. Remote Sens. 2013, 52, 1677–1692. [Google Scholar] [CrossRef]
  28. Tan, Y.; Yuan, H.; Li, L. Manifold-based sparse representation for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2014, 52, 7606–7618. [Google Scholar]
  29. Yang, H.L.; Crawford, M.M. Spectral and spatial proximity-based manifold alignment for multitemporal hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2016, 54, 51–64. [Google Scholar] [CrossRef]
  30. Ma, L.; Crawford, M.M.; Tian, J.W. Anomaly detection for hyperspectral images based on robust locally linear embedding. J. Infrared Millim. THz Waves 2010, 31, 753–763. [Google Scholar] [CrossRef]
  31. Sui, C.; Li, C.; Feng, J.; Mei, X. Unsupervised Manifold-Preserving and Weakly Redundant Band Selection Method for Hyperspectral Imagery. IEEE Trans. Geosci. Remote Sens. 2020, 58, 1156–1170. [Google Scholar] [CrossRef]
  32. Jiang, J.; Ma, J.; Chen, C.; Wang, Z.; Cai, Z.; Wang, L. SuperPCA: A Superpixelwise PCA Approach for Unsupervised Feature Extraction of Hyperspectral Imagery. IEEE Trans. Geosci. Remote Sens. 2018, 56, 4581–4593. [Google Scholar] [CrossRef]
  33. Zhang, L.; Su, H.; Shen, J. Hyperspectral Dimensionality Reduction Based on Multiscale Superpixelwise Kernel Principal Component Analysis. Remote Sens. 2019, 11, 1219. [Google Scholar] [CrossRef]
  34. Cortes, C.; Vapnik, V. Support-Vector Networks. Mach. Learn. 1995, 20, 273–297. [Google Scholar] [CrossRef]
  35. Liao, N.; Guo, B.; Li, C.; Liu, H.; Zhang, C. BACA: Superpixel segmentation with boundary awareness and content adaptation. Remote Sens. 2022, 14, 4572. [Google Scholar] [CrossRef]
  36. Li, C.; Guo, B.; Liao, N.; Gong, J.; He, W. CONIC: Contour Optimized Non-Iterative Clustering Superpixel Segmentation. Remote Sens. 2021, 13, 1061. [Google Scholar] [CrossRef]
  37. Liao, N.; Guo, B.; He, F.; Li, W.; Li, C.; Liu, H. Spherical Superpixel Segmentation with Context Identity and Contour Intensity. Symmetry 2024, 16, 925. [Google Scholar] [CrossRef]
  38. Liu, M.-Y.; Tuzel, O.; Ramalingam, S.; Chellappa, R. Entropy rate superpixel segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Colorado Springs, CO, USA, 20–25 June 2011; pp. 2097–2104. [Google Scholar]
  39. Achanta, R.; Susstrunk, S. Superpixels and polygons using simple non-iterative clustering. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 4651–4660. [Google Scholar]
  40. Hu, Y.; Li, Y.; Song, R.; Rao, P.; Wang, Y. Minimum barrier superpixel segmentation. Image Vis. Comput. 2018, 70, 1–10. [Google Scholar] [CrossRef]
  41. Yuan, Y.; Zhu, Z.; Yu, H.; Zhang, W. Watershed-Based Superpixels with Global and Local Boundary Marching. IEEE Trans. Image Process. 2020, 29, 7375–7388. [Google Scholar] [CrossRef]
  42. Bobbia, S.; Macwan, R.; Benezeth, Y. Iterative Boundaries implicit Identification for superpixels Segmentation: A real-time approach. IEEE Access 2021, 9, 77250–77263. [Google Scholar] [CrossRef]
  43. Li, C.; He, W.; Liao, N.; Gong, J.; Hou, S.; Guo, B. Superpixels with contour adherence via label expansion for image decomposition. Neural Comput. Appl. 2022, 34, 16223–16237. [Google Scholar] [CrossRef]
  44. Hong, D.; Yokoya, N.; Zhu, X.X. Learning a Robust Local Manifold Representation for Hyperspectral Dimensionality Reduction. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2017, 10, 2960–2975. [Google Scholar] [CrossRef]
  45. Achanta, R.; Shaji, A.; Smith, K.; Lucchi, A.; Fua, P.; Susstrunk, S. SLIC superpixels compared to state-of-the-art superpixel methods. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 34, 2274–2282. [Google Scholar] [CrossRef] [PubMed]
  46. Chen, J.; Li, Z.; Huang, B. Linear spectral clustering superpixel. IEEE Trans. Image Process. 2017, 26, 3317–3330. [Google Scholar] [CrossRef] [PubMed]
  47. Chai, D. Rooted Spanning Superpixels. Int. J. Comput. Vis. 2020, 128, 2962–2978. [Google Scholar] [CrossRef]
  48. Van den Bergh, M.; Boix, X.; Roig, G.; Van Gool, L. SEEDS: Superpixels extracted via energy-driven sampling. Int. J. Comput. Vis. 2015, 111, 298–314. [Google Scholar] [CrossRef]
  49. Arbelaez, P.; Maire, M.; Fowlkes, C.; Malik, J. Contour detection and hierarchical image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. (TPAMI) 2011, 33, 898–916. [Google Scholar] [CrossRef] [PubMed]
  50. Anand, R.; Veni, S.; Aravinth, J. Robust Classification Technique for Hyperspectral Images Based on 3D-Discrete Wavelet Transform. Remote Sens. 2021, 13, 1255. [Google Scholar] [CrossRef]
  51. Wang, M.; Liu, X.; Gao, Y.; Ma, X.; Soomro, N. Superpixel segmentation: A benchmark. Signal Process. Image Commun. 2017, 56, 28–39. [Google Scholar] [CrossRef]
  52. Cui, K.; Li, R.; Polk, S.L.; Lin, Y.; Zhang, H.; Murphy, J.M.; Plemmons, R.J.; Chan, R.H. Superpixel-Based and Spatially Regularized Diffusion Learning for Unsupervised Hyperspectral Image Clustering. IEEE Trans. Geosci. Remote Sens. 2024, 62, 1–18. [Google Scholar] [CrossRef]
  53. Polk, S.L.; Cui, K.; Chan, A.H.; Coomes, D.A.; Plemmons, R.J.; Murphy, J.M. Unsupervised diffusion and volume maximization-based clustering of hyperspectral images. Remote Sens. 2023, 15, 1053. [Google Scholar] [CrossRef]
  54. Prabhakar, T.N.; Xavier, G.; Geetha, P.; Soman, K.P. Spatial preprocessing based multinomial logistic regression for hyperspectral image classification. Proc. Comput. Sci. 2015, 46, 1817–1826. [Google Scholar] [CrossRef]
Figure 1. Two typical superpixel algorithms. (a) The segmentation process of SNIC [39] and its variants algorithm; (b) The segmentation process of SEEDS [48].
Figure 1. Two typical superpixel algorithms. (a) The segmentation process of SNIC [39] and its variants algorithm; (b) The segmentation process of SEEDS [48].
Remotesensing 16 03442 g001
Figure 2. Two frequently used seed point methods. (a) Input image; (b) grid-based seed point method; (c) gradient-based seed point method.
Figure 2. Two frequently used seed point methods. (a) Input image; (b) grid-based seed point method; (c) gradient-based seed point method.
Remotesensing 16 03442 g002
Figure 3. Two frequently used seed point methods. (a) Input image; (b) seeds are distributed adaptively according to their entropy density; (c) seeds are evenly distributed throughout the image.
Figure 3. Two frequently used seed point methods. (a) Input image; (b) seeds are distributed adaptively according to their entropy density; (c) seeds are evenly distributed throughout the image.
Remotesensing 16 03442 g003
Figure 4. Dynamic segmentation procedure of SEED. (a) Input image; (b) grid-sampled seeds; (c) gray level image; (d) entropy density distribution diagram Entropy density values range from small to large, indicating colors from blue to yellow; (eg) seed distribution is dynamically adjusted according to different values of the entropy density. Among them, yellow boxes are examples of seed extend paths in regions with larger entropy density, and blue boxes are examples of seed extend paths in regions with smaller entropy density.
Figure 4. Dynamic segmentation procedure of SEED. (a) Input image; (b) grid-sampled seeds; (c) gray level image; (d) entropy density distribution diagram Entropy density values range from small to large, indicating colors from blue to yellow; (eg) seed distribution is dynamically adjusted according to different values of the entropy density. Among them, yellow boxes are examples of seed extend paths in regions with larger entropy density, and blue boxes are examples of seed extend paths in regions with smaller entropy density.
Remotesensing 16 03442 g004
Figure 5. The procedure of the spatial–spectral dimensionality reduction model. (a) Neighbor selection; (b) data normalization; (c) superpixel segmentation; (d) compute weights; (e) calculation of embedding.
Figure 5. The procedure of the spatial–spectral dimensionality reduction model. (a) Neighbor selection; (b) data normalization; (c) superpixel segmentation; (d) compute weights; (e) calculation of embedding.
Remotesensing 16 03442 g005
Figure 6. Visual comparison of segmentation results with 100 expected superpixels. (a) ERS [38]; (b) SNIC [39]; (c) MBS [40]; (d) WSGL [41]; (e) IBIS [42]; (f) SCALE [43]; (g) SEED. Alternating columns show each segmented image followed by the zoom-in performance.
Figure 6. Visual comparison of segmentation results with 100 expected superpixels. (a) ERS [38]; (b) SNIC [39]; (c) MBS [40]; (d) WSGL [41]; (e) IBIS [42]; (f) SCALE [43]; (g) SEED. Alternating columns show each segmented image followed by the zoom-in performance.
Remotesensing 16 03442 g006
Figure 7. Quantitative evaluation of different algorithms on four evaluation indicators. (a) Boundary recall; (b) under-segmentation error; (c) precision recall; (d) achievable segmentation accuracy; (e) compactness; (f) execution time. The expected number of superpixels ranges from 50 to 500.
Figure 7. Quantitative evaluation of different algorithms on four evaluation indicators. (a) Boundary recall; (b) under-segmentation error; (c) precision recall; (d) achievable segmentation accuracy; (e) compactness; (f) execution time. The expected number of superpixels ranges from 50 to 500.
Remotesensing 16 03442 g007
Figure 8. Classification maps for the Indian Pines dataset using all DR methods under comparison with the optimal parameters. (a) Input image; (b) ground truth; (c) PCA [13]; (d) KPCA [16]; (e) LLE [20]; (f) LTSA [22]; (g) SuperPCA [32]; (h) RLMR [44]; (i) S2DL [52]; (j) D-VIC [53]; (k) ERS-LLE; (l) MBS-LLE; (m) SMALE.
Figure 8. Classification maps for the Indian Pines dataset using all DR methods under comparison with the optimal parameters. (a) Input image; (b) ground truth; (c) PCA [13]; (d) KPCA [16]; (e) LLE [20]; (f) LTSA [22]; (g) SuperPCA [32]; (h) RLMR [44]; (i) S2DL [52]; (j) D-VIC [53]; (k) ERS-LLE; (l) MBS-LLE; (m) SMALE.
Remotesensing 16 03442 g008
Figure 9. OA and AA indicators of the proposed method and other methods using the superpixels in the Indian Pines dataset.
Figure 9. OA and AA indicators of the proposed method and other methods using the superpixels in the Indian Pines dataset.
Remotesensing 16 03442 g009
Figure 10. Classification maps for the Salinas A dataset using all DR methods under comparison with the optimal parameters. (a) Input image; (b) ground truth; (c) PCA [13]; (d) KPCA [16]; (e) LLE [20]; (f) LTSA [22]; (g) SuperPCA [32]; (h) RLMR [44]; (i) S2DL [52]; (j) D-VIC [53]; (k) ERS-DR; (l) MBS-DR; (m) SMALE.
Figure 10. Classification maps for the Salinas A dataset using all DR methods under comparison with the optimal parameters. (a) Input image; (b) ground truth; (c) PCA [13]; (d) KPCA [16]; (e) LLE [20]; (f) LTSA [22]; (g) SuperPCA [32]; (h) RLMR [44]; (i) S2DL [52]; (j) D-VIC [53]; (k) ERS-DR; (l) MBS-DR; (m) SMALE.
Remotesensing 16 03442 g010aRemotesensing 16 03442 g010b
Table 1. Classification result of the proposed method and ten comparison algorithms on an Indian Pines dataset.
Table 1. Classification result of the proposed method and ten comparison algorithms on an Indian Pines dataset.
Evaluation
Index
Algorithm
PCAKPCALLELTSASuperPCARLMRS2DLD-VICERS-LLEMBS-LLESMALE
OA (%)64.3567.0368.2372.0689.2680.6573.2552.3578.4180.7890.74
AA (%)73.4276.9875.7180.9693.5589.6665.3253.2786.7790.6595.28
Kappa0.33150.39960.43650.47350.50080.52110.59200.40200.53320.55720.5691
Table 2. Performance of the proposed SMALE model on the Indian Pines dataset, with the expected number of superpixels ranging from 50 to 500.
Table 2. Performance of the proposed SMALE model on the Indian Pines dataset, with the expected number of superpixels ranging from 50 to 500.
Class NamesExpected Superpixel Number
50100150200250300350400450500
Corn-notill68.5772.8978.2293.6587.3685.6480.3770.3666.7562.37
Corn-mintill83.8890.0992.5196.8795.6694.2893.7189.6582.5480.32
Corn92.6595.8896.4199.2198.5597.6297.2194.5690.9890.23
Grass-pasture92.3693.5593.9696.8895.1294.8794.4592.8991.6390.72
Grass-trees82.6389.6890.6396.8795.2194.6692.4585.5280.8778.65
Hay-windrowed97.3297.3297.3210099.6299.2398.7697.3296.7895.62
Soybean-notill85.7490.4391.2595.2292.2291.2591.2589.6680.6570.85
Soybean-mintill90.2591.6594.5597.1096.1095.4395.2290.3689.6788.34
Soybean-clean80.6685.9886.8794.9892.5590.0289.2181.3079.6578.33
Wheat97.6599.5699.5698.5699.2099.2199.5698.1396.2695.13
Woods80.4890.3191.4798.7798.4498.2295.6781.6178.1570.26
Bldg-Gra-Tr-Driv96.3698.0698.0010099.2598.6598.4397.2894.5890.88
Stone-Steel-Towers99.0210010098.9798.9799.3399.5210098.5597.65
Alfalfa96.5297.2297.8610010099.6598.1096.6695.3294.99
Grass-pasture-mowed96.1197.8997.8998.8997.8997.8997.8997.8595.6594.98
Oats98.6599.0199.7710010010010098.7896.0095.02
OA (%)87.5292.1593.0095.8994.6294.2693.4590.3385.6682.26
AA (%)89.9393.0994.1497.8796.6495.9995.1191.3788.3885.89
Table 3. Classification result of the proposed method and ten comparison algorithms on the Salinas A dataset.
Table 3. Classification result of the proposed method and ten comparison algorithms on the Salinas A dataset.
Evaluation Index Algorithm
PCAKPCALLELTSASuperPCARLMRS2DLD-VICERS-LLEMBS-LLESMALE
OA (%)86.5488.6289.2392.0697.2693.5699.1096.5198.2398.4199.28
AA (%)86.9888.9892.7193.9697.5595.2599.6997.2098.8798.6599.74
Kappa0.83150.86960.834150.87650.93080.92110.98400.96520.96320.95720.9915
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Liao, N.; Gong, J.; Li, W.; Li, C.; Zhang, C.; Guo, B. SMALE: Hyperspectral Image Classification via Superpixels and Manifold Learning. Remote Sens. 2024, 16, 3442. https://doi.org/10.3390/rs16183442

AMA Style

Liao N, Gong J, Li W, Li C, Zhang C, Guo B. SMALE: Hyperspectral Image Classification via Superpixels and Manifold Learning. Remote Sensing. 2024; 16(18):3442. https://doi.org/10.3390/rs16183442

Chicago/Turabian Style

Liao, Nannan, Jianglei Gong, Wenxing Li, Cheng Li, Chaoyan Zhang, and Baolong Guo. 2024. "SMALE: Hyperspectral Image Classification via Superpixels and Manifold Learning" Remote Sensing 16, no. 18: 3442. https://doi.org/10.3390/rs16183442

APA Style

Liao, N., Gong, J., Li, W., Li, C., Zhang, C., & Guo, B. (2024). SMALE: Hyperspectral Image Classification via Superpixels and Manifold Learning. Remote Sensing, 16(18), 3442. https://doi.org/10.3390/rs16183442

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop