Next Article in Journal
Data Clustering Utilization Technologies Using Medians of Current Values for Improving Arc Sensing in Unstructured Environments
Next Article in Special Issue
A Dual-Branch Fusion of a Graph Convolutional Network and a Convolutional Neural Network for Hyperspectral Image Classification
Previous Article in Journal
Carbazolyl-Modified Neutral Ir(III) Complexes for Efficient Detection of Picric Acid in Aqueous Media
Previous Article in Special Issue
Pixel-Level Recognition of Trace Mycotoxins in Red Ginseng Based on Hyperspectral Imaging Combined with 1DCNN-Residual-BiLSTM-Attention Model
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Advanced Hyperspectral Image Analysis: Superpixelwise Multiscale Adaptive T-HOSVD for 3D Feature Extraction

1
School of Artificial Intelligence, Hangzhou Dianzi University, Hangzhou 310018, China
2
School of Computer Science and Technology, Hangzhou Dianzi University, Hangzhou 310018, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(13), 4072; https://doi.org/10.3390/s24134072
Submission received: 21 May 2024 / Revised: 18 June 2024 / Accepted: 19 June 2024 / Published: 22 June 2024

Abstract

:
Hyperspectral images (HSIs) possess an inherent three-order structure, prompting increased interest in extracting 3D features. Tensor analysis and low-rank representations, notably truncated higher-order SVD (T-HOSVD), have gained prominence for this purpose. However, determining the optimal order and addressing sensitivity to changes in data distribution remain challenging. To tackle these issues, this paper introduces an unsupervised Superpixelwise Multiscale Adaptive T-HOSVD (SmaT-HOSVD) method. Leveraging superpixel segmentation, the algorithm identifies homogeneous regions, facilitating the extraction of local features to enhance spatial contextual information within the image. Subsequently, T-HOSVD is adaptively applied to the obtained superpixel blocks for feature extraction and fusion across different scales. SmaT-HOSVD harnesses superpixel blocks and low-rank representations to extract 3D features, effectively capturing both spectral and spatial information of HSIs. By integrating optimal-rank estimation and multiscale fusion strategies, it acquires more comprehensive low-rank information and mitigates sensitivity to data variations. Notably, when trained on subsets comprising 2%, 1%, and 1% of the Indian Pines, University of Pavia, and Salinas datasets, respectively, SmaT-HOSVD achieves impressive overall accuracies of 93.31%, 97.21%, and 99.25%, while maintaining excellent efficiency. Future research will explore SmaT-HOSVD’s applicability in deep-sea HSI classification and pursue additional avenues for advancing the field.

Graphical Abstract

1. Introduction

Hyperspectral images (HSIs) are acquired through dedicated sensors such as AVIRIS, Hyperion, and CHRIS, capturing numerous spectral bands across Earth’s geography [1,2]. Distinct species exhibit spectral separability, reflecting unique spectral information characteristics [3]. Unlike conventional images, HSI extends beyond spatial features, offering detailed one-dimensional spectral profiles per pixel [4]. At the same time, hyperspectral imaging has a wide range of potential in various electromagnetic radiation frequency ranges, such as X-ray, ultraviolet, visible and near-infrared, and terahertz [5,6,7,8,9]. With applications spanning agriculture, geology, ecology, and mineralogy [10,11,12,13], HSI faces challenges in classification due to its high spectral resolution and noise, potentially leading to dimensionality disaster and reduced classification accuracy, especially with limited training samples [14,15]. Hence, efficient feature extraction is pivotal for HSI classification tasks.
Various feature extraction methods have emerged, from classical linear techniques like principal component analysis (PCA) and linear discriminant analysis (LDA) to nonlinear approaches such as neighborhood preserving embedding (NPE) and locality preserving projections (LPP) [16,17,18,19]. However, these methods often neglect spatial detail and concentrate solely on spectral information, limiting their effectiveness [20].
To address this, approaches integrating spatial and spectral information have gained prominence, categorized into segmentation-based, decision-based fusion, and feature fusion methods [21]. In segmentation-based methods, superpixelwise PCA (SuperPCA) and superpixelwise adaptive singular spectral analysis (SpaSSA) emphasize localized feature extraction with superpixel segmentation [22,23]. Decision-based fusion methods like robust matrix discriminative analysis (RMDA) combine multiple classification decisions for improved integration [19]. Yet, these methods often treat spectral and spatial information independently, overlooking their intrinsic connection within HSI [20].
Recognizing HSI’s cubic form encompassing spatial and spectral dimensions [21], feature extraction in 3D has emerged, notably through deep learning and tensor-based methods. Deep learning, exemplified by 3D convolutional neural networks (3DCNNs) and transformer-based networks like SpectralFormer and spectral-spatial transformer network (SSTN), has shown promise in extracting deeper features [24,25,26,27]. Despite the substantial progress made by the aforementioned methods, they share common challenges inherent in deep learning, such as the need to determine a large number of hyperparameters and the lack of model interpretability. The tensor-based approach exploits the low-rank structure of HSI, with algorithms such as Tucker decomposition and CANDECOMP/PARAFAC (CP) efficiently extracting 3D features [28,29]. Fu et al. [20] introduced tensor singular spectral analysis (TensorSSA) to extract global and low-rank features of HSI by constructing the global trajectory tensor of HSI. Nonetheless, the low-rank representation of the tensor is sensitive to initialization and parameter selection and is less universally applicable across all types of data, indicating its limitations.
One notable approach is truncated higher-order singular value decomposition (T-HOSVD), a form of Tucker decomposition that compresses tensors for denoising and approximate low-rank representation [30,31]. However, T-HOSVD faces challenges in handling diverse and complex HSI data, along with the difficulty of determining optimal ranks [32]. Therefore, it is necessary to explore an improved T-HOSVD method that can adaptively perform rank estimation while reducing data sensitivity and making the algorithm more robust for better 3D feature extraction of HSI.
To overcome these challenges, this paper proposes a novel method, Superpixelwise Multiscale Adaptive T-HOSVD (SmaT-HOSVD), which integrates local features and low-rank attributes of HSI for 3D feature extraction. Spectral and spatial features are extracted in 3D through superpixel and Multiscale Adaptive T-HOSVD to obtain more comprehensive low-rank information and reduce sensitivity to data, thereby enhancing classification tasks.
The primary contributions of SmaT-HOSVD include the following:
(1)
Introducing SmaT-HOSVD, a 3D feature extraction method utilizing superpixels and Multiscale Adaptive T-HOSVD, which fully utilizes spectral and spatial information and enhances the comprehensiveness of features;
(2)
Employing adaptive rank estimation based on the energy distribution of data, enhancing noise separation and inter-class distinction. In addition, a low-rank fusion strategy is utilized to generate multilevel features to enhance the comprehensiveness of the features and the robustness of the algorithm;
(3)
Demonstrating low time complexity and memory footprint, enhancing computational efficiency, and validating performance across multiple public datasets.

2. Related Works

We adopt italicized letters for scalars (e.g., x and X), bold letters for matrices (e.g., X), and calligraphic letters for tensors (e.g., X ). denotes the real number fields. X I 1 × I n × × I N denotes a third-order tensor.

2.1. Notations and Definitions

Definition 1
(n-mode flattening matrix). An n-mode flattening matrix is obtained by unfolding a tensor along its in dimension and then flattening it to a two-dimensional array. For a tensor  X I 1 × I n × × I N , the resulting matrix is denoted as
X n I n × ( I 1 × I 2 × × I n 1 × I n + 1 × × I N )
Definition 2
(n-mode product [33]). The n-mode product of a tensor X I 1 × I n × × I N with a matrix U J × I n is denoted by γ = X × n U I 1 × × I n 1 × I j × I n + 1 × × I N , whose elements can be denoted by γ i 1 i 2 i n 1 j i n + 1 i N = i n = 1 I n X i 1 i 2 i n 1 i n i n + 1 i N U j i n . This operation can also be expressed through the matrix:
γ = X n U Y ( n ) = U X ( n )
Definition 3
(singular value decomposition (SVD) [34]). For the matrix X n 1 × n 2 , the SVD of X is given by
X = U V *
where U is an n 1 × n 1 complex unitary matrix, is an n 1 × n 2 rectangular diagonal matrix, V is an n 2 × n 2 complex unitary matrix, and V* is the conjugate transpose of V.

2.2. T-HOSVD

The truncated higher-order singular value decomposition (T-HOSVD) [30] is a tensor decomposition technique that has been successfully applied to denoising [35], face recognition [36], and various other applications. Notably, T-HOSVD exhibits outstanding performance in HSI.
For HSI X I 1 × I 2 × I 3 with a multilinear rank denoted by rank- ( R ¯ 1 , R ¯ 2 , R ¯ 3 ) , we can determine the optimal X ¯ that approximates X by solving an optimization problem:
min X ¯ R I 1 × I 2 × I 3 X X ¯ F 2   s . t .   rank - ( R ¯ 1 , R ¯ 2 , R ¯ 3 )
where ( R ¯ 1 , R ¯ 2 , R ¯ 3 ) is the reduced multilinear rank, and the F is the Frobenius norm. To solve this optimization problem, we can employ a simple method, namely, T-HOSVD. The detailed steps of T-HOSVD are as follows.
Unfolding X obtains a matrix X n by n-mode, compute a rank- R ¯ n truncated SVD:
X n = U n n V n T ,   n = 1 , 2 , 3
and store the top R ¯ n left singular vectors U n R I n × R ¯ n , then compute the core tensor G = X × 1 U 1 × 2 U 2 × 3 U 3 by the n-mode product. Finally, the approximation tensor can be reformulated as
X ¯ = G × 1 U 1 × 2 U 2 × 3 U 3
Tucker decomposition offers numerous advantages and finds applications across various fields [37]. De Lathauwer and Vandewalle [38] have investigated the application of the Tucker decomposition in the field of signal processing. Vasilescu and Terzopoulos [39] were pioneers in introducing the Tucker decomposition for applications in computer vision. As a specific instance of orthogonal Tucker decomposition, T-HOSVD serves as a generalization of matrix singular value decomposition. While T-HOSVD does not offer an optimal multilinear rank approximation [40], in practical applications, it tends to produce satisfactory solutions when optimal solutions with small errors exist. However, T-HOSVD encounters two significant challenges in the field of HSI processing. Firstly, HSI often contain diverse targets with multilayered and complex information, making T-HOSVD susceptible to variations in data distribution and reducing its robustness. Secondly, determining the optimal rank for an HSI remains a challenging problem. Typically, the optimal rank is computed by transforming rank estimation into an iterative optimization problem with constraints. However, this approach is time-consuming, and tuning certain parameters can be difficult. To enhance the solution and minimize errors, we incorporate adaptive low-rank estimation and multiscale fusion strategies in our approach. These strategies contribute to a more comprehensive and robust extraction of features from HSI, as detailed in Section 3.2.

3. Methods

The flowchart of the proposed SmaT-HOSVD method is depicted in Figure 1, comprising the following three parts: (1) Superpixel segmentation; (2) Superpixelwise Multiscale Adaptive T-HOSVD; (3) Classification. The details of the proposed SmaT-HOSVD are described as follows.

3.1. Superpixel Segmentation

Superpixel segmentation involves partitioning pixels into several irregular pixel blocks known as superpixels [41]. The boundaries of these superpixels align relatively well with the actual boundaries of the image, facilitating subsequent task processing [42]. Superpixel algorithms can be broadly categorized into two types: graph-based superpixel algorithms and gradient descent-based superpixel algorithms. Classic algorithms include entropy rate superpixel (ERS) [43] and simple linear iterative clustering [41], which are widely used in HSI preprocessing.
ERS transforms the superpixel segmentation problem into solving the objective function with the largest entropy rate on the graph. In ERS, images are mapped to a graph, where each pixel is represented as a vertex, and edge weights represent pairwise similarities [42]. By solving the objective function with the largest entropy rate on the graph, k connected subgraphs are generated, where k is the number of superpixels. SLIC is akin to a clustering method, initializing the superpixel center and conducting local iterations based on the similarity of surrounding pixels, ultimately obtaining the superpixel. However, SLIC ignores global image properties and is insensitive to pixel content [42]. Considering these factors, we opted to use ERS to obtain superpixels.
In our method, principal component analysis (PCA) is applied to the HSI, and the first principal components (PCs), I f , obtained are utilized for superpixel segmentation. This is because the first PCA contains the primary information of the HSI. By applying ERS to the I f , the HSI can be segmented into superpixel blocks:
I f = k S X k ,   s . t .   X K X g = ,   ( k g )
where S denotes the number of superpixels, and X k is the kth superpixel.

3.2. Superpixelwise Multiscale Adaptive T-HOSVD

In reality, determining the rank T-HOSVD is not a straightforward task. Therefore, we prefer to employ a simpler method to adaptively determine the rank- ( R ¯ 1 , R ¯ 2 , R ¯ 3 ) . During the SVD of a matrix, the larger singular values often contain more intrinsic information of the matrix. Based on this principle, we propose a simple strategy to automatically estimate the rank R ¯ n .
For HSI X I 1 × I 2 × I 3 , unfolding X obtains a matrix X n by n-mode. Let σ i be the singular values corresponding to matrix X n , arranged from large to small. Let C be the sum of singular values:
C = sum ( σ i ) ,   i = 1 , 2 I n
The rank R ¯ n can be determined by
R ¯ n = min ( i ) ,   sum ( σ k )   <   T n C ,   k = 1 , 2 i
where T is a given threshold, sum ( σ k ) ,   k = 1 , 2 i is the sum of the first i singular values of X n . Therefore, we can set the threshold- ( T 1 , T 2 , T 3 ) based on experience to adaptively determine the rank- ( R ¯ 1 , R ¯ 2 , R ¯ 3 ) .
For a given HSI X , we can set different thresholds T m ,   m = 1 , 2 , M ,where M is the number of different scales. For T-HOSVD of different scales, different structures and characteristics will be revealed [44]. Ranks of different sizes can reflect various levels of intrinsic information in HSI. Therefore, the rank of one scale often cannot entirely capture the information. It is necessary to integrate sufficiently complementary and relevant information. Multiscale fusion becomes an effective method.
Then, by superimposing the third dimension of X m , the result of Multiscale T-HOSVD can be determined by
Y ˜ = [ X 1 , X 2 , , X m ]   R I 1 × I 2 × ( I 3 × m )
The X m R I 1 × I 2 × I 3 ,   m = 1 , 2 , , M represents the T-HOSVD representation at various scales. However, simply fusing T-HOSVD representations of different scales together will inevitably have negative effects. On one hand, the contributions of different scales are treated equally, while the importance of different scales is ignored; on the other hand, stacking the third dimension will cause a dimensionality disaster, which may significantly impact the classification results [45].
In order to better extract information at different scales and enable multiscale fusion to better display the intrinsic structure and characteristics of HSI, we can perform PCA on Y ˜ to achieve dimensionality reduction:
y = y × 3 W T
where W is the transformation matrix composed of the eigenvectors corresponding to the first n eigenvalues of the covariance matrix Y 3 Y 3 T . Finally, we can obtain the result Y R I 1 × I 2 × n for the Multiscale Adaptive T-HOSVD.
By integrating HSI representations from adaptive T-HOSVD at multiple scales, complementary and relevant information can be effectively captured. Figure 2 illustrates the application of Multiscale Adaptive T-HOSVD to HSI. Among them, we use “Feature Image” color pictures to better represent the extracted features. Although HSI can capture a significant amount of useful information, its practical effectiveness is often limited due to the richness and complexity of the contained data. However, superpixel blocks derived from superpixel segmentation are highly homogeneous regions with similar structures and features. Therefore, applying Multiscale Adaptive T-HOSVD within these homogeneous regions tends to extract more effective features. We refer to this process as SmaT-HOSVD for superpixel blocks, as shown in Figure 3.
Since superpixels are typically irregular regions, preprocessing these irregularly shaped superpixels is necessary. For each superpixel, we need to identify the smallest bounding rectangle that encompasses the superpixel, which may include some adjacent pixels from other superpixels. All bands within this rectangular region are then extracted to form the 3D tensor required for Multiscale Adaptive T-HOSVD. This 3D tensor is processed using Multiscale Adaptive T-HOSVD, and the original superpixel is replaced with the resulting low-rank tensor.
This approach can effectively reduce intra-region differences while highlighting inter-class differences. Additionally, to the best of our knowledge, this is the first time an improved version of T-HOSVD has been applied to superpixel blocks. This method can adaptively estimate the optimal rank for different superpixel blocks and effectively utilize the spatial-spectral correlation of pixels in local images.

3.3. Classification

After feature extraction from the HSI to obtain the feature tensor, the next step is to feed it into the classifier for classification. Support vector machine (SVM) [46] is a classification algorithm whose basic idea is to map the data into a high-dimensional space and find a hyperplane in that space, maximizing the distance of all data points to the hyperplane. SVM has demonstrated good performance in the classification of remote sensing data [47]. Therefore, we decided to use SVM as our classifier.
In addition, the key of SVM is to choose the type of kernel function, mainly linear kernel, polynomial kernel, Gaussian radial basis function (RBF), and so on. Following the suggestion from [48], we decided to use RBF as our kernel function.
Algorithm 1 for the proposed method is shown below.
Algorithm 1: Superpixelwise Multiscale Adaptive T-HOSVD(SmaT-HOSVD)
Input:   Hyperspectral   image   X I 1 × I 2 × I 3 ,   ground   truth   label   L I 1 × I 2 , number of superpixels S, reduction dimensionality n, threshold T1 and T2.
1:
The first principal component X   is   obtained   by   applying   PCA   to   X is obtained by applying PCA to X .
2:
Apply Equation (7) to get superpixels I f
3:
While   X k   in   { I f } do
4:
                  Fill   X k   to   get   the   standard   tensor   X k
5:
                  Apply   Equations   ( 8 )   and   ( 9 )   to   compute   the   optimal   rank   of   X k
6:
                  Apply   Equations   ( 5 )   and   ( 6 )   for   X k   to   obtain   the   low - rank   representations   Y 1 and   Y 2
7:
                  Apply   Equation   ( 10 )   to   get   multiscale   representation   Y ˜ = [ Y 1 , Y 2 ]
8:
          Apply Equation (11) get results y
9:
                  Update   the   superpixel   X k   with   the   result   Y with the result Y
10:
         k ← k+1
11:
End while
12:
Updating   X I 1 × I 2 × n with superpixels I f
13:
Training   and   testing   of   X   and   L using SVM
OUTPUT: Classification results

4. Results

4.1. Datasets

In our experiments, we chose three publicly available hyperspectral datasets to evaluate the reliability of our model.
(1)
Indian Pines: This dataset was acquired by the imaging spectrometer AVRIS over the Indian Pines test field in northwest Indiana. It has a spatial size of 145 × 145 pixels and encompasses a total of 224 bands of reflectance data. The spectral reflectance bands range from 400 nm to 2500 nm. Approximately two-thirds of the hyperspectral scene is dedicated to agricultural areas, while one-third is predominantly covered by forest or vegetation. To focus on relevant information, bands that do not capture water reflections were excluded, resulting in the utilization of the remaining 200 bands for the study.
(2)
Pavia University: Collected by the ROSIS sensor during a flight over Pavia, northern Italy, this dataset comprises 610 × 240 pixels in 115 consecutive bands within the wavelength range of 430 nm to 860 nm. Noise affects 12 of the bands, leading to the selection of the remaining 103 bands for the study.
(3)
Salinas: Captured by the imaging spectrometer AVRIS which is manufactured by NASA’s Jet Propulsion Laboratory (JPL), this dataset represents an image of the Salinas Valley in California, USA. The image size is 512 × 217, and it includes 224 bands. Bands that do not capture water reflections were excluded, resulting in the selection of the remaining 204 bands for the study. The dataset includes areas with vegetables, bare soil, and vineyards.

4.2. Experimental Setup

We employed SVM classifiers to validate the effectiveness of SmaT-HOSVD. Following the methodology of previous work [49], we utilized five evaluation metrics: producer accuracy (PA), overall accuracy (OA), average accuracy (AA), kappa coefficient (Kappa), and running time (s). To mitigate systematic errors and reduce random errors, all experiments were independently conducted ten times, and the average values were computed. Both training and test sets were randomly selected without any repetition in either. The training set represented 2% (Indian Pines), 1% (Pavia University), and 1% (Salinas) of the categories in each sample, respectively.
To validate the effectiveness of the proposed method, we compared it with seven state-of-the-art HSI feature-extraction methods. Considering that SmaT-HOSVD utilizes superpixel segmentation and tensor low-rank representation, we selected two methods based on superpixel segmentation, i.e., SuperPCA and SpaSSA, and one tensor-based method, TensorSSA. In addition, due to the wide application of deep learning, we also consider two deep learning methods, i.e., SpectralFormer and SSTN. Moreover, we also selected T-HOSVD and SVM to highlight the effectiveness of SmaT-HOSVD.
(1)
A support vector machine (SVM) is employed as a classifier with a Gaussian radial basis function (RBF) serving as its kernel function. To determine the hyperparameters, we utilize five-fold cross-validation.
(2)
We utilize T-HOSVD for feature extraction and compare it with our proposed SmaT-HOSVD. The classifier employed is SVM, and the rank- ( R ¯ 1 , R ¯ 2 , R ¯ 3 ) is set to 60.
(3)
SpectralFormer [26] is a neural network based on transformer, utilizing a patch-wise version with a patch size of 1 × 3 and trained for 300 epochs.
(4)
SSTN is employed with a batch size of 32 and trained for 100 epochs, while keeping other hyperparameters consistent with the specifications in [27].
(5)
SuperPCA relies on dimensionality reduction through superpixel segmentation, employing the same parameters as specified in [22].
(6)
SpaSSA is grounded in feature extraction with superpixel segmentation, and the optimal parameters specified in [23] are utilized.
(7)
TensorSSA extracts spectral-spatial features through a 3D approach, and the optimal parameters specified in [20] are employed.
(8)
For our proposed method, the optimal parameters are detailed in IV.C. SVM is employed for classification.
The experiments were configured on MATLAB R2021a on a Windows 10 64-bit platform with an AMD Ryzen 7 5800H (3.2 GHz) CPU and 16 GB of memory which is manufactured by Advanced Micro Devices, Inc. (AMD). Deep learning implementation was carried out on the Ubuntu 20.04 platform using the PyTorch 1.10.2 framework with an NVIDIA GeForce RTX 3050Ti GPU.

4.3. Comparisons with State-of-the-Arts Models

Analyzing the classification results obtained using random selection and different training sets is crucial for understanding the impact of varying training data sizes on model performance. In this subsection, we conduct extensive experiments to compare the proposed method with state-of-the-art methods as detailed in Section 4.2.
(1)
Performance with Different Training Percentages: Figure 4 displays the overall accuracy (OA) achieved by different methods using various training percentages for the three datasets. Specifically, the randomly selected training sample percentages vary as follows: 1%, 2%, 3%, 4%, and 5% for the Indian Pines (IP) dataset; 0.5%, 1%, 2%, 3%, and 5% for the Pavia University (PU) dataset; and 0.2%, 0.4%, 0.5%, 1%, and 2% for the Salinas (SD) dataset.
Figure 4 illustrates that the accuracy of all methods generally improves as the percentage of training samples increases. Notably, the proposed SmaT-HOSVD consistently outperforms seven comparative methods, particularly when dealing with small sample sizes (e.g., 1% IP, 0.5% PU, and 0.2% SD), achieving excellent performance. Comparatively, state-of-the-art deep learning method SSTN and feature extraction method SuperPCA yield suboptimal results across the three datasets, falling short of SmaT-HOSVD, highlighting the effectiveness and advancement of SmaT-HOSVD. Additionally, other methods exhibit varying performance across different datasets. For instance, TensorSSA performs well as the second-best method on the PU dataset after SSTN and SmaT-HOSVD, but its classification performance on the IP and SD datasets is mediocre. In contrast, SmaT-HOSVD consistently achieves the best results across all three datasets, underscoring the robustness and efficacy of the proposed method.
(2)
Quantitative Evaluation: To thoroughly evaluate the effectiveness of the proposed method, we provide a detailed comparison of producer accuracy (PA) for each category in Table 1, Table 2 and Table 3, along with the three evaluation metrics OA, AA, and Kappa. The proposed SmaT-HOSVD consistently achieves the highest accuracy in most classes and outperforms other methods in terms of overall metrics. Taking the IP dataset in Table 1 as an example for analysis, firstly, in terms of OA, SmaT-HOSVD improves 27.83%, 27.78%, 33.08%, 4.24%, 1.81%, 14.35%, and 7.35% compared to SVM, T-HOSVD, SpectralFormer, SSTN, SuperPCA, SpaSSA, and TensorSSA, respectively. Also, on AA and Kappa, SmaT-HOSVD improved to varying degrees. It is worth noting that the PA of SuperPCA is not much different from SmaT-HOSVD, but the processing time of SuperPCA is shorter. This is because during the processing of SmaT-HOSVD, too many neighboring pixels are added to some small superpixel blocks, which introduces noise and results in a degradation of the effect. And compared to SuperPCA, the time complexity of SmaT-HOSVD is higher, but overall, from all three datasets, SmaT-HOSVD achieves better performance. On the PA of each class, SmaT-HOSVD s’ accuracy is basically at the top of the list, above 80% for 5 classes and above 90% for 11 classes. The superior performance of the SmaT-HOSVD method is also observed in Table 2 and Table 3 for the Pavia University and Salinas datasets, respectively.
When comparing methods, both SuperPCA and SpaSSA demonstrate higher classification accuracy compared to the classical SVM. This improvement can be attributed to their utilization of local features, with SuperPCA exhibiting particularly notable performance in this regard. Unfortunately, SuperPCA does not perform well on the PU dataset, possibly due to the limitations of PCA-based feature extraction, especially for small targets. The deep learning method SpectralFormer is less effective than SVM, primarily because of the limited training samples that prevent adequate model training. In contrast, SSTN achieves better results by leveraging spectral-spatial features more comprehensively. The tensor-based method TensorSSA effectively extracts low-rank information but does not effectively utilize important local features, resulting in moderate performance. In contrast, SmaT-HOSVD demonstrates the best performance across all datasets by effectively extracting both local and spectral-space features.
(3)
Qualitative Evaluation: A diagram categorizing IP, PU, and SD data is illustrated in Figure 5, Figure 6 and Figure 7. The analysis primarily focuses on the IP dataset presented in Figure 5. Once again, the emphasis is on the Indian Pines (IP) dataset depicted in Figure 5. SVM and T-HOSVD exhibit significant categorization noise across all categories, characterized by an exceptionally high number of classification errors. The performance of SpectralFormer is unsatisfactory, showing high accuracy only in a few classes while remaining extremely noisy in most classes. This outcome is likely attributed to limited training samples and underfitting of the model. SSTN is significantly more effective, reducing speckle-like classification errors, but still exhibits misclassifications in certain areas. SuperPCA achieves a high accuracy rate primarily because superpixel blocks provide localized features and enhance intra-class similarity. However, some classes are still affected by contamination from speckles. Both SpaSSA and TensorSSA also exhibit significant classification noise.
Our extracted SmaT-HOSVD is able to efficiently extract both spatial and spectral features thanks to the superpixel block and low-rank representation. As shown in Figure 5j, Figure 6j and Figure 7j, good classification results are basically obtained for targets in large or regular categories. This is due to the fact that SmaT-HOSVD effectively utilizes the spatial information through the superpixel blocks and obtains the main spectral information through the low-rank representation, which enhances the intra-class consistency and highlights the inter-class differences. However, the classification results are poor in some class edges and irregular mini-classes, which is due to the fact that too many pixels from other classes are added to introduce noise during processing, making the extracted features not characterize the intra-class information well. Intuitive analysis shows that the SmaT-HOSVD method can effectively reduce the noise and improve the classification accuracy compared with other methods.
(4)
Analysis of Running Time: Table 1, Table 2 and Table 3 likewise give the running times for all the methods. From the table, it can be seen that T-HOSVD, SuperPCA, and SmaT-HOSVD are the fastest and basically belong to the same order of magnitude. The training time for SpectralFormer and SSTN is prolonged due to the model’s complexity and the high number of training iterations, even when utilizing GPUs for training. SpaSSA, on the other hand, exhibits a longer runtime because the feature extraction needs to be carried out individually for each band, resulting in significant time consumption. TensorSSA requires performing the decomposition of the trajectory tensor, which is a time-consuming process. In comparison to SuperPCA and T-HOSVD, SmaT-HOSVD requires more time due to the additional tasks involved. However, despite this, SmaT-HOSVD is relatively efficient and performs well in terms of computational efficiency compared to many other methods.
Overall, the proposed SmaT-HOSVD method strikes a balance between runtime efficiency and classification accuracy, making it a viable choice for HSI classification.

5. Discussion

5.1. Parameter Analysis

In the proposed SmaT-HOSVD method, certain parameters require manual configuration. These parameters include the number of superpixels (S) in ERS-based superpixel segmentation, the dimension (n) for PCA dimensionality reduction, and the threshold (T) in T-HOSVD. To evaluate the impact of these parameters and determine optimal values, we conducted parameter sensitivity experiments on SmaT-HOSVD.
(1)
Number of superpixels (S): Through superpixel segmentation, we can obtain different homogeneous regions where similar pixels are more likely to fall into the same class [22]. The parameter S determines the level of segmentation for the superpixels. A larger S leads to finer segmentation with larger superpixels, whereas a smaller S results in coarser segmentation with smaller superpixels. The degree of superpixel segmentation significantly influences the extracted features. Due to variations in spatial resolution and dataset size, the optimal number of superpixels (S) differs for each dataset. Although ideal classification accuracy requires over-segmentation of the scene, due to the complexity of the distribution of land cover, specific analysis is still required [50]. As shown in Figure 8, the best results are taken at 30, 25, and 90 for S in our approach for Indian Pines, Pavia University, and Salinas datasets, respectively. For example, in the IP dataset, a number of superpixel S below 50 achieves good results, which is due to the fact that the homogeneous regions in the IP dataset are around 30. When S is too large, it makes the image over-segmented, resulting in the obtained superpixel blocks being fragmented and narrow, making the algorithm ineffective.
(2)
Reduction dimensionality (n): The dimensionality reduction n directly influences the amount of information retained in the data. A larger value of n generally retains more information but may also increase noise; conversely, a smaller n emphasizes the main components of information, potentially leading to the loss of important details. Because of the diversity in data, the optimal dimensionality reduction n varies for each dataset. The ideal reduction n effectively filters out noise while preserving essential information. As shown in Figure 9, in our approach, the Indian Pines, University of Pavia, and Salinas datasets work best when the reduced dimension n is set to 20, 20, and 15, respectively. Based on the downscaling results of different datasets, our algorithm achieves the best classification performance in the range of downscaling dimensions from about 15 to 25. This shows that for the obtained low-rank tensor, the dimension reduction is usually around 20 to leave the most dominant information and remove the noise to extract the better features, which is consistent with the conclusion in [45]. When the downscaling dimension is larger, the information retained is cumbersome and contains noise, making the effect decrease. At the same time, a downscaling dimension of around 20 on different datasets achieves better results, which further illustrates the stability of our algorithm.
(3)
Threshold T1 and T2: The choice of different thresholds (T1 and T2) significantly affects the extracted features. The larger threshold- ( T 1 , T 2 , T 3 ) results in a higher rank- ( R ¯ 1 , R ¯ 2 , R ¯ 3 ) , preserving more common features of the HSI. The level of low-rank features extracted varies with different T. As T increases, more dimensions are retained in the tensor decomposition, resulting in the preservation of a larger amount of information. This leads to extracted features that tend to be richer; however, they may also contain some noise. On the contrary, as T decreases, the tensor decomposition retains fewer dimensions, and thus some details and features may be lost. Furthermore, to simplify parameter settings, both T1 and T2 are set to the same value, while T3 is taken relatively independently. And in order to obtain different levels of features, T2 is generally set higher than T1, with T2 as a larger threshold and T1 as a smaller one. In T-HOSVD, the sum of the top 10% or even 1% of the singular values accounted for more than 99% of the sum of all singular values. Based on this principle, we set T1 as [(0.8,0.8,0.8), (0.85,0.85,0.9), (0.8,0.8,0.9)], and T2 as [(0.8,0.8,0.95), (0.85,0.85,0.95), (0.9,0.9,0.95)]. The detailed results of parameter tuning are illustrated in Figure 10.
From Figure 10, it is evident that the larger threshold value of T2 has a more significant impact on the proposed method and is a key parameter, while T1 serves as an auxiliary parameter with less impact. For different datasets, the optimal parameters for T1 and T2 are different. In our method, T1 for the Indian Pines dataset is set to (0.8,0.8,0.9), and T2 is set to (0.85,0.85,0.95); for the Pavia University dataset, T1 is set to (0.8,0.8,0.8), and T2 is set to (0.8,0.8,0.95); and for the Salinas dataset, T1 is set as (0.8,0.8,0.8), and T2 is set as (0.9,0.9,0.95). Our proposed method exhibits robustness to thresholding, as evidenced by the minimal variation in OA on Indian Pines, which is only 0.018. This result underscores the idea that different thresholds can offer complementary structural information.
Determining the number of superpixels and the optimal rank is a critical and open question in the proposed method, which aims to segment HSI and perform low-rank representation. In this paper, we experimentally set the number of superpixels and determined the optimal rank using an adaptive estimation strategy to achieve the best performance. The segmentation scale of HSI varies in terms of the number of homogeneous regions across different datasets. Therefore, in our experiments, we determined the optimal number of superpixel blocks using a continuous adjustment method to adapt to the specific characteristics of each dataset. In fact, we can determine the rank based on the percentage of total energy, as implemented in (8) and (9). The process of determining the optimal rank can be likened to setting a threshold for a given HSI. Given that the initial singular values in singular value decomposition typically contain the majority of the energy, we can fine-tune the threshold to identify the optimal rank. Thus, the above method can determine the number of superpixels and the optimal rank for different datasets.

5.2. Ablation Study

To compare the effectiveness of the proposed SmaT-HOSVD against T-HOSVD, we incorporate two key enhancements: ERS for superpixel segmentation and a multiscale adaptive low-rank representation strategy for T-HOSVD. We aim to validate these operations across three datasets. In the pixel-processing stage, we evaluate two scenarios: using the raw HSI without superpixel segmentation versus using ERS for superpixel segmentation. For the low-rank representation stage, we compare the original T-HOSVD with T-HOSVD employing the multiscale adaptive strategy. Specifically, we set rank- ( R ¯ 1 , R ¯ 2 , R ¯ 3 ) to 60 for the original T-HOSVD. However, due to varying sizes and shapes of superpixel blocks, smaller blocks may not achieve the length dimension of 60. Therefore, for the low-rank representation of superpixel blocks, we set rank- ( R ¯ 1 , R ¯ 2 , R ¯ 3 ) to 10. In our experiments, we select 2% of samples from the IP, 1% of samples from the PU, and 1% of samples from the SD datasets for training.
As shown in Table 4, the combination of superpixel segmentation and multiscale adaptive T-HOSVD (SmaT-HOSVD) achieves the highest classification performance. Comparing the raw HSI with superpixel blocks after using T-HOSVD separately shows that the latter has a higher classification accuracy across all three datasets. This improvement can be attributed to the tendency of superpixel blocks to contain similar or identical class targets, enabling low-rank representation to mitigate noise interference and enhance intra-class similarity. This makes the extracted features more different from one class to another, highlighting the gaps between different classes and facilitating the subsequent classification task. Regarding the enhanced T-HOSVD with adaptive rank estimation and multiscale fusion strategy, the classification accuracies are significantly improved compared with the original T-HOSVD. This is due to the fact that adaptive rank estimation can dynamically adjust the optimal rank according to the target information distribution, while enhancing the classification task by fusing features from different scales to obtain a more comprehensive representation. Moreover, this adaptive rank estimation and multiscale fusion does not add too much computation and has high efficiency.
As depicted in Table 4, Superpixelwise Multiscale Adaptive T-HOSVD outperforms traditional T-HOSVD across the three datasets, showcasing improved classification performance. The combination of superpixel segmentation and low-rank representation can better remove HSI noise and extract effective features [51,52]. The information contained in the superpixel blocks obtained by superpixel segmentation varies in shape and size, and a fixed rank setting may not accurately reflect the information distribution of each superpixel block. Therefore, the use of adaptive rank estimation and multiscale fusion is particularly favorable for the low-rank representation of superpixel blocks, which enables the extracted features to reflect the information of each homogeneous region, thus enhancing the effectiveness of the classification task. Overall, leveraging superpixel blocks and multiscale adaptive T-HOSVD (SmaT-HOSVD) estimation yields superior performance in HSI classification.

5.3. Analysis between SmaT-HOSVD and Classification

In this paper, we utilize a combination of SmaT-HOSVD and SVM classifiers to achieve ground object classification. SmaT-HOSVD is employed for 3D feature extraction to obtain spectral-spatial features, aiming to reduce inter-class noise and enhance intra-class consistency for improved classification effectiveness. SmaT-HOSVD operates as an unsupervised 3D feature extraction method on HSI superpixel blocks without the need for labeled training data. This method effectively enhances inter-class differences and promotes inter-class consistency, thereby enhancing classification performance. Whereas supervised feature-extraction methods have been shown to further improve classification performance [53,54], and the sample prior information included in SmaT-HOSVD is also worth exploring.
However, there is a potential risk associated with SmaT-HOSVD. It may inadvertently diminish differences between similar objects while increasing agreement between dissimilar objects, which can introduce noise and lead to misclassification. The superpixel blocks obtained from superpixel segmentation may contain a higher number of different objects, and the introduction of noise in the algorithm can unintentionally reduce the differences between similar objects while increasing the consistency between dissimilar objects, leading to errors in classification. This requires us to avoid the superpixel blocks containing too much noise by finding a more appropriate superpixel segmentation method or designing an adaptive superpixel number-estimation algorithm. Meanwhile, for narrow and irregular superpixel blocks, dividing them into regular tensors may contain too many dissimilar pixels, resulting in the extracted features containing noise and leading to poor classification. In the subsequent improvement, we will seek an adaptive division of superpixels strategy instead of simple division to avoid excessive noise leading to the degradation of results.

6. Conclusions

HSIs, with their inherent three-dimensional nature, possess both spectral and spatial characteristics. Traditional methods often focus on either spatial or spectral features, or they attempt a simple fusion of the two. Consequently, there is a growing trend towards the utilization of 3D or tensor methods in HSI classification. However, these approaches frequently concentrate on global features and may overlook homogeneous regions, which could provide more effective spatial and spectral features. To address these limitations, this paper introduces the SmaT-HOSVD method for 3D feature extraction in HSI.
The proposed SmaT-HOSVD method classifies HSI into homogeneous clusters through superpixel segmentation to obtain superpixel blocks. It then performs adaptive rank estimation and multiscale fusion on these superpixel blocks to acquire low-order representations for each homogeneous cluster, facilitating 3D feature extraction of qualitative areas. SmaT-HOSVD effectively utilizes local features and low-order representations to capture the low-rank intrinsic features of HSI, thereby reducing redundancy and noise. This enhances the similarity within similar classes and widens the gap between different classes. By extracting features from HSI in 3D, SmaT-HOSVD obtains spectral-spatial information that facilitates subsequent HSI classification tasks. SmaT-HOSVD surpasses several state-of-the-art methods on three public datasets with an impressive overall accuracy improvement of 12% to 20%. Moreover, even with very little training, SmaT-HOSVD can achieve good overall accuracy while maintaining low time complexity, which has great potential in practical applications.
In future work, we plan to explore enhanced methods for incorporating irregular superpixels into HOSVD. This involves avoiding the use of excessive noise pixels to fill irregular superpixels, ensuring appropriate handling of irregularities. In addition, in the deep-sea manganese nodule classification task, due to the spectral spatial changes of HSI and the instability of illumination in the deep sea [55], we will also study the application of SmaT-HOSVD in the deep-sea manganese nodule classification task to improve the classification accuracy.

Author Contributions

Conceptualization, Q.D.; methodology, Q.D.; software, Q.D.; validation, Q.D. and C.M.; formal analysis, Q.D.; investigation, C.M.; resources, Q.D.; data curation, C.M.; writing—original draft preparation, Q.D.; writing—review and editing, Q.Z.; visualization, Q.D.; supervision, Q.Z.; project administration, Q.Z.; funding acquisition, Q.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Landgrebe, D. Hyperspectral image data analysis. IEEE Signal Process. Mag. 2002, 19, 17–28. [Google Scholar] [CrossRef]
  2. Deng, B.; Jia, S.; Shi, D. Deep Metric Learning-Based Feature Embedding for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2020, 58, 1422–1435. [Google Scholar] [CrossRef]
  3. Zhao, C.; Li, X.; Ren, J.; Marshall, S. Improved sparse representation using adaptive spatial support for effective target detection in hyperspectral imagery. Int. J. Remote Sens. 2013, 34, 8669–8684. [Google Scholar] [CrossRef]
  4. Luo, Y.; Zou, J.; Yao, C.; Zhao, X.; Li, T.; Bai, G. HSI-CNN: A Novel Convolution Neural Network for Hyperspectral Image. In Proceedings of the 2018 International Conference on Audio, Language and Image Processing (ICALIP), Shanghai, China, 16–17 July 2018; pp. 464–469. [Google Scholar]
  5. Carpenter, M.H.; Croce, M.P.; Baker, Z.K.; Batista, E.R.; Caffrey, M.P.; Fontes, C.J.; Koehler, K.E.; Kossmann, S.E.; McIntosh, K.G.; Rabin, M.W.; et al. Hyperspectral X-ray Imaging with TES Detectors for Nanoscale Chemical Speciation Mapping. J. Low Temp. Phys. 2020, 200, 437–444. [Google Scholar] [CrossRef]
  6. Al Ktash, M.; Stefanakis, M.; Englert, T.; Drechsel, M.S.L.; Stiedl, J.; Green, S.; Jacob, T.; Boldrini, B.; Ostertag, E.; Rebner, K.; et al. UV Hyperspectral Imaging as Process Analytical Tool for the Characterization of Oxide Layers and Copper States on Direct Bonded Copper. Sensors 2021, 21, 7332. [Google Scholar] [CrossRef] [PubMed]
  7. Batshev, V.I.; Krioukov, A.V.; Machikhin, A.S.; Zolotukhina, A.A. Multispectral video camera optical system. J. Opt. Technol. 2023, 90, 706–712. [Google Scholar] [CrossRef]
  8. Adesokan, M.; Alamu, E.O.; Otegbayo, B.; Maziya-Dixon, B. A Review of the Use of Near-Infrared Hyperspectral Imaging (NIR-HSI) Techniques for the Non-Destructive Quality Assessment of Root and Tuber Crops. Appl. Sci. 2023, 13, 5226. [Google Scholar] [CrossRef]
  9. Kulya, M.; Petrov, N.V.; Tsypkin, A.; Egiazarian, K.; Katkovnik, V. Hyperspectral data denoising for terahertz pulse time-domain holography. Opt. Express 2019, 27, 18456–18476. [Google Scholar] [CrossRef] [PubMed]
  10. Fabiyi, S.D.; Vu, H.; Tachtatzis, C.; Murray, P.; Harle, D.; Dao, T.K.; Andonovic, I.; Ren, J.; Marshall, S. Varietal Classification of Rice Seeds Using RGB and Hyperspectral Images. IEEE Access 2020, 8, 22493–22505. [Google Scholar] [CrossRef]
  11. Cloutis, E.A. Review Article Hyperspectral geological remote sensing: Evaluation of analytical techniques. Int. J. Remote Sens. 1996, 17, 2215–2242. [Google Scholar] [CrossRef]
  12. Cochrane, M.A. Using vegetation reflectance variability for species level classification of hyperspectral data. Int. J. Remote Sens. 2000, 21, 2075–2087. [Google Scholar] [CrossRef]
  13. Murphy, R.J.; Monteiro, S.T.; Schneider, S. Evaluating Classification Techniques for Mapping Vertical Geology Using Field-Based Hyperspectral Sensors. IEEE Trans. Geosci. Remote Sens. 2012, 50, 3066–3080. [Google Scholar] [CrossRef]
  14. Melgani, F.; Bruzzone, L. Classification of hyperspectral remote sensing images with support vector machines. IEEE Trans. Geosci. Remote Sens. 2004, 42, 1778–1790. [Google Scholar] [CrossRef]
  15. Qiao, T.; Ren, J.; Wang, Z.; Zabalza, J.; Sun, M.; Zhao, H.; Li, S.; Benediktsson, J.A.; Dai, Q.; Marshall, S. Effective Denoising and Classification of Hyperspectral Images Using Curvelet Transform and Singular Spectrum Analysis. IEEE Trans. Geosci. Remote Sens. 2017, 55, 119–133. [Google Scholar] [CrossRef]
  16. Zabalza, J.; Ren, J.; Yang, M.; Zhang, Y.; Wang, J.; Marshall, S.; Han, J. Novel Folded-PCA for improved feature extraction and data reduction with hyperspectral imaging and SAR in remote sensing. ISPRS J. Photogramm. Remote Sens. 2014, 93, 112–122. [Google Scholar] [CrossRef]
  17. Bandos, T.V.; Bruzzone, L.; Camps-Valls, G. Classification of Hyperspectral Images with Regularized Linear Discriminant Analysis. IEEE Trans. Geosci. Remote Sens. 2009, 47, 862–873. [Google Scholar] [CrossRef]
  18. Hang, R.; Liu, Q.; Sun, Y.; Yuan, X.; Pei, H.; Plaza, J.; Plaza, A. Robust Matrix Discriminative Analysis for Feature Extraction From Hyperspectral Images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2017, 10, 2002–2011. [Google Scholar] [CrossRef]
  19. He, X.; Niyogi, P. Locality Preserving Projections. Adv. Neural Inf. Process. Syst. 2003, 16. [Google Scholar]
  20. Fu, H.; Sun, G.; Zhang, A.; Shao, B.; Ren, J.; Jia, X. Tensor Singular Spectrum Analysis for 3-D Feature Extraction in Hyperspectral Images. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5403914. [Google Scholar] [CrossRef]
  21. Imani, M.; Ghassemian, H. An overview on spectral and spatial information fusion for hyperspectral image classification: Current trends and challenges. Inf. Fusion 2020, 59, 59–83. [Google Scholar] [CrossRef]
  22. Jiang, J.; Ma, J.; Chen, C.; Wang, Z.; Cai, Z.; Wang, L. SuperPCA: A Superpixelwise PCA Approach for Unsupervised Feature Extraction of Hyperspectral Imagery. IEEE Trans. Geosci. Remote Sens. 2018, 56, 4581–4593. [Google Scholar] [CrossRef]
  23. Sun, G.; Fu, H.; Ren, J.; Zhang, A.; Zabalza, J.; Jia, X.; Zhao, H. SpaSSA: Superpixelwise Adaptive SSA for Unsupervised Spatial–Spectral Feature Extraction in Hyperspectral Image. IEEE Trans. Cybern. 2022, 52, 6158–6169. [Google Scholar] [CrossRef] [PubMed]
  24. Paoletti, M.E.; Haut, J.M.; Plaza, J.; Plaza, A. Deep learning classifiers for hyperspectral imaging: A review. ISPRS J. Photogramm. Remote Sens. 2019, 158, 279–317. [Google Scholar] [CrossRef]
  25. Li, Y.; Zhang, H.; Shen, Q. Spectral–Spatial Classification of Hyperspectral Imagery with 3D Convolutional Neural Network. Remote Sens. 2017, 9, 67. [Google Scholar] [CrossRef]
  26. Hong, D.; Han, Z.; Yao, J.; Gao, L.; Zhang, B.; Plaza, A.; Chanussot, J. SpectralFormer: Rethinking Hyperspectral Image Classification with Transformers. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5518615. [Google Scholar] [CrossRef]
  27. Zhong, Z.; Li, Y.; Ma, L.; Li, J.; Zheng, W.S. Spectral–Spatial Transformer Network for Hyperspectral Image Classification: A Factorized Architecture Search Framework. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5514715. [Google Scholar] [CrossRef]
  28. Li, Q.; Schonfeld, D. Multilinear Discriminant Analysis for Higher-Order Tensor Data Classification. IEEE Trans. Pattern Anal. Mach. Intell. 2014, 36, 2524–2537. [Google Scholar] [CrossRef] [PubMed]
  29. Peng, J.; Sun, W.; Li, H.C.; Li, W.; Meng, X.; Ge, C.; Du, Q. Low-Rank and Sparse Representation for Hyperspectral Image Processing: A review. IEEE Geosci. Remote Sens. Mag. 2022, 10, 10–43. [Google Scholar] [CrossRef]
  30. De Lathauwer, L.; De Moor, B.; Vandewalle, J. A Multilinear Singular Value Decomposition. SIAM J. Matrix Anal. Appl. 2000, 21, 1253–1278. [Google Scholar] [CrossRef]
  31. Bernardi, A.; Iannacito, M.; Rocchini, D. High order singular value decomposition for plant diversity estimation. Boll. Dell’unione Mat. Ital. 2021, 14, 557–591. [Google Scholar] [CrossRef]
  32. An, J.; Zhang, X.; Zhou, H.; Jiao, L. Tensor-Based Low-Rank Graph with Multimanifold Regularization for Dimensionality Reduction of Hyperspectral Images. IEEE Trans. Geosci. Remote Sens. 2018, 56, 4731–4746. [Google Scholar] [CrossRef]
  33. Zhang, Z.; Aeron, S. Exact Tensor Completion Using t-SVD. IEEE Trans. Signal Process. 2017, 65, 1511–1526. [Google Scholar] [CrossRef]
  34. Quijorna, E.P. Nuclear instruments and methods in physics research section B: Beam interactions with materials and atoms. Nucl. Instrum. Methods Phys. Res. B 2024, 549, 165287. [Google Scholar] [CrossRef]
  35. Wang, Q.; Kang, W. Adaptive tensor compressive sensing based on noise estimation: Application in three-dimensional images. Multimed. Tools Appl. 2017, 76, 6175–6188. [Google Scholar] [CrossRef]
  36. Bergqvist, G.; Larsson, E.G. The Higher-Order Singular Value Decomposition: Theory and an Application [Lecture Notes]. IEEE Signal Process. Mag. 2010, 27, 151–154. [Google Scholar] [CrossRef]
  37. Kolda, T.G.; Bader, B.W. Tensor Decompositions and Applications. SIAM Rev. 2009, 51, 455–500. [Google Scholar] [CrossRef]
  38. De Lathauwer, L.; Vandewalle, J. Dimensionality reduction in higher-order signal processing and rank-(R1,R2,…,RN) reduction in multilinear algebra. Linear Algebra Its Appl. 2004, 391, 31–55. [Google Scholar] [CrossRef]
  39. Vasilescu, M.A.O.; Terzopoulos, D. Multilinear Analysis of Image Ensembles: TensorFaces. In Computer Vision—ECCV 2002, Proceedings of the 7th European Conference on Computer Vision, Copenhagen, Denmark, 28–31 May 2002; Springer: Berlin/Heidelberg, Germany, 2002; pp. 447–460. [Google Scholar]
  40. Ahmadi-Asl, S.; Abukhovich, S.; Asante-Mensah, M.G.; Cichocki, A.; Phan, A.H.; Tanaka, T.; Oseledets, I. Randomized Algorithms for Computation of Tucker Decomposition and Higher Order SVD (HOSVD). IEEE Access 2021, 9, 28684–28706. [Google Scholar] [CrossRef]
  41. Achanta, R.; Shaji, A.; Smith, K.; Lucchi, A.; Fua, P.; Süsstrunk, S. SLIC Superpixels Compared to State-of-the-Art Superpixel Methods. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 34, 2274–2282. [Google Scholar] [CrossRef]
  42. Wang, M.; Liu, X.; Gao, Y.; Ma, X.; Soomro, N.Q. Superpixel segmentation: A benchmark. Signal Process. Image Commun. 2017, 56, 28–39. [Google Scholar] [CrossRef]
  43. Liu, M.Y.; Tuzel, O.; Ramalingam, S.; Chellappa, R. Entropy rate superpixel segmentation. In Proceedings of the CVPR 2011, Colorado Springs, CO, USA, 20–25 June 2011; pp. 2097–2104. [Google Scholar]
  44. Fang, L.; Li, S.; Kang, X.; Benediktsson, J.A. Spectral–Spatial Hyperspectral Image Classification via Multiscale Adaptive Sparse Representation. IEEE Trans. Geosci. Remote Sens. 2014, 52, 7738–7749. [Google Scholar] [CrossRef]
  45. An, J.; Lei, J.; Song, Y.; Zhang, X.; Guo, J. Tensor Based Multiscale Low Rank Decomposition for Hyperspectral Images Dimensionality Reduction. Remote Sens. 2019, 11, 1485. [Google Scholar] [CrossRef]
  46. Cherkassky, V.; Ma, Y. Practical selection of SVM parameters and noise estimation for SVM regression. Neural Netw. 2004, 17, 113–126. [Google Scholar] [CrossRef] [PubMed]
  47. Pal, M.; Foody, G.M. Feature Selection for Classification of Hyperspectral Data by SVM. IEEE Trans. Geosci. Remote Sens. 2010, 48, 2297–2307. [Google Scholar] [CrossRef]
  48. Tarabalka, Y.; Fauvel, M.; Chanussot, J.; Benediktsson, J.A. SVM- and MRF-Based Method for Accurate Classification of Hyperspectral Images. IEEE Geosci. Remote Sens. Lett. 2010, 7, 736–740. [Google Scholar] [CrossRef]
  49. Guo, X.; Huang, X.; Zhang, L.; Zhang, L. Hyperspectral image noise reduction based on rank-1 tensor decomposition. ISPRS J. Photogramm. Remote Sens. 2013, 83, 50–63. [Google Scholar] [CrossRef]
  50. Li, J.; Zhang, H.; Zhang, L. Efficient Superpixel-Level Multitask Joint Sparse Representation for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2015, 53, 5338–5351. [Google Scholar] [CrossRef]
  51. Fan, F.; Ma, Y.; Li, C.; Mei, X.; Huang, J.; Ma, J. Hyperspectral image denoising with superpixel segmentation and low-rank representation. Inf. Sci. 2017, 397–398, 48–68. [Google Scholar] [CrossRef]
  52. Yang, S.; Hou, J.; Jia, Y.; Mei, S.; Du, Q. Superpixel-Guided Discriminative Low-Rank Representation of Hyperspectral Images for Classification. IEEE Trans. Image Process. 2021, 30, 8823–8835. [Google Scholar] [CrossRef]
  53. Huang, H.; Shi, G.; He, H.; Duan, Y.; Luo, F. Dimensionality Reduction of Hyperspectral Imagery Based on Spatial–Spectral Manifold Learning. IEEE Trans. Cybern. 2020, 50, 2604–2616. [Google Scholar] [CrossRef]
  54. Luo, F.; Zou, Z.; Liu, J.; Lin, Z. Dimensionality Reduction and Classification of Hyperspectral Image via Multistructure Unified Discriminative Embedding. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5517916. [Google Scholar] [CrossRef]
  55. Ye, P.; Han, C.; Zhang, Q.; Gao, F.; Yang, Z.; Wu, G. An Application of Hyperspectral Image Clustering Based on Texture-Aware Superpixel Technique in Deep Sea. Remote Sens. 2022, 14, 5047. [Google Scholar] [CrossRef]
Figure 1. Schematic of the proposed SmaT-HOSVD-based HSI classification.
Figure 1. Schematic of the proposed SmaT-HOSVD-based HSI classification.
Sensors 24 04072 g001
Figure 2. Schematic diagram of Multiscale Adaptive T-HOSVD usage on HSI Global.
Figure 2. Schematic diagram of Multiscale Adaptive T-HOSVD usage on HSI Global.
Sensors 24 04072 g002
Figure 3. Schematic diagram of SmaT-HOSVD on HSI.
Figure 3. Schematic diagram of SmaT-HOSVD on HSI.
Sensors 24 04072 g003
Figure 4. OA obtained by different methods with different training percentages over (a) IP, (b) PU, and (c) SD datasets.
Figure 4. OA obtained by different methods with different training percentages over (a) IP, (b) PU, and (c) SD datasets.
Sensors 24 04072 g004
Figure 5. Classification results for the Indian Pines dataset (10% training samples): (a) Original; (b) Ground Truth; (c) SVM (66.09); (d) T-HOSVD (65.53); (e) SpectralFormer (60.23); (f) SSTN (89.79); (g) SuperPCA (90.9); (h) SpaSSA (75.14); (i) TensorSSA (85.22); (j) Ours (93.31).
Figure 5. Classification results for the Indian Pines dataset (10% training samples): (a) Original; (b) Ground Truth; (c) SVM (66.09); (d) T-HOSVD (65.53); (e) SpectralFormer (60.23); (f) SSTN (89.79); (g) SuperPCA (90.9); (h) SpaSSA (75.14); (i) TensorSSA (85.22); (j) Ours (93.31).
Sensors 24 04072 g005
Figure 6. Classification results for the Pavia University dataset (1% training samples): (a) Origin; (b) Ground Truth; (c) SVM (88.81); (d) T-HOSVD (89.33); (e) SpectralFormer (77.94); (f) SSTN (95.12); (g) SuperPCA (85.34); (h) SpaSSA (93.61); (i) TensorSSA (94.78); (j) Ours (97.34).
Figure 6. Classification results for the Pavia University dataset (1% training samples): (a) Origin; (b) Ground Truth; (c) SVM (88.81); (d) T-HOSVD (89.33); (e) SpectralFormer (77.94); (f) SSTN (95.12); (g) SuperPCA (85.34); (h) SpaSSA (93.61); (i) TensorSSA (94.78); (j) Ours (97.34).
Sensors 24 04072 g006
Figure 7. Classification results for the Salinas dataset (1% training samples): (a) Original; (b) Ground Truth; (c) SVM (86.47); (d) T-HOSVD (85.54); (e) SpectralFormer (86.55); (f) SSTN (97.66); (g) SuperPCA (98.89); (h) SpaSSA (97.06); (i) TensorSSA (90.00); (j) Ours (99.24).
Figure 7. Classification results for the Salinas dataset (1% training samples): (a) Original; (b) Ground Truth; (c) SVM (86.47); (d) T-HOSVD (85.54); (e) SpectralFormer (86.55); (f) SSTN (97.66); (g) SuperPCA (98.89); (h) SpaSSA (97.06); (i) TensorSSA (90.00); (j) Ours (99.24).
Sensors 24 04072 g007aSensors 24 04072 g007b
Figure 8. The influence of parameter S on overall accuracy (OA) concerning (a) IP with 2% training, (b) PU with 1% training, and (c) SD with 1% training.
Figure 8. The influence of parameter S on overall accuracy (OA) concerning (a) IP with 2% training, (b) PU with 1% training, and (c) SD with 1% training.
Sensors 24 04072 g008
Figure 9. The influence of reduction dimensionality n on overall accuracy (OA) concerning (a) IP with 2% training, (b) PU with 1% training, and (c) SD with 1% training.
Figure 9. The influence of reduction dimensionality n on overall accuracy (OA) concerning (a) IP with 2% training, (b) PU with 1% training, and (c) SD with 1% training.
Sensors 24 04072 g009
Figure 10. The influence of parameter T1 and T2 on overall accuracy (OA) concerning (a) IP with 2% training, (b) PU with 1% training, and (c) SD with 1% training.
Figure 10. The influence of parameter T1 and T2 on overall accuracy (OA) concerning (a) IP with 2% training, (b) PU with 1% training, and (c) SD with 1% training.
Sensors 24 04072 g010
Table 1. Classification results obtained by different methods for the IP dataset (2% training percentage).
Table 1. Classification results obtained by different methods for the IP dataset (2% training percentage).
ClassSamplesSvmT-hosvdSpectralformerSstnSuperpcaSpassaTensorssaOurs
1465.338.880.082.61100.038.6789.1180.00
2142857.1368.4051.7593.9890.2675.1681.4987.59
383045.8738.6238.5091.0985.0074.6183.9284.76
423718.3221.5533.6290.2961.2581.1667.9880.52
548369.2866.3820.0882.1991.8676.5385.1292.01
673084.7781.5389.3793.0196.2490.0494.8099.89
72827.0459.250.0100.096.374.4495.1995.93
847894.9287.3999.78100.096.9490.5693.27100.0
92024.7431.570.0100.095.2628.9599.4784.74
1097256.1744.4338.4489.8188.0574.1481.8290.50
11245574.8277.9674.5180.9395.6779.7887.5595.40
1259332.7527.5316.7087.5281.5156.9969.5090.65
1320591.4590.5066.5100.099.579.6595.599.55
14126589.9993.9497.3498.1897.2792.7197.3799.98
1538625.9093.9415.6179.0294.3769.7974.8995.27
169349.7839.5683.5249.4654.2986.8185.3896.70
OA65.4865.5360.2389.0791.5078.9685.9693.31
AA53.0253.7645.3688.6388.9973.1286.4092.09
Kappa60.1860.0753.7787.5990.3076.0184.0192.37
Time (s)3.0171.645224.73961.11109.4332.552.667
Bold represents the best result, and the color represents the different categories.
Table 2. Classification results obtained by different methods for the PU dataset (1% training percentage).
Table 2. Classification results obtained by different methods for the PU dataset (1% training percentage).
ClassSamplesSvmT-hosvdSpectralformerSstnSuperpcaSpassaTensorssaOurs
1663185.688.8488.0189.4780.7291.2894.2696.47
218,64996.096.4198.1299.8996.4799.5898.8299.32
3209967.878.723.9084.7586.5777.8284.2295.2
4306484.887.1074.8189.5555.9485.0089.1987.28
5134598.998.8798.8799.7098.1499.1799.1998.43
6502975.578.8821.7295.6493.6197.2691.1598.91
7133078.775.306.1699.8479.2975.5992.0396.43
8368280.175.4791.8586.9773.4285.8588.5296.83
994799.899.67100.097.5244.2172.5497.0189.47
OA88.189.3377.9495.1286.6893.3994.2797.21
AA85.386.5964.8393.7078.7187.1492.9495.37
Kappa84.185.7669.2393.5182.1591.2192.9796.30
Time (s)12.24.731513.21183.27859.034.3311.298
Bold represents the best result, and the color represents the different categories.
Table 3. Classification results obtained by different methods for the SA dataset (1% training percentage).
Table 3. Classification results obtained by different methods for the SA dataset (1% training percentage).
ClassSamplesSvmT-hosvdSpectralformerSstnSuperpcaSpassaTensorssaOurs
1200997.4198.3498.4499.69100.098.4098.27100.0
2372699.1297.9397.2396.1099.8098.7399.6999.82
3197689.7992.9978.4399.5497.8899.1997.75100.0
4139498.4995.2899.2099.0693.1192.8698.3998.99
5267896.6796.1597.1069.3496.5196.0497.2498.69
6395999.1199.4399.6297.4199.9499.8799.2999.95
7357999.3899.3798.0875.2899.2797.2399.3399.29
81127173.5279.4778.6499.1899.5494.4883.0499.59
9620398.8198.9798.37100.099.6799.5799.1399.92
10327884.9979.3588.41100.096.9094.4490.5597.28
11106887.0883.3471.1497.4690.6598.1590.2593.29
12192795.5895.4393.18100.097.8399.4099.1499.88
1391697.0598.3497.02100.098.3897.7397.7498.54
14107089.4987.1591.5097.4697.0796.0291.4795.94
15726858.6949.1560.6797.8899.4393.6474.3999.49
16180789.1689.1486.69100.099.2495.9394.0099.15
OA86.0285.5486.5597.6698.7496.5591.2899.25
AA90.9089.9989.6194.6997.8396.9894.3598.74
Kappa84.4383.8485.0197.4098.5996.1690.3099.17
Time (s)19.8349.214554.51233.602821.9437.18814.681
Bold represents the best result, and the color represents the different categories.
Table 4. OAs (%) of different combinations of pixel processing and low-rank representation on three datasets.
Table 4. OAs (%) of different combinations of pixel processing and low-rank representation on three datasets.
Pixel ProcessingLow-Rank Representation StageAccuracy (OA)
Raw
HSI
Superpixel
Blocks (ERS)
T-HOSVDMultiscale Adaptive
T-HOSVD
IP
(2%)
PU
(1%)
SD
(1%)
62.987.5884.76
66.3989.4486.64
90.7891.6198.90
93.3197.2199.25
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Dai, Q.; Ma, C.; Zhang, Q. Advanced Hyperspectral Image Analysis: Superpixelwise Multiscale Adaptive T-HOSVD for 3D Feature Extraction. Sensors 2024, 24, 4072. https://doi.org/10.3390/s24134072

AMA Style

Dai Q, Ma C, Zhang Q. Advanced Hyperspectral Image Analysis: Superpixelwise Multiscale Adaptive T-HOSVD for 3D Feature Extraction. Sensors. 2024; 24(13):4072. https://doi.org/10.3390/s24134072

Chicago/Turabian Style

Dai, Qiansen, Chencong Ma, and Qizhong Zhang. 2024. "Advanced Hyperspectral Image Analysis: Superpixelwise Multiscale Adaptive T-HOSVD for 3D Feature Extraction" Sensors 24, no. 13: 4072. https://doi.org/10.3390/s24134072

APA Style

Dai, Q., Ma, C., & Zhang, Q. (2024). Advanced Hyperspectral Image Analysis: Superpixelwise Multiscale Adaptive T-HOSVD for 3D Feature Extraction. Sensors, 24(13), 4072. https://doi.org/10.3390/s24134072

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop