Next Article in Journal
A Cross-Resolution Surface Net Radiative Inversion Based on Transfer Learning Methods
Next Article in Special Issue
DBI-Attack:Dynamic Bi-Level Integrated Attack for Intensive Multi-Scale UAV Object Detection
Previous Article in Journal
Integration of 3D Gaussian Splatting and Neural Radiance Fields in Virtual Reality Fire Fighting
Previous Article in Special Issue
Learning SAR-Optical Cross Modal Features for Land Cover Classification
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Spectral-Spatial Mamba for Hyperspectral Image Classification

by
Lingbo Huang
,
Yushi Chen
* and
Xin He
School of Electronics and Information Engineering, Harbin Institute of Technology, Harbin 150001, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2024, 16(13), 2449; https://doi.org/10.3390/rs16132449
Submission received: 16 May 2024 / Revised: 25 June 2024 / Accepted: 1 July 2024 / Published: 3 July 2024
(This article belongs to the Special Issue Intelligent Remote Sensing Data Interpretation)

Abstract

:
Recently, transformer has gradually attracted interest for its excellence in modeling the long-range dependencies of spatial-spectral features in HSI. However, transformer has the problem of the quadratic computational complexity due to the self-attention mechanism, which is heavier than other models and thus has limited adoption in HSI processing. Fortunately, the recently emerging state space model-based Mamba shows great computational efficiency while achieving the modeling power of transformers. Therefore, in this paper, we first proposed spectral-spatial Mamba (SS-Mamba) for HSI classification. Specifically, SS-Mamba mainly includes a spectral-spatial token generation module and several stacked spectral-spatial Mamba blocks. Firstly, the token generation module converts any given HSI cube to spatial and spectral tokens as sequences. And then these tokens are sent to stacked spectral-spatial mamba blocks (SS-MB). Each SS-MB includes two basic mamba blocks and a spectral-spatial feature enhancement module. The spatial and spectral tokens are processed separately by the two basic mamba blocks, correspondingly. Moreover, the feature enhancement module modulates spatial and spectral tokens using HSI sample’s center region information. Therefore, the spectral and spatial tokens cooperate with each other and achieve information fusion within each block. The experimental results conducted on widely used HSI datasets reveal that the proposed SS-Mamba requires less processing time compared with transformer. The Mamba-based method thus opens a new window for HSI classification.

1. Introduction

Hyperspectral images (HSIs) can capture hundreds of narrow spectral bands across the electromagnetic spectrum, containing both rich spatial and spectral information [1]. Compared with traditional RGB images and multispectral images, HSIs can provide richer information about land-covers, and thus they have been widely exploited in many applications, such as environmental monitoring [2], precision agriculture [3], mineral exploration [4], and military reconnaissance [5]. Aiming at pixel-level category labeling, classification is a fundamental task for HSI processing and applications. It has become a hot topic in remote sensing research, drawing significant academic and practical interest [6,7].
HSI classification mainly consists of feature extraction and classifier classification. Researchers mainly focused on spectral features in the early time period. They usually mapped the original spectral features to a new space through linear or nonlinear transformations such as principal component analysis (PCA) [8], linear discriminant analysis (LDA) [9], and manifold learning methods [10,11]. These methods only used spectral features without considering spatial information, which limited the classification performance. Therefore, spectral-spatial feature extraction techniques have attracted lots of attention, including extended morphological profiles (EMP) [12], extended multi-attribute profiles (EMAP) [13], Gabor filtering [14], sparse representations [15], etc. The commonly used classifiers included support vector machines [16], logistic regression [17], etc. Different combinations of feature extraction techniques and classifiers form a variety of methods.
However, these methods relied on manually crafted features, requiring careful design tailored to specific scenarios. Additionally, their ability to extract high-level image features is also limited [18].
In recent years, the rapid development of deep learning has greatly propelled the advancement of HSI classification research. Deep learning networks can automatically learn high-level and discriminative features based on the data characteristics and have been widely applied in the field of HSI classification [19,20]. Typical deep learning models include stacked autoencoder (SAE) [21], deep belief networks (DBN) [22], convolutional neural network (CNN) [23], recurrent neural networks (RNNs) [24], etc. From these many models, convolutional neural network has achieved many successes and gained much interest [25]. Various CNN architectures have been proposed for extracting spectral and spatial features, including one-dimensional (1D) CNN [26], 2D CNN [27], 1D-2D CNN [28], and 3D CNN [29]. For example, Zhong et al., proposed an end-to-end spectral-spatial 3D CNN, deepening the network using residual connections [30]. Zhang et al., proposed a 3D dense connected CNN for classification, utilizing dense connections to ensure effective information transmission between layers [31]. Gong et al., introduced a novel CNN, which could extract deep multi-scale features from HSI and enhance classification performance by diversified metrics [32]. In [33], a double-branch dual-attention mechanism was combined with CNN for enhanced feature extraction. Reference [34] introduced a two-stream residual separable convolution network, specifically designed to address issues of redundant information, data scarcity, and class imbalance typically encountered in HSIs. Researchers have also extensively investigated the integration of CNN with various machine learning techniques, such as transfer learning, ensemble learning, and few-shot learning, to enhance the performance of CNN models under different conditions. Yang et al. proposed an effective transfer learning approach to deal with images with different sensors and a different number of bands [35]. The proposed method could use multi-sensor HSIs to jointly train a robust and general CNN model. In [36], a pixel-pair feature generation mechanism was specifically designed to augment the training dataset size, while ensemble learning strategies were also employed to mitigate the overfitting problem and bolster the robustness of CNN-based classifiers. Yu et al. [37] utilized the scheme of prototype learning to address few-shot HSI classification, emphasizing the efficiency in using training samples.
Compared with CNNs, which mainly focus on modeling locality, transformers have demonstrated proficiency in capturing long-range dependencies, enabling a comprehensive understanding of spatial and spectral features’ relationships in HSI [38]. Consequently, transformer-based HSI classification methods have emerged as a promising approach [39,40]. For instance, He et al. proposed HSI-BERT, where each pixel within a given HSI cube sample is treated as a token for the transformer to capture global context. It was viewed as the initial application of a transformer-based model for classification with competitive accuracies [41]. Recognizing the significance of long-range dependencies in spectral bands, Hong et al. introduced SpectralFormer, utilizing a pure transformer to process spectral features [42]. Tang et al. devised a double-attention transformer encoder to separately capture spatial and spectral features of HSIs, which were subsequently fused to enhance discriminative capabilities [43]. In [44], a cross spatial-spectral dense transformer was proposed to extract spatial and spectral features in a dense learning manner. And the cross-attention mechanism was used for feature fusion. In addition to utilizing purely a transformer as described above, researchers have explored the fusion of transformers with CNNs for feature extraction to leverage the strengths of both models effectively [45]. For example, Sun et al. [46] proposed a network comprising 2D and 3D convolution layers to preprocess input HSI samples, followed by a Gaussian-weighted feature tokenizer to generate input tokens for transformer blocks. To extract multiscale spectral-spatial features in HSIs, Wu et al., introduced an enhanced transformer utilizing multiscale CNNs to generate spectral-spatial tokens with hash-based positional embeddings [47]. Experimental results demonstrated that leveraging multiscale CNN features is advantageous in improving the classification performance. In [48], a lightweight CNN and transformers in a dual-branch manner were integrated to a basic spectral-spatial block to extract hierarchical features. The proposed model outperformed comparison methods by a significant margin.
However, the traditional transformer architecture introduces its own set of challenges, primarily due to its quadratic computational complexity driven by the self-attention mechanism [49]. This complexity becomes prohibitive when dealing with the high-dimensional data typical of HSI, which contains both spatial and spectral information. Therefore, fully modeling the long-range dependency of spatial-spectral features make a transformer model computationally heavy and, hence, impractical even when compared with other deep learning models. However, structured state space models (SSMs) have recently been developed to address transformers’ computational inefficiency on long sequences [50]. As an advanced SSM, Mamba has emerged as a promising model that outperforms in both computational efficiency and feature extraction capability [51]. Like transformers, Mamba models can capture long-range dependencies but with greater computational efficiency, making them well-suited for high-dimensional datasets like HSI.
Therefore, in this paper, we explored the application of the Mamba model for HSI classification and designed a spatial-spectral learning framework based on the Mamba model, named spectral-spatial mamba (SS-Mamba). The SS-Mamba model mainly consists of a spectral-spatial token generation module and several stacked spectral-spatial Mamba blocks to extract the deep and discriminant features of HSI. The token generation module transforms HSI cubes into sequences of spatial and spectral tokens. These tokens are then processed through several stacked spectral-spatial Mamba blocks. Each spectral-spatial Mamba block employs a double-branch structure that includes spatial mamba feature extraction, spectral mamba feature extraction, and a spectral-spatial feature enhancement module. The obtained spatial and spectral tokens are processed by the two basic Mamba blocks, correspondingly. After that, the spectral tokens and spatial tokens cooperate with each other in the designed spectral-spatial feature enhancement module. The dual-branch architecture and spectral-spatial feature enhancement effectively maximize spectral–spatial information fusion and thus improve the performance of HSI classification.
The main contributions are summarized as follows:
(1)
A spectral-spatial Mamba-based learning framework is proposed for HSI classification, which can effectively utilize Mamba’s computational efficiency and powerful long-range feature extraction capability.
(2)
We designed a spectral-spatial token generation mechanism to convert any given HSI cube to spatial and spectral tokens as sequences for input. It improves and combines the spectral and spatial patch partition to fully exploit the spectral-spatial information contained in HSI samples.
(3)
A feature enhancement module is designed to enhance the spectral-spatial features and achieve information fusion. By modulating the spatial and spectral tokens using the HSI sample’s center region information, the model can focus on the informative region and conduct spectral-spatial information interaction and fusion within each block.
The rest of the paper is organized as follows. Section 2 presents the proposed SS-Mamba for HSI classification. In Section 3, the experiments are conducted on four widely used HSI datasets. The results are presented and discussed. Section 4 provides a discussion about the proposed method. In Section 5, the conclusion is briefly summarized.

2. Methodology

The flowchart of the proposed spectral-spatial Mamba model for HSI classification is depicted in Figure 1. It can be seen that the proposed SS-Mamba mainly consists of a spectral-spatial token generation module and several stacked spectral-spatial Mamba blocks to extract the deep and discriminant features of HSI. Each spectral-spatial Mamba block employs a double-branch structure that includes spatial mamba feature extraction, spectral mamba feature extraction, and a spectral-spatial feature enhancement module. To start with, an HSI cube is generated as the model’s input for any given pixel using the spatial neighborhood region, following most of the HSI classification methods. The HSI input samples are then processed to generate spectral and spatial tokens. And these tokens are sent to several spectral-spatial Mamba blocks. Each of the blocks is composed of two basic Mamba blocks and a spectral-spatial feature enhancement module. The obtained spatial and spectral tokens are processed by the two basic Mamba blocks, correspondingly. After that, the spectral tokens and spatial tokens cooperate with each other in the designed spectral-spatial feature enhancement module. After feature extraction by these spectral-spatial Mamba blocks, the spectral and spatial tokens are averaged and then added to form the final spectral-spatial feature for the given HSI sample. Finally, the obtained feature is sent to a fully connected layer to accomplish classification.

2.1. Overview of the State Space Models

The state space model serves as a framework for the modeling relationship between input and output sequences. Specifically, it maps a one-dimensional input x ( t ) R to an output x ( t ) R through the hidden state h ( t ) R N . This procedure can be formulated through the following ordinary differential equation:
h ( t ) = A h ( t ) + B x ( t ) ,
y ( t ) = C h ( t ) ,
Here, A R N × N represents the state matrix, while B R N × 1 and C R N × 1 are the system parameters. To adapt this continuous system for deep learning applications in discrete sequences like image and text, the structural state space model (S4) further employs discretization. Specifically, a timescale parameter is introduced, and a consistent discretization rule is applied to convert A and B into discrete parameters A ¯ and B ¯ , respectively. The discretization employed in the S4 model uses the zero order holding (ZOH) rule, defined as
    A ¯ = e A ,
B ¯ = A 1 e A I B ,
And the discretization SSM can then be calculated in recurrence denoted by the following equations:
h t = A ¯ h t 1 + B ¯ x t ,
y t = C h t .
For fast and efficient parallel training, the above process in recurrence can also be reformulated using convolution:
y = x K ¯ ,
where K ¯ denotes the structured convolutional kernels. It is obtained by
K ¯ = ( C B ¯ ,   C A ¯ B ¯ , , C A ¯ N 1 B ¯ ) ,
where N denoted the input sequence’s length.
In Mamba, the selective scan mechanism is further used. Specifically, the matrices B , C, and are generated from the input data x , allowing the model to dynamically model the contextual relationship within the input sequence. Inspired by the transformer and Hungry Hungry Hippo (H3) architectures [52], the normalization, residual connectivity, gated MLP, and SSM are combined to form a basic Mamba block, which constitutes the fundamental component of a Mamba network. The structure of a basic Mamba block is shown in Figure 1.

2.2. Spectral-Spatial Token Generation

To utilize Mamba’s power sequence modeling ability, the HSI input is needed to be transformed to sequence as Mamba’s input. Given an HSI training sample x R H × W × B , where H and W are the height and width of the input, respectively, and B is the number of spectral bands, the proposed method converts them to spatial token sequence and spatial token sequence, respectively. Figure 2 illustrates the spectral and spatial token generation process.
For spatial token generation, there are mainly three steps, namely, spectral mapping, spatial partition, and patch embedding.
We firstly process spectral features before performing spatial partition for the spatial Mamba. The purpose is to fully exploit the spectral-spatial information contained in HSI samples in the case that the spatial-partition-based Mamba feature extraction mainly focuses on spatial tokens’ relationships. Specifically, the input HSI sample is firstly reshaped to a tensor with a shape of H W × B , and then a lightweight multilayer perception (MLP) is used for spectral feature mapping:
x s p a R H W × D = M L P ( R e s h a p e x ) ,
where M L P ( · ) denotes the corresponding mapping function and D is the mapped feature dimension.
The spectrally mapped HSI sample is then spatially partitioned into N = H W / P s p a 2 non-overlapped patches p ¯ s p a , where p ¯ s p a R N × P 2 D , and P s p a is the spatially partitioned patch size. After that, a linear layer is used to project these patches to a given dimension and the obtained spatial input tokens:
Z s p a 0 = p ¯ s p a E s p a ,
where E s p a R P 2 D × D denotes the learnable matrix in the linear layer and D represents the tokens’ dimension.
Similar to spatial token generation, the spectral token generation process also contains three steps including spatial mapping, spectral partition, and patch embedding.
To start with, we firstly extract the center region as the input x ^ R S × S × B , where S is a small integer, i.e., 3. A small center region can make the spectral feature robust when compared with only the center pixel. Like the spatial token generation process, a spatial feature is processed before performing spatial partition. The obtained input x ^ is reshaped to a tensor with a shape of B × S 2 , and then another lightweight MLP is used for spatial feature mapping:
x s p e R B × D = M L P ( R e s h a p e x ^ ) .
The spatially mapped HSI sample x s p e is then spectral partitioned into M = B / P s p e non-overlapped patches p ¯ s p e , where p ¯ s p e R M × P s p e D , and P s p e is the spectrally partitioned patch size. After that, the spectral input token can be obtained by patch embedding:
Z s p e 0 = p ¯ s p e E s p e R M × D ,
where E s p e R P s p e D × D denotes the learnable matrix in the linear layer.
Following the process of transformers, the obtained tokens are added with positional embedding to provide location information within the HSI sample. It is worth noting that the spatial tokens are provided with 2D sinusoidal position embedding, while the spectral tokens use 1D positional embedding.

2.3. Spectral-Spatial Mamba Block

The spectral-spatial feature extraction is mainly achieved by several spectral-spatial Mamba blocks, which consist of two distinct basic Mamba blocks and a feature enhancement module. The two basic Mamba blocks are mainly used for spatial and spectral feature extraction, correspondingly. And in the feature enhancement module, the spatial or spectral tokens are modulated using the HSI sample’s center region information from the tokens of the other type. By this way, the spectral tokens and spatial tokens cooperate with each other and achieve information fusion in each block.
Specifically, in the l-th block, the spectral and spatial tokens are firstly processed by Mamba blocks:
Z s p e l = M B s p e l ( Z s p e l 1 ) ,
Z s p a l = M B s p a l ( Z s p a l 1 ) ,
where Z s p e l and Z s p a l denote the l-th block’s output spectral and spatial tokens, respectively, and M B s p e l and M B s p a l represent the mapping function of basic Mamba blocks introduced in Section 2.1 for spectral and spatial tokens, respectively.
After the mamba’s feature extraction, the obtained spectral and spatial tokens are then sent to the feature enhancement module for information interaction, just as Figure 3 shows.
For the spatial tokens, we first take out the token corresponding to the central patch, denoted as f 1 l :
f 1 l = Z s p a l [ ( N + 1 ) / 2 , : ] ,
And for the spectral tokens, these N tokens are averaged to obtain the center region’s spectral feature from the view of the spectral branch, denoted as f 2 l :
f 2 l = 1 M j = 1 M Z s p e l [ j , : ]           j = 1 , , M .
The obtained features are then fused by average:
f l = f 1 l + f 2 l / 2 .
An MLP is used for further feature extraction, and the sigmoid activation is also used for scaling:
s l = σ ( M L P f l ) ,
where σ represents the sigmoid activation function. The obtained feature s l can be viewed as the modulated weights, which is to be multiplied with the spectral and spatial tokens. Before multiplication, it should cope with matching dimensions. We made N and M copies of s for spatial and spatial tokens, respectively:
A s p a l = [ s l ; s l ; s l ] R N × D ,
A s p e l = [ s l ; s l ; s l ] R M × D .
The obtained modulated matrices A s p a l and A s p e l are then multiplied with the spectral and spatial tokens for forcing the model to focus more on the most informative central region as the input HSI image’s labels are mainly determined by the center pixel’s label, and the spectral information is mainly needed here for enhancement. The procedure can be formulated as
Z s p a l = A s p a l Z s p a l ,
Z s p e l = A s p e l Z s p e l ,
where ⊙ means the element-wise multiplication.
After feature extraction by L spectral-spatial Mamba blocks, the obtained spectral and spatial tokens are averaged and then added to form the final spectral-spatial feature for the given HSI sample, which can be formulated by
f s p a = 1 N j = 1 M Z s p a L [ j , : ]             j = 1 , , N ,
f s p e = 1 M j = 1 M Z s p e L j , :             j = 1 , , M ,
f = f s p a + f s p e .
And the representation f of the input HSI sample is sent into a fully connected layer to obtain the final logit prediction:
y ^ R K = F C f ,
where F C · denotes the mapping function of a fully connected layer and y ^ is the predicted label vector for the input HSI sample x . Moreover, K is the number of classes.
The cross-entropy loss is used for optimizing the designed model:
L c l s = k = 1 K y k l o g y ^ k ,
where y ^ k is the k-th element of y ^ , and y k is the k-th element of the one-hot label vector.

3. Results

3.1. Datasets Description

The performance of the proposed Mamba-based method was assessed using four widely used HSI datasets: Indian Pines, Pavia University, Houston, and Chikusei.
(1) Indian Pines: The Indian Pines dataset primarily records agricultural areas in Northwestern Indiana, USA, captured in June 1992 by the Airborne Visible/Infrared Imaging Spectrometer (AVIRIS). The dataset comprises 145 × 145 pixels with a spatial resolution of 20 m. It includes 220 spectral bands, covering wavelengths from 400 nm to 2500 nm. For application, 200 bands are retained after removing 20 bands with low signal-to-noise ratio. The dataset includes 10,249 labeled samples belonging to 16 distinct land-cover categories. The dataset’s false color and ground-truth maps are illustrated in Figure 4, and the distribution of training and test samples is detailed in Table 1.
(2) Pavia University: The Pavia University dataset, captured by the Reflective Optics System Imaging Spectrometer (ROSIS) over the University of Pavia, Italy, contains 610 × 340 pixels at a fine spatial resolution of 1.3 m. This dataset records 103 spectral bands and labels 42,776 pixels from nine land-cover classes. Figure 5 shows the false color and ground-truth maps, while Table 2 shows the distribution of training and test samples of each class.
(3) Houston: The Houston dataset focuses on an urban area around the University of Houston campus, USA. It was captured by the National Center for Airborne Laser Mapping (NCALM) for the 2013 IEEE GRSS Data Fusion Contest [53]. The dataset contains 349 × 1905 pixels with a spatial resolution of 2.5 m per pixel and includes 144 spectral bands spanning 380 nm to 1050 nm. It consists of 15,029 labeled pixels categorized into 15 land-cover classes. Figure 6 illustrates both the false color and ground-truth maps, and Table 3 details the class-wise distribution of training and test samples.
(4) Chikusei: The Chikusei dataset is an aerial hyperspectral dataset acquired using the Headwall Hyperspec-VNIR-C sensor in Chikusei, Japan, on 29 July 2014 [54]. It comprises 128 spectral bands spanning wavelengths from 343 nm to 1018 nm. The dataset spatial size is 2517 × 2335, with a resolution of 2.5 m. It includes 19 types of land-covers, encompassing urban and rural areas. In Figure 7, the true color composite image and the corresponding ground truth map of the Chikusei dataset are depicted. Table 4 shows the distribution of training and test samples of each class.

3.2. Experimental Setup

To demonstrate the effectiveness of the proposed SS-Mamba, various kinds of HSI classification methods were selected and evaluated as comparison methods. These comparison methods are listed as follows:
(1)
EMP-SVM [16]: this method utilizes EMP for spatial feature extraction followed by a classic SVM for final classification. This approach is commonly employed as a benchmark against deep learning-based methodologies.
(2)
CNN: it is a vanilla CNN that simply contains four convolutional layers. It is viewed as a basic spectral-spatial deep learning-based model for HSI classification.
(3)
SSRN [30]: it is a 3D deep learning framework that uses three-dimensional convolutional kernels and residual blocks to improve the CNN’s performance of HSI classification.
(4)
DBDA [33]: it is an advanced CNN model that integrates a double-branch dual-attention mechanism for enhanced feature extraction. It is used for comparison with transformer-based methodologies, which rely on self-attention mechanisms.
(5)
MSSG [55]: it employs a super-pixel structured graph U-Net to learn multiscale features across multilevel graphs. As a graph CNN and global learning model, MSSG is contrasted with the proposed Mamba and patch-based methods.
(6)
SSFTT [46]: SSFTT is a spatial-spectral transformer that designs a unique tokenization method and uses CNN to provide local features for the transformer.
(7)
LSFAT [56]: it is a local semantic feature aggregation-based transformer that has the advantages of learning multiscale features.
(8)
CT-Mixer [45]: it is an aggregated framework of CNN and transformer, which is hoped to effectively utilize both of the advantages of the above two classic models.
The listed comparison methods encompass a diverse kind of methods, including the traditional method, CNN-based methods, transformer-based models, attention-based methods, and mixture methods, thus offering a comprehensive and exhaustive basis for comparison. For details, one can refer to the original papers.
When using SS-Mamba, a spatial window size of 27 × 27 (i.e., H = W = 27 ) was used for any given pixel to generate the model’s HSI input sample. As for the hyper-parameters in the proposed SS-Mamba, the spatial partition size was set to three, that is P s p a = 3. And the spectral partition size was set to two, that is P s p e = 2. The embedding dimension for each Mamba block was 64, that is D = 64. The widely used multi-step learning rate scheduler was employed for training. Specifically, the initial learning rate was 0.0005, and it was divided by two every 80 epochs. The total number of epochs was set to 180 for all the four datasets. And the mini-batch strategy with a batch size of 256 was used for all the datasets.
The classification performance was primarily evaluated based on metrics including overall accuracy (OA), average accuracy (AA), and the Kappa coefficient (K). To ensure the reliability and robustness of the classification methods’ output results, the experiments were repeated ten times with varying random initializations. Moreover, the training samples were randomly selected from all labeled samples each time. However, the training samples remained the same for all the involved methods each time.

3.3. Ablation Experiments

3.3.1. Ablation Experiment with Basic Sequence Model

To demonstrate the classification ability of the Mamba model, the ablation experiments with different sequence models were conducted on the four datasets. These sequence models included long short-term memory network (LSTM), gated recurrent unit (GRU), and transformer. The overall framework of the model remained unchanged, with only the basic sequence model varying. The related results are shown in Table 5, Table 6, Table 7 and Table 8. From the obtained results, one can see the following:
(i)
Spectral-spatial models achieved the highest accuracies with the same basic sequence model, followed by the spatial model, with the spectral model proven to be the least effective. For example, as shown in Table 5, spectral-spatial LSTM achieved better results than spatial LSTM and spectral LSTM, with improvements of 3.76 percentage points and 32.85 percentage points in terms of OA, respectively. Moreover, the spectral-spatial mamba outperformed spatial mamba by 2.77 percentage points, 5.14 percentage points, and 0.0364 in terms of OA, AA, and K on the Pavia University dataset, respectively. The results indicate that the designed spectral-spatial learning framework is effective for different sequence models.
(ii)
With the learning framework, the Mamba-based models achieved higher accuracies than the classical sequence models such as LSTM, GRU, and transformer. For example, the spectral-spatial Mamba outperformed spectral-spatial GRU by 0.45 percentage points, 0.52 percentage points, and 0.0046 in terms of OA, AA, and K on the Pavia University dataset, respectively. On the Houston dataset, spectral-spatial Mamba yielded better results than transformer, GRU, and LSTM, with improvements of 0.92 percentage points, 0.49 percentage points, and 0.80 percentage points for OA, respectively. One can also draw similar conclusions for spatial or spectral learning methods on the four datasets. The results indicate that the used Mamba-based sequence models are effective for different learning frameworks.
Notably, we find that all the spectral models seem harder to be trained well compared with the other models, especially on the Indian Pines dataset. Specifically, the spectral models need larger learning rates and more epochs to train. Moreover, these models are all much harder to train and perform the worst on the Indian Pines dataset, probably due to the low quality of the dataset.

3.3.2. Ablation Experiment with Feature Enhancement Module

For the spectral-spatial Mamba model, we designed the feature enhancement module, which can use the HSI sample’s center region information from the tokens to enhance spectral-spatial features. To demonstrate the effectiveness of this module, the ablation experiment was conducted. Table 9 shows the ablation experimental results. From the results, one can see that the spectral-spatial Mamba with feature enhancement module achieved higher accuracies when compared with the model without a feature enhancement module. For example, the feature enhancement module improved the classification results by 2.09 percentage points, 1.78 percentage points, and 0.0226 in terms of OA, AA, and K on the Houston dataset, respectively. The results indicate that the designed feature enhancement module is effective.

3.4. Classification Results

The classification results of different methods on the four datasets are shown in Table 10, Table 11, Table 12 and Table 13. From these results, one can see that the proposed SS-Mamba achieved superior classification performance over other comparison methods when using twenty training samples per class across all the four datasets.
Specifically, the following can be seen: (1) The CNN-based methods usually achieve higher classification accuracies than the transformer-based methods. As an example, on the Pavia University dataset, the MSSG performed better than the CT-Mixer with an improvement of 1.46 percentage points for OA, 3.44 percentage points for AA, and 0.0194 for K. On the Houston dataset, the overall accuracies of the comparison transformer-based methods were lower than 93%. In contrast, the overall accuracies achieved by CNN-based methods were generally higher than 93%. And the OA of MSSG was even close to 94%. It seems that it was necessary to improve transformer-based methods in the case of limited training samples. (2) The designed spectral-spatial Mamba model obtained the highest accuracies when compared with other methods on the four datasets. For example, in Table 9, SS-Mamba yielded better results than the DBDA, with an improvement of 0.53 percentage points for OA, 3.68 percentage points for AA, and 0.0073 for K. On the Houston dataset, the proposed method achieved better results than MSSG, DBDA, and SSFTT, with improvements of 0.38 percentage points, 0.63 percentage points, and 4.12 percentage points in terms of OA, respectively. Compared with transformer, Mamba, which is also a sequence model, had better classification performance, which shows the potential application of the sequence model in HSI classification.

3.5. Classification Maps

As qualitative classification results, the classification maps for different methods on the four datasets are shown in Figure 8, Figure 9, Figure 10 and Figure 11. From these maps, it can be clearly seen that the proposed SS-Mamba achieved better classification performance than the comparison methods. For example, on the Pavia dataset, many methods tended to misclassify some pixels from Asphalt, Trees, and Bricks. The reason may be that they were located close to each other. This suggests that these models may focus more on spatial information brought about by the spatial window, potentially neglecting the spectral characteristics for accurate classification. In contrast, the proposed method demonstrated good performance in this context. Thanks to the designed spectral-spatial learning framework and the feature extraction ability of Mamba, it can make full use of spectral–spatial information and improve the performance of HSI classification.

4. Discussion

4.1. Complexity Analysis

Table 14 shows the number of parameters (i.e., Param.) and consuming time (i.e., Test Time) during testing on a batch of 100 samples on the Pavia University dataset. The SS-LSTM/GRU/Transformer means the corresponding spectral-spatial sequence model in the ablation experiment.
The results reveal that the Mamba model exhibited faster inference times compared to the transformer model. Additionally, sequential models like Mamba generally necessitate longer inference times but also achieve higher classification accuracies when compared to CNN models. Furthermore, the newly proposed Mamba-2 (i.e., the second generation of Mamba) has faster training and inference processes, which will further benefit our work in the future.

4.2. Features Maps

Taking the two input samples of Pavia University as an example, Figure 12 shows the corresponding spatial feature maps of each block. And different columns show the feature maps on different feature channels (i.e., the first eight channels). It can be seen that the model still retained some of the image details in the token of each block. On the one hand, different channels focused on different image information, and their feature maps were not the same. On the other hand, the feature maps of different blocks had some differences. And the image details (semantic structure) were gradually blurred with the depth. However, probably because the model was shallow with a few blocks, the change of feature maps with depth was not particularly obvious.

4.3. Comparison of the Proposed Classification Method with Spectral Unmixing-Based Methods

The spectral unmixing technique plays an important role in HSI processing and application, and there are many classification methods that utilize spectral unmixing for improving performance. Therefore, it is necessary to discuss the proposed SS-Mamba’s advantages over the spectral-unmixing-based classification methods. In [57], hard examples were firstly selected based on specific criteria. Then, spectral unmixing was used to improve their quality, facilitating subsequent classification tasks. In [58], the researchers focused on designing autoencoders for unmixing tasks based on the linear mixing model and then using the encoder’s features for classification. Li et al. collected the abundance representations from multiple HSIs to form an enlarged dataset and then used the enlarged abundance dataset to train a classifier like CNN, which could alleviate the overfitting problem [59]. Based on these relevant works, we would like to emphasize the distinctive advantages of our classification approach over these spectral unmixing-based methods.
While spectral unmixing methods are effective in enhancing image quality and extracting spectral information, the whole classification methods typically involve a multi-step process that includes spectral unmixing followed by separate feature extraction and classification steps. This can be time-consuming and requires prior knowledge of hyperspectral mixing models, which may vary in accuracy and impact the final classification results.
In contrast, our proposed classification method is designed as an end-to-end framework. This streamlined approach eliminates the need for intermediate processing stages and expert knowledge of spectral unmixing models, making it more efficient and user-friendly. By leveraging the capabilities of Mamba for feature extraction and classification simultaneously, our method offers a more seamless and accessible solution for achieving high classification accuracy. In the feature, we can try to combine the proposed method and spectral unmixing technique to obtain better classification performance.

4.4. Limitations and Feature Work

The SS-Mamba divides the HSI samples as tokens, which are the input of the SS-Mamba. The token generation process may disrupt the semantic structure of the HSI samples to some extent. Specifically, the current partitioning method does not take object’s orientation and shape into consideration, potentially causing pixels belonging to the same object to be distributed across different patches. Therefore, future work will focus on the combination with traditional architectures like CNN for enhancing local feature extraction and improving the generation of tokens.

5. Conclusions

In this study, we developed an effective exploration to build a spectral-spatial model for HSI classification, which solely used an emerging sequence model named Mamba. The proposed model converts any given HSI cube to spatial and spectral tokens as sequences. And then stacked Mamba blocks are used to effectively model the relationships between the tokens. Moreover, the spectral and spatial tokens are enhanced and fused for more discriminant features. The proposed SS-Mamba was then evaluated on four widely used datasets (i.e., Indian Pines, Pavia University, Houston, and Chikusei datasets), and some main conclusions can be drawn from the experimental results as follows:
(1)
Through a comparative analysis of classification results, it is evident that the proposed SS-Mamba can make full use of spatial-spectral information, and it can achieve superior performance for HSI classification tasks.
(2)
The ablation experiments show that as a sequence model, Mamba is effective and can gain competitive classification performance for HSI classification when compared with other sequence models like transformer, LSTM, and GRU.
(3)
The ablation experiments also show that the designed spectral-spatial learning framework is effective for different sequence models, when compared with spectral-only or spatial-only models.
(4)
The designed feature enhancement module is effective to enhance spectral and spatial features and improve the SS-Mamba’s classification performance.
This research explored the prospects of the utilization Mamba model in HSI classification and provides new insights for further studies.

Author Contributions

Conceptualization: Y.C.; methodology: L.H. and Y.C.; writing—original draft preparation: L.H., Y.C. and X.H. All authors have read and agreed to the published version of the manuscript.

Funding

This work was funded by the Postdoctoral Fellowship Program of CPSF under GZB20230968, and the National Natural Science Foundation of China under Grant 62371169.

Data Availability Statement

The Houston dataset is available at https://hyperspectral.ee.uh.edu/ (accessed on 1 September 2020). The Indian Pines and Pavia University datasets are available at http://www.ehu.eus/ccwintco/index.php?title=Hyperspectral_Remote_Sensing_Scenes (accessed on 1 September 2020). The Chikusei dataset is available at https://naotoyokoya.com/Download.html (accessed on 1 September 2020).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Paoletti, M.E.; Haut, J.M.; Plaza, J.; Plaza, A. Deep learning classifiers for hyperspectral imaging: A review. ISPRS J. Photogramm. Remote Sens. 2019, 158, 279–317. [Google Scholar] [CrossRef]
  2. Camps-Valls, G.; Tuia, D.; Bruzzone, L.; Benediktsson, J.A. Advances in hyperspectral image classification: Earth monitoring with statistical learning methods. IEEE Signal Process. Mag. 2013, 31, 45–54. [Google Scholar] [CrossRef]
  3. Gevaert, C.M.; Suomalainen, J.; Tang, J.; Kooistra, L. Generation of spectral–temporal response surfaces by combining multispectral satellite and hyperspectral UAV imagery for precision agriculture applications. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2015, 8, 3140–3146. [Google Scholar] [CrossRef]
  4. Murphy, R.J.; Schneider, S.; Monteiro, S.T. Consistency of measurements of wavelength position from hyperspectral imagery: Use of the ferric iron crystal field bbsorption at 900 nm as an indicator of mineralogy. IEEE Trans. Geosci. Remote Sens. 2013, 52, 2843–2857. [Google Scholar] [CrossRef]
  5. Ardouin, J.-P.; Lévesque, J.; Rea, T.A. A demonstration of hyperspectral image exploitation for military applications. In Proceedings of the 2007 10th International Conference on Information Fusion, Quebec, QC, Canada, 9–12 July 2007; pp. 1–8. [Google Scholar]
  6. Chang, C.-I. Hyperspectral Data Exploitation: Theory and Applications; John Wiley & Sons: Hoboken, NJ, USA, 2007. [Google Scholar]
  7. Bioucas-Dias, J.M.; Plaza, A.; Camps-Valls, G.; Scheunders, P.; Nasrabadi, N.; Chanussot, J. Hyperspectral remote sensing data analysis and future challenges. IEEE Geosci. Remote Sens. Mag. 2013, 1, 6–36. [Google Scholar] [CrossRef]
  8. Kang, X.; Xiang, X.; Li, S.; Benediktsson, J.A. PCA-based edge-preserving features for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 7140–7151. [Google Scholar] [CrossRef]
  9. Fu, L.; Li, Z.; Ye, Q.; Yin, H.; Liu, Q.; Chen, X.; Fan, X.; Yang, W.; Yang, G. Learning Robust Discriminant Subspace Based on Joint L2, p-and L2, s-Norm Distance Metrics. IEEE Trans. Neural Netw. Learn. Syst. 2020, 33, 130–144. [Google Scholar] [CrossRef]
  10. Lunga, D.; Prasad, S.; Crawford, M.M.; Ersoy, O. Manifold-learning-based feature extraction for classification of hyperspectral data: A review of advances in manifold learning. IEEE Signal Process. Mag. 2013, 31, 55–66. [Google Scholar] [CrossRef]
  11. Huang, H.; Shi, G.; He, H.; Duan, Y.; Luo, F. Dimensionality reduction of hyperspectral imagery based on spatial–spectral manifold learning. IEEE Trans. Cybern. 2019, 50, 2604–2616. [Google Scholar] [CrossRef]
  12. Benediktsson, J.A.; Palmason, J.A.; Sveinsson, J.R. Classification of hyperspectral data from urban areas based on extended morphological profiles. IEEE Trans. Geosci. Remote Sens. 2005, 43, 480–491. [Google Scholar] [CrossRef]
  13. Dalla Mura, M.; Atli Benediktsson, J.; Waske, B.; Bruzzone, L. Extended profiles with morphological attribute filters for the analysis of hyperspectral data. Int. J. Remote Sens. 2010, 31, 5975–5991. [Google Scholar] [CrossRef]
  14. Jia, S.; Shen, L.; Li, Q. Gabor feature-based collaborative representation for hyperspectral imagery classification. IEEE Trans. Geosci. Remote Sens. 2014, 53, 1118–1129. [Google Scholar]
  15. Chen, Y.; Nasrabadi, N.M.; Tran, T.D. Hyperspectral image classification using dictionary-based sparse representation. IEEE Trans. Geosci. Remote Sens. 2011, 49, 3973–3985. [Google Scholar] [CrossRef]
  16. Melgani, F.; Bruzzone, L. Classification of hyperspectral remote sensing images with support vector machines. IEEE Trans. Geosci. Remote Sens. 2004, 42, 1778–1790. [Google Scholar] [CrossRef]
  17. Khodadadzadeh, M.; Li, J.; Plaza, A.; Bioucas-Dias, J.M. A subspace-based multinomial logistic regression for hyperspectral image classification. IEEE Geosci. Remote Sens. Lett. 2014, 11, 2105–2109. [Google Scholar] [CrossRef]
  18. Yu, H.; Xu, Z.; Zheng, K.; Hong, D.; Yang, H.; Song, M. MSTNet: A multilevel spectral–spatial transformer network for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5532513. [Google Scholar] [CrossRef]
  19. Li, S.; Song, W.; Fang, L.; Chen, Y.; Ghamisi, P.; Benediktsson, J.A. Deep learning for hyperspectral image classification: An overview. IEEE Trans. Geosci. Remote Sens. 2019, 57, 6690–6709. [Google Scholar] [CrossRef]
  20. Zhang, W.-T.; Li, Y.-B.; Liu, L.; Bai, Y.; Cui, J. Hyperspectral Image Classification Based on Spectral-Spatial Attention Tensor Network. IEEE Geosci. Remote Sens. Lett. 2024, 21, 5500305. [Google Scholar] [CrossRef]
  21. Chen, Y.; Lin, Z.; Zhao, X.; Wang, G.; Gu, Y. Deep learning-based classification of hyperspectral data. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2014, 7, 2094–2107. [Google Scholar] [CrossRef]
  22. Chen, C.; Ma, Y.; Ren, G. Hyperspectral classification using deep belief networks based on conjugate gradient update and pixel-centric spectral block features. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 13, 4060–4069. [Google Scholar] [CrossRef]
  23. Yue, G.; Zhang, L.; Zhou, Y.; Wang, Y.; Xue, Z. S2TNet: Spectral-Spatial Triplet Network for Few-Shot Hyperspectral Image Classification. IEEE Geosci. Remote Sens. Lett. 2024, 21, 5501705. [Google Scholar] [CrossRef]
  24. Mou, L.; Ghamisi, P.; Zhu, X.X. Deep recurrent neural networks for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 3639–3655. [Google Scholar] [CrossRef]
  25. Yu, C.; Han, R.; Song, M.; Liu, C.; Chang, C.-I. A simplified 2D-3D CNN architecture for hyperspectral image classification based on spatial–spectral fusion. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 13, 2485–2501. [Google Scholar] [CrossRef]
  26. Sun, H.; Zheng, X.; Lu, X.; Wu, S. Spectral–spatial attention network for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2019, 58, 3232–3245. [Google Scholar] [CrossRef]
  27. Zhao, W.; Du, S. Spectral–spatial feature extraction for hyperspectral image classification: A dimension reduction and deep learning approach. IEEE Trans. Geosci. Remote Sens. 2016, 54, 4544–4554. [Google Scholar] [CrossRef]
  28. Huang, L.; Chen, Y. Dual-path siamese CNN for hyperspectral image classification with limited training samples. IEEE Geosci. Remote Sens. Lett. 2020, 18, 518–522. [Google Scholar] [CrossRef]
  29. Xu, Q.; Xiao, Y.; Wang, D.; Luo, B. CSA-MSO3DCNN: Multiscale octave 3D CNN with channel and spatial attention for hyperspectral image classification. Remote Sens. 2020, 12, 188. [Google Scholar] [CrossRef]
  30. Zhong, Z.; Li, J.; Luo, Z.; Chapman, M. Spectral–spatial residual network for hyperspectral image classification: A 3-D deep learning framework. IEEE Trans. Geosci. Remote Sens. 2017, 56, 847–858. [Google Scholar] [CrossRef]
  31. Zhang, C.; Li, G.; Du, S.; Tan, W.; Gao, F. Three-dimensional densely connected convolutional network for hyperspectral remote sensing image classification. J. Appl. Remote Sens. 2019, 13, 016519. [Google Scholar] [CrossRef]
  32. Gong, Z.; Zhong, P.; Yu, Y.; Hu, W.; Li, S. A CNN with multiscale convolution and diversified metric for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2019, 57, 3599–3618. [Google Scholar] [CrossRef]
  33. Li, R.; Zheng, S.; Duan, C.; Yang, Y.; Wang, X. Classification of hyperspectral image based on double-branch dual-attention mechanism network. Remote Sens. 2020, 12, 582. [Google Scholar] [CrossRef]
  34. Zahisham, Z.; Lim, K.M.; Koo, V.C.; Chan, Y.K.; Lee, C.P. 2SRS: Two-stream residual separable convolution neural network for hyperspectral image classification. IEEE Geosci. Remote Sens. Lett. 2023, 20, 5501505. [Google Scholar] [CrossRef]
  35. Yang, B.; Hu, S.; Guo, Q.; Hong, D. Multisource domain transfer learning based on spectral projections for hyperspectral image classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 3730–3739. [Google Scholar] [CrossRef]
  36. Dong, S.; Feng, W.; Quan, Y.; Dauphin, G.; Gao, L.; Xing, M. Deep ensemble CNN method based on sample expansion for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5531815. [Google Scholar] [CrossRef]
  37. Yu, C.; Gong, B.; Song, M.; Zhao, E.; Chang, C.-I. Multiview calibrated prototype learning for few-shot hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5544713. [Google Scholar] [CrossRef]
  38. Peng, Y.; Zhang, Y.; Tu, B.; Li, Q.; Li, W. Spatial–spectral transformer with cross-attention for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5537415. [Google Scholar] [CrossRef]
  39. Qi, W.; Huang, C.; Wang, Y.; Zhang, X.; Sun, W.; Zhang, L. Global-local three-dimensional convolutional transformer network for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5510820. [Google Scholar] [CrossRef]
  40. Zou, J.; He, W.; Zhang, H. Lessformer: Local-enhanced spectral-spatial transformer for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5535416. [Google Scholar] [CrossRef]
  41. He, J.; Zhao, L.; Yang, H.; Zhang, M.; Li, W. HSI-BERT: Hyperspectral image classification using the bidirectional encoder representation from transformers. IEEE Trans. Geosci. Remote Sens. 2019, 58, 165–178. [Google Scholar] [CrossRef]
  42. Hong, D.; Han, Z.; Yao, J.; Gao, L.; Zhang, B.; Plaza, A.; Chanussot, J. SpectralFormer: Rethinking hyperspectral image classification with transformers. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5518615. [Google Scholar] [CrossRef]
  43. Tang, P.; Zhang, M.; Liu, Z.; Song, R. Double attention transformer for hyperspectral image classification. IEEE Geosci. Remote Sens. Lett. 2023, 20, 5502105. [Google Scholar] [CrossRef]
  44. Xu, H.; Zeng, Z.; Yao, W.; Lu, J. CS2DT: Cross spatial–spectral dense transformer for hyperspectral image classification. IEEE Geosci. Remote Sens. Lett. 2023, 20, 5510105. [Google Scholar] [CrossRef]
  45. Zhang, J.; Meng, Z.; Zhao, F.; Liu, H.; Chang, Z. Convolution transformer mixer for hyperspectral image classification. IEEE Geosci. Remote Sens. Lett. 2022, 19, 6014205. [Google Scholar] [CrossRef]
  46. Sun, L.; Zhao, G.; Zheng, Y.; Wu, Z. Spectral–spatial feature tokenization transformer for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5522214. [Google Scholar] [CrossRef]
  47. Wu, K.; Fan, J.; Ye, P.; Zhu, M. Hyperspectral image classification using spectral–spatial token enhanced transformer with hash-based positional embedding. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5507016. [Google Scholar] [CrossRef]
  48. Wang, W.; Liu, L.; Zhang, T.; Shen, J.; Wang, J.; Li, J. Hyper-ES2T: Efficient spatial–spectral transformer for the classification of hyperspectral remote sensing images. Int. J. Appl. Earth Obs. Geoinf. 2022, 113, 103005. [Google Scholar] [CrossRef]
  49. Gao, Z.; Shi, X.; Wang, H.; Zhu, Y.; Wang, Y.B.; Li, M.; Yeung, D.-Y. Earthformer: Exploring space-time transformers for earth system forecasting. Adv. Neural Inf. Process. Syst. 2022, 35, 25390–25403. [Google Scholar]
  50. Gu, A.; Goel, K.; Ré, C. Efficiently modeling long sequences with structured state spaces. arXiv 2021, arXiv:2111.00396. [Google Scholar]
  51. Gu, A.; Dao, T. Mamba: Linear-time sequence modeling with selective state spaces. arXiv 2023, arXiv:2312.00752. [Google Scholar]
  52. Fu, D.Y.; Dao, T.; Saab, K.K.; Thomas, A.W.; Rudra, A.; Ré, C. Hungry hungry hippos: Towards language modeling with state space models. arXiv 2022, arXiv:2212.14052. [Google Scholar]
  53. Debes, C.; Merentitis, A.; Heremans, R.; Hahn, J.; Frangiadakis, N.; van Kasteren, T.; Liao, W.; Bellens, R.; Pižurica, A.; Gautama, S. Hyperspectral and LiDAR data fusion: Outcome of the 2013 GRSS data fusion contest. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2014, 7, 2405–2418. [Google Scholar] [CrossRef]
  54. Yokoya, N.; Iwasaki, A. Airborne hyperspectral data over Chikusei. Space Appl. Lab. Univ. Tokyo Tokyo Jpn. Tech. Rep. 2016, 5, 5. [Google Scholar]
  55. Liu, Q.; Xiao, L.; Yang, J.; Wei, Z. Multilevel superpixel structured graph U-Nets for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5516115. [Google Scholar] [CrossRef]
  56. Tu, B.; Liao, X.; Li, Q.; Peng, Y.; Plaza, A. Local semantic feature aggregation-based transformer for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5536115. [Google Scholar] [CrossRef]
  57. Fang, B.; Bai, Y.; Li, Y. Combining spectral unmixing and 3D/2D dense networks with early-exiting strategy for hyperspectral image classification. Remote Sens. 2020, 12, 779. [Google Scholar] [CrossRef]
  58. Guo, A.J.; Zhu, F. Improving deep hyperspectral image classification performance with spectral unmixing. Signal Process. 2021, 183, 107949. [Google Scholar] [CrossRef]
  59. Li, C.; Cai, R.; Yu, J. An attention-based 3D convolutional autoencoder for few-shot hyperspectral unmixing and classification. Remote Sens. 2023, 15, 451. [Google Scholar] [CrossRef]
Figure 1. The illustration for the design details of the proposed SS-Mamba—working flow. For illustrative purposes, a single image flow instead of a batch is shown here. The proposed SS-Mamba mainly consists of a spectral-spatial token generation module and several stacked spectral-spatial Mamba blocks to extract the deep and discriminant features of HSI. And the spectral-spatial information interaction and fusion occurs in the token generation process (early stage), spectral-spatial Mamba blocks (middle stage), and mean token addition (last stage).
Figure 1. The illustration for the design details of the proposed SS-Mamba—working flow. For illustrative purposes, a single image flow instead of a batch is shown here. The proposed SS-Mamba mainly consists of a spectral-spatial token generation module and several stacked spectral-spatial Mamba blocks to extract the deep and discriminant features of HSI. And the spectral-spatial information interaction and fusion occurs in the token generation process (early stage), spectral-spatial Mamba blocks (middle stage), and mean token addition (last stage).
Remotesensing 16 02449 g001
Figure 2. The procedure of spectral and spatial token generation. The spatial branch processes the entire region by partitioning it along the spatial dimension after spectral mapping, while the spectral branch focuses on the center region, partitioning it along the spectral dimension after spatial mapping.
Figure 2. The procedure of spectral and spatial token generation. The spatial branch processes the entire region by partitioning it along the spatial dimension after spectral mapping, while the spectral branch focuses on the center region, partitioning it along the spectral dimension after spatial mapping.
Remotesensing 16 02449 g002
Figure 3. The illustration of the spectral-spatial feature enhancement module.
Figure 3. The illustration of the spectral-spatial feature enhancement module.
Remotesensing 16 02449 g003
Figure 4. Indian Pines dataset: (a) true color map (645 nm, 547 nm, 458 nm), (b) ground truth.
Figure 4. Indian Pines dataset: (a) true color map (645 nm, 547 nm, 458 nm), (b) ground truth.
Remotesensing 16 02449 g004
Figure 5. Pavia University dataset: (a) false color map (643 nm, 547 nm, 455 nm), (b) ground truth.
Figure 5. Pavia University dataset: (a) false color map (643 nm, 547 nm, 455 nm), (b) ground truth.
Remotesensing 16 02449 g005
Figure 6. Houston dataset: (a) true color map (650 nm, 550 nm, 450 nm), (b) ground truth.
Figure 6. Houston dataset: (a) true color map (650 nm, 550 nm, 450 nm), (b) ground truth.
Remotesensing 16 02449 g006
Figure 7. Chikusei dataset: (a) false color map (648 nm, 549 nm, 448 nm), (b) ground truth.
Figure 7. Chikusei dataset: (a) false color map (648 nm, 549 nm, 448 nm), (b) ground truth.
Remotesensing 16 02449 g007
Figure 8. Classification maps using different methods on the Indian Pines dataset. (a) SS-Mamba, (b) CT-Mixer, (c) SSFTT, (d) LSFAT, (e) MSSG, (f) DBDA, (g) SSRN, (h) EMP-SVM.
Figure 8. Classification maps using different methods on the Indian Pines dataset. (a) SS-Mamba, (b) CT-Mixer, (c) SSFTT, (d) LSFAT, (e) MSSG, (f) DBDA, (g) SSRN, (h) EMP-SVM.
Remotesensing 16 02449 g008
Figure 9. Classification maps using different methods on the Pavia University dataset. (a) SS-Mamba, (b) CT-Mixer, (c) SSFTT, (d) LSFAT, (e) MSSG, (f) DBDA, (g) SSRN, (h) EMP-SVM.
Figure 9. Classification maps using different methods on the Pavia University dataset. (a) SS-Mamba, (b) CT-Mixer, (c) SSFTT, (d) LSFAT, (e) MSSG, (f) DBDA, (g) SSRN, (h) EMP-SVM.
Remotesensing 16 02449 g009
Figure 10. Classification maps using different methods on the Houston dataset. (a) SS-Mamba, (b) CT-Mixer, (c) SSFTT, (d) LSFAT, (e) MSSG, (f) DBDA, (g) SSRN, (h) EMP-SVM.
Figure 10. Classification maps using different methods on the Houston dataset. (a) SS-Mamba, (b) CT-Mixer, (c) SSFTT, (d) LSFAT, (e) MSSG, (f) DBDA, (g) SSRN, (h) EMP-SVM.
Remotesensing 16 02449 g010
Figure 11. Classification maps using different methods on the Chikusei dataset. (a) SS-Mamba, (b) CT-Mixer, (c) SSFTT, (d) LSFAT, (e) MSSG, (f) DBDA, (g) SSRN, (h) EMP-SVM.
Figure 11. Classification maps using different methods on the Chikusei dataset. (a) SS-Mamba, (b) CT-Mixer, (c) SSFTT, (d) LSFAT, (e) MSSG, (f) DBDA, (g) SSRN, (h) EMP-SVM.
Remotesensing 16 02449 g011
Figure 12. The spatial feature maps of each block. For simplicity of illustration, only the spatial feature maps on the first 8 channels are given. Each row corresponds to the feature maps of different channels, while each column represents the feature maps outputted by different blocks.
Figure 12. The spatial feature maps of each block. For simplicity of illustration, only the spatial feature maps on the first 8 channels are given. Each row corresponds to the feature maps of different channels, while each column represents the feature maps outputted by different blocks.
Remotesensing 16 02449 g012
Table 1. Land cover classes and numbers of samples in the Indian Pines dataset.
Table 1. Land cover classes and numbers of samples in the Indian Pines dataset.
No.Class NameTraining SamplesTest SamplesTotal Samples
1Alfalfa202646
2Corn-notill2014081428
3Corn-mintill20810830
4Corn20217237
5Grass-pasture20463483
6Grass-trees20710730
7Grass-pasture-mowed20828
8Hay-windrowed20458478
9Oats15520
10Soybean-notill20952972
11Soybean-mintill2024352455
12Soybean-clean20573593
13Wheat20185205
14Woods2012451265
15Buildings-Grass-Trees20366386
16Stone-Steel-Towers207393
Total315993410,249
Table 2. Land cover classes and numbers of samples in the Pavia University dataset.
Table 2. Land cover classes and numbers of samples in the Pavia University dataset.
No.Class NameTraining SamplesTest SamplesTotal Samples
1Asphalt2066116631
2Meadows2018,62918,649
3Gravel2020792099
4Trees2030443064
5Mental sheets2013251345
6Bare soil2050095029
7Bitumen2013101330
8Bricks2036623682
9Shadow20927947
Total18042,59642,776
Table 3. Land cover classes and numbers of samples in the Houston dataset.
Table 3. Land cover classes and numbers of samples in the Houston dataset.
No.Class NameTraining SamplesTest SamplesTotal Samples
1Grass-healthy2012311251
2Grass-stressed2012341254
3Grass-synthetic20677697
4Tree2012241244
5Soil2012221242
6Water20305325
7Residential2012481268
8Commercial2012241244
9Road2012321252
10Highway2012071227
11Railway2012151235
12Parking-lot-12012131233
13Parking-lot-220449469
14Tennis-court204008428
15Running-track20640660
Total30014,72915,029
Table 4. Land cover classes and numbers of samples in the Chikusei dataset.
Table 4. Land cover classes and numbers of samples in the Chikusei dataset.
No.Class NameTraining SamplesTest SamplesTotal Samples
1Water528402845
2Bare soil (school)528542859
3Bare soil (park)5281286
4Bare soil (farmland)548474852
5Natural plants542924297
6Weeds511031108
7Forest520,51120,516
8Grass565106515
9Rice field (grown)513,36413,369
10Rice field (first stage)512631268
11Row crops559565961
12Plastic house521882193
13Manmade-1512151220
14Manmade-2576597664
15Manmade-35426431
16Manmade-45217222
17Manmade grass510351040
18Asphalt5796801
19Paved ground5140145
Total9577,49777,592
Table 5. Ablation experiment with different sequence models on the Indian Pines dataset.
Table 5. Ablation experiment with different sequence models on the Indian Pines dataset.
LSTMGRUTransformerMamba
Spectral onlyOA (%)56.9359.4154.5360.95
AA (%)69.8472.0967.5073.31
K × 10051.8154.6549.1256.27
Spatial onlyOA (%)86.0284.3983.6087.40
AA (%)91.1790.1789.4192.35
K × 10084.1582.3181.4385.71
Spectral-spatialOA (%)89.7890.3888.0391.59
AA (%)94.1594.7493.1895.46
K × 10088.2489.0586.3890.42
Table 6. Ablation experiment with different sequence models on the Pavia University dataset.
Table 6. Ablation experiment with different sequence models on the Pavia University dataset.
LSTMGRUTransformerMamba
Spectral onlyOA (%)73.4073.5374.5575.52
AA (%)81.5181.8381.9583.11
K × 10066.3466.5567.7868.88
Spatial onlyOA (%)89.3990.6592.6293.63
AA (%)89.1190.0595.2893.49
K × 10086.2487.8490.3691.67
Spectral-spatialOA (%)95.5195.9594.9996.40
AA (%)97.9797.9197.2698.43
K × 10094.1794.8593.4795.31
Table 7. Ablation experiment with different sequence models on the Houston dataset.
Table 7. Ablation experiment with different sequence models on the Houston dataset.
LSTMGRUTransformerMamba
Spectral onlyOA (%)81.3283.8184.6184.86
AA (%)82.6485.1785.7085.92
K × 10079.8182.5183.3783.51
Spatial onlyOA (%)87.8888.5289.1690.21
AA (%)89.4289.8790.2091.31
K × 10086.9187.6088.2889.42
Spectral-spatialOA (%)93.5093.8193.3894.30
AA (%)94.1794.3293.4894.96
K × 10092.9793.2492.8493.84
Table 8. Ablation experiment with different sequence models on the Chikusei dataset.
Table 8. Ablation experiment with different sequence models on the Chikusei dataset.
LSTMGRUTransformerMamba
Spectral onlyOA (%)68.7370.0068.2878.38
AA (%)79.8081.7382.6885.58
K × 10064.3865.8364.2775.39
Spatial onlyOA (%)92.0193.3093.1393.83
AA (%)92.7693.7793.3793.84
K × 10090.8592.3092.1392.92
Spectral-spatialOA (%)94.3194.3894.2194.97
AA (%)94.1894.2194.0294.83
K × 10093.5993.6293.5494.22
Table 9. Ablation experiment for the feature enhancement module.
Table 9. Ablation experiment for the feature enhancement module.
Indian PinesPavia UniversityHouston
w/w/ow/w/ow/w/o
OA (%)91.5989.0196.4095.9794.3092.21
AA (%)95.4693.3598.4398.0494.9693.18
K × 10090.4287.5195.3194.7593.8491.58
w/: with, w/o: without.
Table 10. Testing data classification results (mean ± standard deviation) on the Indian Pines dataset.
Table 10. Testing data classification results (mean ± standard deviation) on the Indian Pines dataset.
ClassEMP-SVMCNNSSRNDBDAMSSGLSFATSSFTTCT-MixerSS-Mamba
Alfalfa 96.54   ± 2.07100.0 ± 0.00 86.04   ± 9.73 79.35   ± 12.80 98.46   ± 2.55100.0 ± 0.00100.0 ± 0.00100.0 ± 0.00100.0 ± 0.00
Corn-notill 63.74   ± 7.45 78.96   ± 6.22 86.85   ± 6.00 86.89   ± 10.7487.42 ± 6.17 81.13   ± 6.11 85.38   ± 3.92 81.94   ± 3.95 80.30   ± 4.58
Corn-mintill 76.56   ± 4.40 87.75   ± 5.96 86.54   ± 5.65 86.88   ± 8.79 87.80   ± 6.91 86.90   ± 5.32 87.63   ± 5.69 87.47   ± 4.2988.54 ± 5.07
Corn 81.94   ± 5.14 98.53   ± 2.56 73.29   ± 13.10 81.64   ± 11.47 96.22   ± 7.07 99.07   ± 1.23 98.06   ± 2.43 99.35   ± 1.1199.49 ± 0.76
Grass-pasture 86.57   ± 3.85 92.18   ± 3.2198.00 ± 1.96 97.02   ± 3.74 88.70   ± 6.28 92.89   ± 2.90 94.15   ± 2.09 93.24   ± 1.54 93.69   ± 1.99
Grass-trees 92.85   ± 4.63 95.41   ± 2.53 97.79   ± 1.40 96.54   ± 1.25 97.69   ± 1.78 95.35   ± 3.89 97.98   ± 1.26 95.30   ± 3.4798.34 ± 1.53
Grass-pasture-mowed 92.50   ± 6.12100.0 ± 0.00 73.78   ± 18.70 58.08   ± 28.22100.0 ± 0.00100.0 ± 0.00100.0 ± 0.00100.0 ± 0.00100.0 ± 0.00
Hay-windrowed 94.10   ± 3.25 99.80   ± 0.59 99.46   ± 0.83 99.62   ± 0.58 99.93   ± 0.20 99.56   ± 0.97 99.34   ± 1.22 99.98   ± 0.07 100.0   ± 0.00
Oats 98.00   ± 6.00100.0 ± 0.00 44.13   ± 20.56 32.65   ± 11.22100.0 ± 0.00100.0 ± 0.00100.0 ± 0.00100.0 ± 0.00100.0 ± 0.00
Soybean-notill 68.28   ± 7.20 88.93   ± 3.10 79.53   ± 6.45 81.08   ± 8.08 84.79   ± 8.70 87.58   ± 4.8989.66 ± 4.47 87.88   ± 4.44 89.61   ± 3.92
Soybean-mintill 59.22   ± 4.69 87.41   ± 5.47 92.94   ± 3.5095.43 ± 4.19 86.75   ± 8.83 85.26   ± 3.98 84.72   ± 5.24 86.89   ± 3.17 89.51   ± 5.87
Soybean-clean 66.61   ± 7.68 83.84   ± 6.23 82.78   ± 10.73 90.57   ± 13.45 83.58   ± 12.48 83.70   ± 4.94 77.40   ± 4.99 82.57   ± 6.8793.32 ± 4.74
Wheat 97.08   ± 1.81 99.57   ± 0.89 95.49   ± 3.82 91.42   ± 5.35100.0 ± 0.00 99.89   ± 0.32 98.86   ± 2.15 98.97   ± 1.89 99.46   ± 1.62
Woods 86.39   ± 5.68 97.39   ± 1.71 97.87   ± 1.14 98.43   ± 1.0599.32 ± 0.96 97.69   ± 1.18 98.75   ± 1.18 97.07   ± 1.48 98.63   ± 1.40
Buildings-Grass-Trees 71.48   ± 8.55 97.19   ± 2.60 87.30   ± 7.75 88.99   ± 5.4697.32 ± 3.96 95.63   ± 3.74 96.91   ± 2.02 96.97   ± 2.59 96.67   ± 3.16
Stone-Steel-Towers 95.48   ± 4.71 99.18   ± 0.91 79.51   ± 4.48 68.09   ± 12.75 99.32   ± 0.68 98.49   ± 1.67100.0 ± 0.00 99.59   ± 0.63 99.73   ± 0.82
OA (%) 73.32   ± 2.25 89.62   ± 1.72 89.44   ± 1.38 90.26   ± 3.06 90.60   ± 2.03 89.35   ± 1.38 90.10   ± 1.58 89.87   ± 1.2891.59 ± 1.85
AA (%) 82.96   ± 1.31 94.14   ± 0.72 85.08   ± 2.08 83.29   ± 2.47 94.21   ± 1.46 93.95   ± 0.62 94.30   ± 0.83 94.20   ± 0.7295.46 ± 0.90
K × 100 69.93   ± 2.50 88.20   ± 1.93 88.00   ± 1.55 88.95   ± 3.43 89.27   ± 2.30 87.90   ± 1.55 88.76   ± 1.78 88.49   ± 1.4490.42 ± 2.08
Table 11. Testing data classification results (mean ± standard deviation) on the Pavia University dataset.
Table 11. Testing data classification results (mean ± standard deviation) on the Pavia University dataset.
ClassEMP-SVMCNNSSRNDBDAMSSGLSFATSSFTTCT-MixerSS-Mamba
Asphalt 81.27   ± 6.60 90.85   ± 4.53 97.84   ± 1.7198.74 ± 1.20 97.06   ± 3.69 87.16   ± 3.78 86.16   ± 6.15 90.59   ± 5.26 95.70   ± 3.41
Meadows 83.13   ± 3.26 93.83   ± 3.49 97.72   ± 0.8199.51 ± 0.36 92.29   ± 5.91 95.05   ± 3.17 94.50   ± 3.60 94.62   ± 4.88 94.05   ± 4.27
Gravel 81.60   ± 4.51 98.12   ± 1.27 83.71   ± 8.17 90.81   ± 12.1099.97 ± 0.10 93.95   ± 4.43 93.94   ± 4.13 94.39   ± 4.19 99.61   ± 0.54
Trees 95.29   ± 2.44 96.29   ± 1.39 97.70   ± 2.01 92.74   ± 7.92 97.12   ± 1.38 92.76   ± 4.46 88.97   ± 5.25 84.61   ± 6.8098.92 ± 0.55
Mental sheets 99.26   ± 0.26 99.33   ± 0.52 99.86   ± 0.27 99.53   ± 0.63100.0 ± 0.00 98.85   ± 0.89 98.95   ± 1.06 99.18   ± 0.67100.0 ± 0.00
Bare soil 80.27   ± 6.31 99.47   ± 0.63 91.98   ± 3.69 90.96   ± 5.45 99.40   ± 1.58 99.19   ± 0.82 96.07   ± 3.9099.53 ± 1.08 99.19   ± 1.63
Bitumen 93.11   ± 1.56 99.39   ± 0.68 88.49   ± 12.03 93.80   ± 8.83100.0 ± 0.00 99.08   ± 0.76 99.58   ± 0.67 99.34   ± 0.67 99.93   ± 0.20
Bricks 83.86   ± 3.96 98.94   ± 0.80 84.79   ± 7.36 89.83   ± 6.4798.99 ± 1.43 91.74   ± 3.98 86.63   ± 7.99 97.10   ± 2.75 98.50   ± 0.95
Shadow99.85 ± 0.12 96.66   ± 1.33 99.41   ± 0.94 96.82   ± 1.69 99.47   ± 0.93 95.49   ± 2.30 95.54   ± 2.17 93.95   ± 3.14 99.96   ± 0.05
OA (%) 84.53   ± 2.22 95.26   ± 1.74 94.72   ± 1.17 95.87   ± 1.85 95.79   ± 2.55 94.06   ± 1.29 92.61   ± 1.92 94.33   ± 2.8596.40 ± 2.27
AA (%) 88.63   ± 1.57 96.99   ± 0.74 93.50   ± 2.07 94.75   ± 2.36 98.25   ± 0.83 94.81   ± 0.54 93.37   ± 1.24 94.81   ± 1.7798.43 ± 0.77
K × 100 80.00   ± 2.76 93.80   ± 2.23 93.02   ± 1.53 94.58   ± 2.41 94.54   ± 3.24 92.22   ± 1.63 90.29   ± 2.47 92.60   ± 3.6395.31 ± 2.92
Table 12. Testing data classification results (mean ± standard deviation) on the Houston dataset.
Table 12. Testing data classification results (mean ± standard deviation) on the Houston dataset.
ClassEMP-SVMCNNSSRNDBDAMSSGLSFATSSFTTCT-MixerSS-Mamba
Grass-healthy 92.99   ± 4.30 92.59   ± 4.5696.25 ± 2.94 93.48   ± 5.59 93.44   ± 4.63 92.67   ± 4.33 95.36   ± 3.62 90.90   ± 5.13 92.88   ± 4.33
Grass-stressed 93.06   ± 5.72 97.00   ± 2.21 97.65   ± 2.48 95.10   ± 3.78 95.37   ± 2.36 97.57   ± 1.9898.58 ± 1.13 95.11   ± 2.60 95.99   ± 2.91
Grass-synthetic 98.97   ± 1.10 98.66   ± 1.41 99.93   ± 0.22100.0 ± 0.00 98.67   ± 1.26 98.17   ± 1.20 99.20   ± 0.60 97.10   ± 4.20100.0 ± 0.00
Tree 94.75   ± 2.94 97.66   ± 1.81 95.98   ± 4.13 97.13   ± 2.17 97.69   ± 1.74 96.58   ± 1.78 97.82   ± 3.19 92.27   ± 4.2099.17 ± 1.67
Soil 96.51   ± 4.52 97.47   ± 5.11 95.41   ± 2.36 97.66   ± 2.42 98.61   ± 3.22 99.78   ± 0.2999.99 ± 0.02 98.58   ± 3.08 98.29   ± 5.08
Water 94.72   ± 3.42 95.38   ± 3.52 97.31   ± 7.83 97.37   ± 2.23 95.08   ± 3.64 97.47   ± 3.6698.42 ± 3.86 96.42   ± 3.53 96.39   ± 3.63
Residential 85.54   ± 4.67 90.54   ± 2.4192.10 ± 2.47 91.93   ± 3.62 91.46   ± 3.66 86.43   ± 2.43 84.32   ± 5.32 89.18   ± 3.91 91.60   ± 2.06
Commercial 69.36   ± 4.90 78.48   ± 6.64 93.23   ± 3.4994.88 ± 3.19 78.02   ± 5.81 83.25   ± 6.19 80.80   ± 5.19 83.90   ± 4.83 83.33   ± 4.35
Road 75.81   ± 6.81 90.60   ± 3.94 89.87   ± 3.83 88.57   ± 2.2792.26 ± 3.16 82.61   ± 5.52 78.82   ± 6.79 86.58   ± 2.31 92.05   ± 2.02
Highway 87.63   ± 4.01 96.06   ± 4.12 86.49   ± 6.78 89.76   ± 3.83 97.61   ± 2.10 96.50   ± 2.86 95.10   ± 3.3399.09 ± 1.52 98.05   ± 1.94
Railway 85.58   ± 7.72 91.52   ± 4.71 90.45   ± 1.94 95.44   ± 2.10 93.82   ± 5.63 93.46   ± 5.3895.82 ± 3.23 92.34   ± 6.17 92.48   ± 6.14
Parking-lot-1 76.18   ± 6.14 91.81   ± 5.48 89.91   ± 4.7393.15 ± 3.27 92.45   ± 6.26 89.55   ± 5.62 89.65   ± 4.87 91.27   ± 5.74 91.24   ± 5.57
Parking-lot-2 56.44   ± 5.90 96.08   ± 2.62 93.52   ± 5.65 82.75   ± 6.06 95.86   ± 2.48 92.98   ± 4.0197.75 ± 2.10 91.22   ± 5.39 92.87   ± 3.30
Tennis-court 97.94   ± 2.53 99.93   ± 0.22 97.67   ± 3.29 98.13   ± 2.59 99.95   ± 0.15 99.88   ± 0.25 99.85   ± 0.22100.0 ± 0.00100.0 ± 0.00
Running-track 99.08   ± 0.46 99.59   ± 1.06 96.93   ± 1.96 95.05   ± 2.47 99.45   ± 1.59 99.97   ± 0.09100.0 ± 0.00 98.63   ± 2.38100.0 ± 0.00
OA (%) 86.56   ± 1.36 93.36   ± 0.92 93.32   ± 1.05 93.67   ± 0.92 93.92   ± 1.02 92.85   ± 0.93. 92.88   ± 0.97 92.73   ± 1.2194.30 ± 1.10
AA (%) 86.97   ± 1.26 94.23   ± 0.67 94.18   ± 1.12 94.03   ± 0.87 94.65   ± 0.85 93.79   ± 0.74 94.10   ± 0.85 93.51   ± 1.0494.96 ± 0.89
K × 100 85.47   ± 1.47 92.82   ± 0.99 92.78   ± 1.13 93.16   ± 1.00 93.43   ± 1.10 92.27   ± 1.01 92.30   ± 1.05 92.14   ± 1.3193.84 ± 1.20
Table 13. Testing data classification results (mean ± standard deviation) on the Chikusei dataset.
Table 13. Testing data classification results (mean ± standard deviation) on the Chikusei dataset.
ClassEMP-SVMCNNSSRNDBDAMSSGLSFATSSFTTCT-MixerSS-Mamba
Water83.55 ± 10.6092.99 ± 4.4083.51 ± 12.9483.44 ± 13.894.58 ± 4.9594.77 ± 4.8293.86 ± 6.3590.75 ± 5.9696.00 ± 2.62
Bare soil (school)93.83 ± 3.8499.54 ± 0.5398.07 ± 2.0299.65 ± 0.5199.72 ± 0.3299.73 ± 0.2999.38 ± 0.5399.76 ± 0.39100.0 ± 0.00
Bare soil (park)98.01 ± 2.6299.57 ± 0.9828.93 ± 10.7531.63 ± 15.77100.0 ± 0.0099.50 ± 0.9999.03 ± 1.7099.89 ± 0.3299.72 ± 0.85
Bare soil (farmland)50.19 ± 20.782.66 ± 16.1090.14 ± 11.3887.22 ± 10.7383.77 ± 14.781.21 ± 15.682.93 ± 16.1383.00 ± 18.1384.44 ± 17.5
Natural plants96.70 ± 2.7699.95 ± 0.0295.10 ± 3.3296.53 ± 3.2499.99 ± 0.02100.0 ± 0.0099.97 ± 0.0599.49 ± 0.5999.98 ± 0.03
Weeds87.28 ± 12.1395.62 ± 3.6473.53 ± 22.8985.41 ± 24.1495.69 ± 3.7294.89 ± 3.5595.17 ± 3.7595.26 ± 3.8595.01 ± 3.60
Forest82.13 ± 7.4999.97 ± 0.0595.66 ± 3.7099.37 ± 0.8799.96 ± 0.0399.67 ± 0.5496.73 ± 9.2499.92 ± 0.07100.0 ± 0.00
Grass91.93 ± 2.7293.05 ± 2.9996.71 ± 4.9699.90 ± 0.2792.91 ± 3.5291.95 ± 2.2092.40 ± 3.6093.78 ± 3.2597.42 ± 3.12
Rice field (grown)79.34 ± 20.9794.59 ± 10.5896.57 ± 3.7799.43 ± 0.4691.72 ± 12.093.43 ± 1.1288.42 ± 13.3386.40 ± 13.2094.08 ± 11.0
Rice field (first stage)99.26 ± 0.5599.94 ± 0.1781.93 ± 9.8689.73 ± 5.23100.0 ± 0.00100.0 ± 0.00100.0 ± 0.00100.0 ± 0.00100.0 ± 0.00
Row crops66.40 ± 14.4782.22 ± 10.9094.58 ± 11.3797.42 ± 3.2983.82 ± 11.082.52 ± 10.4083.43 ± 14.0085.10 ± 10.4983.96 ± 9.60
Plastic house69.20 ± 11.5084.48 ± 9.1391.50 ± 6.2096.78 ± 4.4886.69 ± 8.3068.08 ± 22.079.03 ± 12.1890.79 ± 10.6385.37 ± 8.23
Manmade-195.09 ± 1.9795.97 ± 1.4896.16 ± 7.6298.75 ± 2.2796.22 ± 1.6696.93 ± 1.1297.15 ± 1.7996.44 ± 1.7896.36 ± 1.62
Manmade-286.85 ± 11.2489.49 ± 10.8099.80 ± 0.3399.60 ± 7.8294.45 ± 6.9692.73 ± 9.5891.19 ± 10.6093.04 ± 8.3892.95 ± 7.95
Manmade-391.01 ± 17.2391.78 ± 8.4393.87 ± 9.6998.12 ± 5.2097.96 ± 4.2096.31 ± 10.9893.73 ± 12.6295.59 ± 10.3893.97 ± 130
Manmade-493.73 ± 7.8595.67 ± 6.0493.60 ± 7.3298.24 ± 3.4894.70 ± 7.9695.48 ± 8.0594.19 ± 4.4994.52 ± 8.0397.33 ± 7.71
Manmade grass93.39 ± 6.3896.06 ± 8.3998.35 ± 1.6596.62 ± 2.3299.71 ± 3.4698.40 ± 4.1299.74 ± 0.75100.0 ± 0.00100.0 ± 0.00
Asphalt88.52 ± 12.1783.98 ± 11.269.53 ± 13.8572.33 ± 13.8285.43 ± 12.2079.53 ± 18.7378.37 ± 19.9576.36 ± 17.4685.14 ± 13.50
Paved ground88.07 ± 7.6998.86 ± 3.4324.50 ± 16.2735.81 ± 35.81100.0 ± 0.0099.93 ± 0.2199.29 ± 0.95100.0 ± 0.00100.0 ± 0.00
OA (%)81.58 ± 4.6493.87 ± 2.2891.46 ± 3.6294.39 ± 2.3994.28 ± 2.6493.37 ± 2.6192.05 ± 3.2693.17 ± 3.2594.97 ± 2.34
AA (%)86.02 ± 3.0693.50 ± 1.4084.32 ± 2.9888.39 ± 2.2494.59 ± 1.6392.90 ± 1.7392.84 ± 1.7193.69 ± 2.1494.83 ± 1.58
K × 10078.97 ± 5.3192.95 ± 2.6090.18 ± 4.1393.55 ± 2.7393.43 ± 3.0192.39 ± 2.9790.87 ± 3.7092.16 ± 3.6994.22 ± 2.67
Table 14. Complexity analysis for different models.
Table 14. Complexity analysis for different models.
CT-Mixer SS-LSTM SS-GRU SS-Transformer SS-Mamba
Param.0.77 M1.00 M0.81 M0.48 M0.47 M
Test Time8.61 ms7.77 ms9.14 ms11.05 ms10.45 ms
OA (%)94.3395.5195.9594.9996.40
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Huang, L.; Chen, Y.; He, X. Spectral-Spatial Mamba for Hyperspectral Image Classification. Remote Sens. 2024, 16, 2449. https://doi.org/10.3390/rs16132449

AMA Style

Huang L, Chen Y, He X. Spectral-Spatial Mamba for Hyperspectral Image Classification. Remote Sensing. 2024; 16(13):2449. https://doi.org/10.3390/rs16132449

Chicago/Turabian Style

Huang, Lingbo, Yushi Chen, and Xin He. 2024. "Spectral-Spatial Mamba for Hyperspectral Image Classification" Remote Sensing 16, no. 13: 2449. https://doi.org/10.3390/rs16132449

APA Style

Huang, L., Chen, Y., & He, X. (2024). Spectral-Spatial Mamba for Hyperspectral Image Classification. Remote Sensing, 16(13), 2449. https://doi.org/10.3390/rs16132449

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop