Next Article in Journal
Research and Design of BPM Shortwave Time Signal Modulation Technology Based on Chirp
Previous Article in Journal
A Cropland Disturbance Monitoring Method Based on Probabilistic Trajectories
Previous Article in Special Issue
An Automated Machine Learning Framework for Adaptive and Optimized Hyperspectral-Based Land Cover and Land-Use Segmentation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

AFA–Mamba: Adaptive Feature Alignment with Global–Local Mamba for Hyperspectral and LiDAR Data Classification

by
Sai Li
1,2 and
Shuo Huang
3,4,*
1
College of Mechanical and Electrical Engineering, Zaozhuang University, Zaozhuang 277160, China
2
Zaozhuang Robot Autonomous Positioning and Navigation Technology Innovation Center, Zaozhuang 277160, China
3
Key Laboratory of Infrared System Detection and Imaging Technology, Chinese Academy of Sciences, Shanghai 200083, China
4
Shanghai Institute of Technical Physics, Chinese Academy of Sciences, Shanghai 200083, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2024, 16(21), 4050; https://doi.org/10.3390/rs16214050
Submission received: 1 September 2024 / Revised: 18 October 2024 / Accepted: 28 October 2024 / Published: 30 October 2024
(This article belongs to the Special Issue Deep Learning for the Analysis of Multi-/Hyperspectral Images II)

Abstract

:
The joint classification of hyperspectral imagery (HSI) and LiDAR data is an important task in the field of remote sensing image interpretation. Traditional classification methods, such as support vector machine (SVM) and random forest (RF), have difficulty capturing the complex spectral–spatial–elevation correlation information. Recently, important progress has been made in HSI-LiDAR classification using Convolutional Neural Networks (CNNs) and Transformers. However, due to the large spatial extent of remote sensing images, the vanilla Transformer and CNNs struggle to effectively capture global context. Moreover, the weak misalignment between multi-source data poses challenges for their effective fusion. In this paper, we introduce AFA–Mamba, an Adaptive Feature Alignment Network with a Global–Local Mamba design that achieves accurate land cover classification. It contains two main core designs: (1) We first propose a Global–Local Mamba encoder, which effectively models context through a 2D selective scanning mechanism while introducing local bias to enhance the spatial features of local objects. (2) We also propose an SSE Adaptive Alignment and Fusion (A2F) module to adaptively adjust the relative positions between multi-source features. This module establishes a guided subspace to accurately estimate feature-level offsets, enabling optimal fusion. As a result, our AFA–Mamba consistently outperforms state-of-the-art multi-source fusion classification approaches across multiple datasets.

1. Introduction

Remote sensing and imaging technologies have developed rapidly in recent years, as more types of data sources have been applied to the extraction and analysis of geomorphological information [1,2,3,4,5,6]. The joint classification of hyperspectral imagery (HSI) and LiDAR data has attracted more and more interest from researchers [7]. Hyperspectral remote sensing is a form of technology that is capable of acquiring fine spectral information, which contains data in tens or even hundreds of spectral bands for achieving accurate identification of the ground [8]. LiDAR technology determines the distance and position of a target object by transmitting laser pulses and measuring their return time, thus obtaining elevation information from the ground [9,10,11]. With the increasing improvements of both forms of imaging technology, HSI-LiDAR joint classification plays an important role in precision agriculture, mineral development [12], environmental monitoring [13], urban planning [14], and other fields. By improving the accuracy of land cover and object classification, this joint method can enhance decision-making processes in resource management, disaster prevention, and land-use planning.
Early remote sensing technologies [15,16] were relatively limited and the type of data acquired was homogenous. As an important field of remote sensing applications, landscape classification and recognition are largely restricted in accuracy and efficiency by their data sources [16]. For example, HSI captures the radiation variations of a target object within a continuous spectral range and combines them with spatial information to achieve an accurate response to the physical and chemical properties of the object [17]. However, in some complex scenarios, different objects in the image may have similar spectral curves, and the lower spatial resolution of HSI makes it challenging to distinguish structures [18]. In such cases, combining elevation information from LiDAR data can enable the accurate identification of different land cover categories [19]. Therefore, studying how to effectively leverage the strengths of both data types for the joint classification of HSI and LiDAR data is significant for raising the accuracy and efficiency of land cover classification.
In past decades, many researchers have focused on transferring modeling methods commonly used in machine learning and pattern recognition to remote sensing image classification tasks [20,21], such as support vector machine (SVM) [22], random forest (RF) [23], Extreme Learning Machine (ELM) [24], etc. In addition, to better utilize spectral–spatial information, researchers have designed a series of ingenious classification models. Hang et al. [25] pioneered a novel approach termed matrix-based spatial spectral feature representation. Wang et al. [26] introduced an advanced approach known as discriminative multi-kernel learning, which learns to determine the optimal combination of kernels from a set of predefined basic kernels by maximizing the divisibility of the reproduced kernels in Hilbert space. It can greatly improve HSI classification performance without strictly limiting the selection of basic kernels. However, these traditional methods often require a certain amount of prior knowledge and complex parameter tuning, which can lead to suboptimal performance when dealing with complicated samples and strongly non-linear data [18].
Recently, Deep Neural Networks (DNNs) have emerged as a dominant force, demonstrating remarkable efficacy in the realm of remote sensing image analysis and processing [8,27,28,29,30]. In particular, Convolutional Neural Networks (CNNs), which are able to accurately distinguish between different feature classes by automatically learning and extracting the deep features of images, greatly improve the discriminability of remote sensing target features. Compared with traditional classification methods based on manual feature extraction, DNNs can handle more complex data as well as adapt to different shooting angles and scale variations [31]. For example, Paoletti et al. [30] proposed a novel convolutional residual pyramid network to achieve fast convergence and high accuracy classification models under complex HSI remote sensing data. Furthermore, Roy et al. [29] introduced a novel superimposed hybrid architecture comprising both 3D-CNN and 2D-CNN, which significantly enhanced the fusion of spectral and spatial information. However, the local receptive field of CNN limits its ability to capture the global information of the spectrum, resulting in suboptimal classification accuracy. To further enhance the model’s global feature extraction capability, researchers [32,33,34,35] introduced the Vision Transformer (ViT) for HSI classification, capturing global spectral information and better representing the relationship between spectral and spatial features. For instance, Mei et al. [35] presented the Group-Aware Hierarchical Transformer. This novel model underscores both local and global interactions of spectral–spatial information by introducing a novel grouped pixel embedding. Although ViT can represent the global dependence of spectra well, its computational quadratic complexity makes it difficult to use for large-scale datasets. Recently, a new method known as Mamba [36] has been introduced, employing state-space models (SSMs) to efficiently capture global semantic information with minimal computational overhead, thus achieving linear complexity operations. Mamba-based methods [37,38,39] have shown great potential in remote sensing classification tasks. For example, RS–Mamba [39] introduced an innovative omnidirectional selective scanning module, which selectively scans remote sensing images in multiple directions. This enables the extraction of large-scale spatial features from various orientations. Zhou’s team proposes a novel centralized Mamba–Cross-Scan (MCS) mechanism for converting HSI images into sequence data to enhance feature generation and focusing for fine-grained recognition of feature categories.
Nevertheless, in recent years, single-source data classification has become increasingly unable to meet the identification needs of complex landform scenarios in terms of data comprehensiveness, credibility, and prediction accuracy. Benefiting from the powerful capabilities of deep learning models in feature extraction and fusion, CNN-based and ViT-based methods show great advantages in the joint classification of HSI and LiDAR data. Hang’s team [40] introduced a coupled CNN model capable of learning distributed spectral–spatial features from HSI while simultaneously capturing elevation information from LiDAR data. This innovative model integrates heterogeneous features through a parameter sharing strategy, marking a substantial departure from conventional approaches. Zhang et al. [41] designed an Interleaved Perceptual Convolutional Neural Network (IP-CNN), which can introduce HSI perception constraints and LiDAR perception constraints into the integration of multi-source structural information, and achieved satisfactory results in small sample training. Lu et al. [42] unveiled a revolutionary method called Coupled Adversarial Learning-based Classification (CALC). Their approach pioneers an adversarial setup between a dual generator and a discriminator. This mechanism adeptly extracts similar higher-order semantic information and modality-specific complementary details, introducing a paradigm shift in classification methodologies. Zhao et al. [43] proposed a new two-branch approach that combines a CNN and Transformer encoder in a hierarchical form to achieve effective joint classification of heterogeneous information from multiple sources. Additionally, Sun and colleagues [44] developed a groundbreaking multi-scale lightweight fusion network based on this, distinctively free from attention mechanisms. This novel architecture not only drastically reduces training parameters but also effectively captures multi-scale depth and high-order features.
Although DNN-based methods have made significant progress in the joint classification of HSI and LiDAR data, the existing methods still face a series of challenges when dealing with complex feature environments [45,46,47]. Initially, given the varied imaging mechanisms of distinct sensors, the model necessitates the separate processing of diverse source data to ensure the effective representation of specific information. Second, the application scenarios of HSI and LiDAR sensors are distinct [48], resulting in different performance focuses for different data types. This requires the training of classification models to fully consider differences and flexibly capture fine-grained local details and global information. Finally, for formal variations in different data features, the classification model also requires a strong adaptive complementary ability to realize the effective fusion and complementarity of spectral–spatial–elevation features, so as to achieve accurate scene classification.
To tackle the aforementioned challenges, we introduce a comprehensive joint classification method for HSI and LiDAR data, termed Adaptive Feature Alignment Network with a Global–Local Mamba. It contains two branches: an HSI branch and a LiDAR feature processing branch. Specifically, we propose an SSE feature extraction module to process HSI and LiDAR data separately to mine more spectral–spatial and elevation features. It can fully extract adaptive feature information while considering data discretization. In addition, to capture semantic information at multiple scales, we also propose Global–Local Mamba modules for the two types of data sources to achieve dynamic awareness of spectral–space–elevation information. Finally, to eliminate the inherent distinction between the expressions of the two types of data in the upper and lower branches, we design an SSE Adaptive Alignment and Fusion ( A 2 F) module that alleviates the feature differences and spatial mismatches between the data sources by learning the discriminative features of the both types of data to adapt to the calibration differences.
To summarize, the main contributions are as follows:
(1)
We propose a joint HSI-LiDAR classification method called Adaptive Feature Alignment Network with a Global–Local Mamba (AFA-Mamba), which uses a hybrid two-branch CNN architecture to accurately extract 3D spectral–spatial information from HSI data while simultaneously capturing 2D elevation information from LiDAR data.
(2)
The proposed Global–Local Mamba module is designed to be both efficient and effective in processing data. It operates by dynamically capturing and analyzing spectral, spatial, and elevation information, allowing the model to adaptively focus on different types of information depending on the context.
(3)
We design a novel spectral–spatial–elevation Adaptive Alignment and Fusion module to adaptively recalibrate the differences by learning the discriminative features of the two types of data, thereby effectively mitigating the problems due to feature differences and spatial mismatches between the data sources. It ensures the accuracy and consistency of HSI and LiDAR information in the fusion process.
(4)
AFA–Mamba demonstrates superior classification performance compared to several existing SOTA methods. The experimental results across all three datasets consistently validate the outstanding performance of our approach.
This paper’s subsequent sections are structured as follows: Section 2 offers a comprehensive overview of our AFA–Mamba, elucidating its core components and operational principles. Moving forward, Section 3 meticulously outlines the experimental datasets, delineates the experimental setup, and conducts a comprehensive analysis of the classification outcomes. Lastly, Section 4 succinctly summarizes the study’s conclusions and paves the way for future research directions.

2. Methodology

Figure 1 illustrates the comprehensive workflow of our proposed AFA–Mamba. It includes HSI and LiDAR Data Preprocessing, a spectral–spatial–elevation (SSE) Feature Extraction Module, a Global–Local Mamba, and an SSE Adaptive Alignment and Fusion ( A 2 F) Module. We introduce the details of each module in the following sections.

2.1. HSI and LiDAR Data Preprocessing

This module is provided with HSI data, denoted as X H R m × n × l , and LiDAR data, denoted as X L R m × n , covering the same area of the Earth’s surface, where m and n represent the spatial dimensions, and l represents the number of spectral bands in the HSI data. The HSI data typically contain numerous spectral bands capable of conveying more valuable information, with each pixel being representable by a one-hot category vector. However, the large size of spectral data leads to expensive computational costs. Therefore, we extracted the first k principal components from HSI through PCA to reduce the number of the spectral band from l to k while keeping its spatial dimensions unchanged, which is defined as X H p c a R m × n × k .
Next, for X H p c a R m × n × k and X L R m × n , we employed a window with a patch size of s × s , performed 3D and 2D patch extraction, and obtained the 3D small patches X H P R s × s × k and 2D small patches X L P R s × s , respectively. We determined the identity of each patch by its central pixel. For edge pixels that may not meet the window size, we performed a padding operation on these pixels and defined the width as ( s 1 ) / 2 . Lastly, we filtered out the pixel blocks with label 0, and then partitioned the remaining samples into training and test sets.

2.2. Spectral–Spatial–Elevation Feature Extraction Module

By exploiting the distinct advantage of CNNs in context modeling and feature extraction, they excel in managing vast remote sensing datasets. CNNs efficiently capture spectral–spatial correlations in HSI data and comprehensively extract elevation details from LiDAR data. Hence, our approach employs a 3D-CNN to extract intricate spectral–spatial features from high-dimensional 3D patches, enabling precise local modeling. Concurrently, we utilize a separate 2D-CNN to focus specifically on extracting elevation features from LiDAR data.
As illustrated in Figure 1, for the HSI data X H p c a R m × n × k , we first used Conv3-D to extract discriminative spectral–spatial features from the HSI data, and convert the spatial dimensions of the generated feature cube into a two-dimensional vector. Subsequently, we utilized Conv2-D to mitigate the redundancy inherent in both spectral and spatial information. Unlike HSI data processing, we employed two Conv2-D convolutions to extract the surface elevation information for the LiDAR data X L R m × n , whose dimensions and convolution kernel size are 16 @ 3 × 3 and 64 @ 3 × 3 , respectively. To speed up the training process and increase non-linear learning capabilities, we added layer normalization and ReLU activation functions after each convolution layer.

2.3. Global–Local Mamba Encoder

In hyperspectral classification tasks, remote sensing images are typically captured by satellites from a top-down perspective, resulting in large spatial features with arbitrary orientations. Therefore, the global context modeling of remote sensing images is crucial for classification tasks. However, relying solely on global features is often insufficient for identifying ground vegetation, as it usually requires local representation enhancement to handle small regions with severe boundary adhesion. Although state-space models (SSMs) have shown promise in long-sequence modeling, they face challenges in effectively combining local invariants and global context in visual data. To address this, we propose Global–Local Mamba, which introduces local bias to enhance the spatial features of local objects.
State-space models (SSMs) are a type of continuous system that map one one-dimensional function or sequence to another through specific implicit latent states. In other words, it is a model that uses latent states to transform and model sequential data. Given a one-dimensional sequence x ( t ) R , which is projected through a hidden state h ( t ) R m to form a new one-dimensional sequence y ( t ) R , the entire system can be represented as follows:
h ( t ) = A h ( t ) + B x ( t ) , y ( t ) = C h ( t ) + D x ( t ) .
where A is the state matrix, which defines how the current hidden state affects its own rate of change. B and C are the input matrix and output matrix, respectively. D is the feed-through (or direct transmission) matrix.
In practice, the continuous system described above should be discretized under the zero-order hold assumption, converting the matrices A and B into their discrete forms for a time scale Δ R + :
A ¯ = exp ( Δ A ) , B ¯ = ( Δ A ) 1 ( exp ( Δ A ) I ) · Δ B .
where Δ is the step size. Therefore, the discretized version of SSMs can be expressed as:
h t = A ¯ h t 1 + B ¯ x t , y t = C h t + D x t .
However, the current system remains static when handling different inputs. To overcome this limitation, Mamba introduces a selective state-space model that allows parameters to adapt based on the input, enhancing selective information processing across sequences. This parameter selection mechanism can be expressed as:
B ¯ = f B ( x t ) , C ¯ = f C ( x t ) , Δ = θ A ( P + f A ( x t ) .
Here, f B ( x t ) , f C ( x t ) , and f A ( x t ) are linear functions that expand features to the dimension of the hidden state. Since SSMs are tailored for long sequences, they have limitations in capturing detailed local information. In addition, VMamba [25] and Vim, propose specific position-aware scanning strategies to preserve the structure of 2D images. However, these directed sequences overlook the visual information within the pixel neighborhood. We explore a Global–Local Mamba Encoder, where global perception is received before focusing on details, thereby compensating for the lack of local information.
As shown in Figure 2, our Global–Local Mamba Encoder adopts the linear and state-space model flow, inspired by the usage of similar structures in Transformers and Mamba. Furthermore, we introduce a component, eg ( L p ) , to enhance local bias and correct the causal relationships between neighborhood data, thereby improving the original SSM by maintaining local 2D dependencies:
h t = A ¯ h t 1 + B ¯ x t , y t = C h t + D x t + eg ( L p ) .
where eg ( L p ) operates independently of the hidden state space.

2.4. SSE Adaptive Alignment and Fusion Module

Efficiently fusing multi-source remote sensing data stands as a pivotal factor in significantly enhancing remote sensing classification performance. However, the fusion of discriminative features from different sources remains a major challenge due to the inherent feature variation and spatial misalignment. To this end, we propose a spectral–spatial–elevation (SSE) Adaptive Alignment and Fusion ( A 2 F) module to adaptively adjust the relative positions between multi-source features. It achieves optimal fusion by building a guided subspace to accurately estimate the feature-level offset.
Figure 3 illustrates the detailed workflow of the SSE A 2 F module. Specifically, we extract modality-invariant features and modality-specific features from both the HSI and elevation feature. To mitigate misalignment in the multi-source feature space, we use the modality-invariant features to predict spatial offsets in the multi-source data. These predicted offsets then adaptively adjust the spatial positions of the modality-specific features to achieve optimal feature alignment.
For each input of the spectral–spatial–elevation features F h s i R H × W × C and F l i d a r R H × W × C , we fuse their modality-invariant features to form a guided subspace C = [ C h s i , C l i d a r ] . Then, through channel attention projection to this guided subspace, we amplify the information-rich bands while reducing the influence of irrelevant bands.
C = C o n c a t ( F h s i , F l i d a r ) , C ^ = C m ( [ C h s i , C l i d a r ] ; θ c ) .
where C m is the channel attention operation. θ c denotes the assigning of a separate parameter to each feature.
To learn the spatial shift between modalities from the guided subspace as a strong prior for subsequent alignment and fusion, we use deformable convolutions to achieve implicit offset compensation. Taking into account the varying contributions of each spectral and sub-band region to classification, we first introduce a modulation scalar Δ t k learned from C ^ , which dynamically aggregates information from the surrounding area of the corresponding position p. As depicted in Figure 3, given the center-sampled value x ( p ) in the modality-specific feature F s hsi or F s lidar and the modulation scalar Δ t k , the corresponding value y ( p ) in the aligned feature F h s i l i d a r can be derived as follows:
y ( p ) = k = 1 K w k · x ( p + p k + Δ p k ) · Δ t k
where K, p k , and w k represent the number of kernel weights, and the fixed offset of the k-th position, respectively. Finally, the two aligned features F h s i l i d a r and F l i d a r h s i are fused and sent to the classification head.

2.5. Classification Block

Following the SSE A 2 F module processing, the output is passed through a multi-layer perceptron (MLP) layer for the final classification. This MLP comprises two linear layers featuring Gaussian Error Linear Units (GELUs). Notably, the last linear layer integrates a softmax function to derive conclusive labels for classification.

3. Experiment and Analysis

In this section, we elaborate on the various configurations and results of our experiments. First, we give the specific dataset, experimental configurations, and evaluation indicators for fair comparison. Then, we conduct quantitative experiments on three representative multi-modal datasets to demonstrate the advanced performance of our AFA–Mamba. In addition, comprehensive ablation studies are performed to explain the role of each component in the proposed AFA–Mamba. Finally, the quantitative and visual results demonstrate the superior performance of our AFA–Mamba over existing SOTA methods in remote sensing classification tasks.

3.1. Dataset Description

Following the previous methods [28,42], we select three datasets that are widely utilized for remote sensing classification tasks: MUUFL Gulfport, Trento, and the Augsburg datasets, which contain include two types of data (HSI and LiDAR). Table 1 reports the details of all datasets used, including the number of samples and the land cover ground truth categories.
(1) MUUFL Gulfport Dataset: Gathered at the University of Southern Mississippi Gulf Park campus, this dataset utilizes a reflective optical system spectrometer sensor for imaging. The dataset integrates both hyperspectral imaging (HSI) and LiDAR data. The HSI data feature 72 spectral bands spanning from 0.38 to 1.05 μ m. Additionally, the LiDAR data are composed of two rasters that function at a wavelength of 1.06 μ m. The MUUFL Gulfport dataset has a pixel size of 325 × 220 and covers 11 distinct land cover categories. Due to the significant noise present in the first and last 8 spectral bands, we deleted these bands during training to improve data quality. Figure 4 presents a visual representation of the MUFFL dataset, including pseudo-color composite images (HSI data), grayscale images (LiDAR data), and depictions of the various land cover categories within the dataset.
(2) Trento Dataset: This dataset originates from the rural area surrounding Trento, a city in southern Italy, and includes six distinct scenes. The hyperspectral imaging data and LiDAR data were acquired using the AISA Eagle system’s hyperspectral imaging Eagle sensor and the Optech ALTM 3100EA sensor, respectively. Both sensors provide a spatial resolution of 1 m, with a pixel size of 600 × 166 . The HSI data includes 63 spectral channels spanning from 402.89 nm to 989.09 nm, offering a spectral resolution between 0.42 μ m and 0.99 μ m. The LiDAR data are composed of a single raster. Figure 5 displays the image types for the HSI and LiDAR data and the ground truth land cover categories.
(3) Augsburg Dataset: The Augsburg dataset captures land cover in the German city of Augsburg by integrating data from hyperspectral imaging and LiDAR sources. Both data sources use 30 m spatial resolution. The HSI data contain 180 spectral bands with wavelengths ranging from 0.4 μ m (UV and visible) to 2.5 μ m (NIR). The LiDAR data consist of a single raster layer, both captured at a wavelength of 1.06 μ m. As shown in Figure 6, it has a pixel size of 332 × 485 and consists of seven different categories.

3.2. Experimental Setting

(1) Evaluation Metrics: To assess the performance of our proposed AFA–Mamba, we utilized four commonly used evaluation metrics for a fair comparison: average accuracy (AA), overall accuracy (OA), Kappa coefficient (k), and per-class accuracy. We want the score of each metric to be as high as possible, which means the classification accuracy of our model is more accurate. To ensure fair comparisons, all experiments were conducted using separate training and test sets.
(2) Environment Configuration: We implemented our AFA–Mamba by PyTorch 2.2.0. Training was conducted using the Adam optimizer with the parameters β 1 = 0.9 and β 2 = 0.999 . The model trained for 100 epochs, with a batch size of 64 and an initial learning rate of 1 × 10 3 . The training process utilized an NVIDIA Geforce RTX 4070ti 16 GB GPU. Regarding the traditional method among the comparison methods, we completed the experiment on the MATLAB platform.
Parameter Setting Adjustment: (1) Patch Size: For joint HSI and LiDAR data classification methods, choosing an appropriate patch size means that the model can consider more neighborhood pixels with limited computational cost, which may better capture the spatial distribution, band changes, and elevation information characteristics of ground objects. In the dataset presented above, we fixed all parameter settings except patch size. In addition, in order to select the appropriate patch size, we evaluated the accuracy effect of the patch sizes in the set {7, 9, 11, 13, 15, 17} in sequence. As shown in Figure 7, an excessively large patch size will cause the method, especially the scalar fusion module, to have more complex inputs, resulting in a reduced network fitting effect and thus reduced accuracy. A patch size that is too small will result in insufficient data context information, thereby affecting global information retention and resulting in insufficient accuracy. It can be seen from this experiment that a patch size of 11 is the most appropriate.
Although HSI data contain hundreds of continuous spectral bands and provide rich ground object information, due to their wide range of spectral imaging and high response characteristics, many redundant bands are often retained in the image, which can easily cause dimensional disaster. Therefore, we utilized PCA to retain the most important spectral data to achieve a balance between spectral expression ability and computational efficiency. As shown in Figure 8, too small a number of retained spectra will result in a large loss of spectral information, and low-resolution spatial data and elevation information cannot support sufficient classification accuracy. However, an excessively large number of retained spectra will introduce a certain amount of redundant noise, accompanied by high computational costs, resulting in a decrease in classification accuracy. We selected the number of retained spectral bands from the set {5, 10, 15, 20, 25, 30, 35, 40} for accuracy effect evaluation. When we retained the main 30 bands of HSI data, they could be effectively combined with LiDAR data to improve classification accuracy.
Learning rate: The learning rate controls the magnitude of the weight updates during model training, influencing how quickly or slowly the model learns. A higher learning rate can lead to oscillations during training, potentially hindering the model’s ability to converge consistently. On the contrary, a lower learning rate will slow down the training process, leading to increased training time and higher computational cost. Therefore, in the hyperspectral image classification task, an appropriate learning rate helps to stabilize the fitting effect of the training process, thus affecting the final classification performance. We selected the learning rate from the set {1 × 10 5 , 5 × 10 5 , 1 × 10 4 , 5 × 10 4 , 1 × 10 3 , 5 × 10 3 } for accuracy effect evaluation. As shown in Figure 9, it can be seen from this experiment that for the MUUFL Gulfport and Augsburg datasets, setting the learning rate to 5 × 10 4 and the Trento dataset to 1 × 10 3 can achieve better classification results.

3.3. Ablation Study

To assess the impact of each component within AFA–Mamba, a series of experiments was conducted using the MUFFL dataset. The sample setup is shown in Table 1. The experiments included PCA, the spectral–spatial–elevation feature extraction module (SSE-FE), the HSI-Global–Local Mamba module (HSI-GL–Mamba), the LiDAR–Global–Local Mamba module (LiDAR-GL–Mamba), and the Adaptive Alignment and Fusion Module ( A 2 F) analysis. Table 2 reports the classification performance resulting from combining different modules.
No PCA was used in Case 1, and the raw spectra of the data were directly fed into the model. Due to the massive amount of invalid band information brought in, the accuracy of the model degraded severely. The OA, AA, and K accuracies all degraded by more than 3%. This illustrates the importance of filtering bands for de-redundancy in hyperspectral tasks.
Case 2 replaced the SSE-FE module with a regular 2D-CNN, which coarsely extracted spectral–spatial–elevation information. The OA, AA, and K accuracies were somewhat affected, decreasing by 1.86%, 1.57%, and 1.55. This illustrates the improvement of our designed feature enhancement extraction module with respect to a single CNN.
Case 3 removed the SSE-FE module and directly used the most basic patch for the features of two different types of data. We can observe that the model extraction ability dropped significantly, which is reflected in three important indicators, among which the OA, AA, and K accuracy all dropped by more than 2%. This illustrates the importance of the SSE-FE module in improving classification accuracy.
Case 4 and Case 5 eliminated the HSI-GL–Mamba and LiDAR-GL–Mamba modules, respectively, which are of great significance in our feature retention. Both modules can differentiate between global and local information based on the feature map. Here, there is only one serial convolutional branch for replacement, which does not allow for the efficient extraction of more comprehensive features from HSI and LiDAR data. Therefore, the accuracy of the three classifications dropped significantly. This demonstrates that the HSI-GL–Mamba and LiDAR-GL–Mamba modules explore deep spectral–spatial–elevation semantic information, facilitating the effective integration and interaction of features across multiple branches.
Case 6 removed the A 2 F module and adds the spectral–spatial and elevation features directly into the classifier. It is obvious that although the Mamba structure is able to obtain sufficiently rich features, due to the inherent differences and spatial offsets in the two features, direct addition cannot effectively fuse the discriminative features of these features, thus leading to a decrease in classification accuracy. In Case 7, the A 2 F module we proposed was able to eliminate this challenge, thus achieving the optimal classification accuracy and greatly improving the three classification indicators.
Cases 7 and 8 replace the GL–Mamba of the upper-branch HSI and the GL–Mamba of the lower-branch LiDAR with the original Mamba. It is evident that while the original Mamba structure can capture global contextual information, it lacks the local bias necessary to refine spatial details. As a result, the classification accuracy suffers, as the model is unable to fully exploit the fine-grained spatial and spectral variations present in both data sources.

3.4. Classification Result and Analysis

To demonstrate the superiority of our proposed AFA–Mamba compared to other advanced methods, we carefully selected some representative and excellent classification methods and divided them into two categories: the HSI-based classification methods, including RF [20], SVM [21], 2D-CNN [49], HybridSN [29], GAHT [35], and MiM [50], and the joint HSI and LiDAR data fusion classification methods, including CoupledCNN [40], CALC [42], HCTnet [43], M2FNet [44], and HLMamba [51]. For a fair comparison, the default parameters of all methods followed the corresponding references, and the training set partitioning and other parameters were consistent with the configuration described above. We repeated the experiment ten times to obtain its mean and variance.
(1) Quantitative Results and Analysis:
Table 3, Table 4 and Table 5 report the quantitative comparison results of our AFA–Mamba against other advanced methods across the three datasets. We conducted each experiment 10 times and calculated the average and standard deviation of the results to ensure the reliability and fairness of the comparison. In the table, the best results are highlighted in bold red. Our method consistently achieved the highest classification performance across all datasets for the OA, AA, and Kappa metrics.
For example, Table 3 provides the detailed results of each comparison method alongside our proposed AFA–Mamba using the MUUFL Gulfport dataset. Our method achieves the highest performance in all aspects. In contrast to traditional approaches like RF and SVM, deep learning methods can learn a wider range of features to improve classification accuracy. The 2D-CNN, HyBridSN, GAHT, and MiM methods based on HSI data classification extract effective features in HSI data, but their OA results are 1.32%, 2.84%, 4.00%, and 1.48% lower than our AFA–Mamba, respectively. The main reason for this is that their limited receptive fields ignore global features. Moreover, our method outperforms joint HSI and LiDAR data classification approaches in key metrics such as OA, AA, and k. It also demonstrates competitive performance in the average accuracy of each class. Specifically, our AFA–Mamba surpasses CoupledCNN, CALC, HCTnet, M2FNet, and HLMamba in AA by 2.53%, 4.79%, 3.58%, 2.57%, and 1.1%, respectively. This is attributed to the multiple receptive fields of the local–global branch of our method and the efficient spectral–elevation feature correction fusion strategy. Table 4 and Table 5 report the results of all compared methods on the Trento and Augsburg datasets. We can draw similar conclusions as above from the two tables. Obviously, our proposed method still maintains an advanced level in all aspects.
The Trento dataset has a small number of categories and an uneven distribution of sample numbers. Therefore, the OA and Kappa of each method achieved high accuracy, while AA had greater room for improvement. It can be observed from Table 4 that the joint classification methods of HSI and LiDAR data achieved good results, especially HCTnet and M2FNet. The dual Transformer encoder branch used by HCTnet better combines the characteristics of the two types of remote sensing data. However, due to its more complex fusion mode, it is prone to under-fitting in the Trento dataset with a small number of samples. As for our proposed AFA–Mamba, the MFR module can better retain the detailed information of the data even in scenarios where the number of samples is insufficient. Therefore, our OA and Kappa are 0.56% and 0.75% higher than HCTnet, respectively. In addition, M2FNet benefits from its multi-scale feature extraction and retains more spectral–spatial–elevation information, and the Global–Local Mamba modules we proposed fully enhance the awareness of global–local information, so it is ultimately better than M2FNet and achieved OA, AA, and Kappa values that are 0.77%, 0.19%, and 1.03% higher, respectively.
The Augsburg dataset has a high spatial resolution and complex ground object information, which places higher requirements on the model’s local–global information balance capability. In addition, due to the addition of LiDAR data elevation information, the joint classification method has additional advantages over the single-source data classification methods. Specifically, as shown in Table 5, CALC uses dual adversarial networks to conduct adversarial training on ground object space and achieves high OA accuracy. This is due to the fact that the adversarial strategy is conducive to the effective combination of space and elevation information. However, our proposed SFCF module achieves effective correction of the fused data by introducing additional offsets, which enhances the coupling degree of spatial–elevation information. Therefore, it is still 0.12%, 1.36%, and 0.18% higher than the CALC method in terms of OA, AA, and Kappa, respectively. Similarly, compared with other methods, the efficient and stable fusion caused by the offset correction of our method can also be proven to be more robust. In summary, by comparing with current classic and advanced methods in three different types of datasets, it is proved that our AFA–Mamba has efficient joint classification performance.
(2) Visual Evaluation and Analysis: Figure 10, Figure 11 and Figure 12 show that the visualization results of various methods can be used for qualitative comparison. From the results, we can significantly observe the classification differences between different methods. It is worth mentioning that our AFA–Mamba can generate more accurate noise-free feature classification maps.
Specifically, the classification method that only uses HSI data has unclear boundaries and introduces more noise. The multi-source data joint classification method significantly reduces this phenomenon, but performs poorly in some ground areas. On the other hand, the classification results obtained by our method have clear boundaries and high classification accuracy. For example, the visualization results of the MUUFL Gulfport dataset are presented in Figure 10, and it can be easily seen that most methods have blur and a lot of noise when distinguishing small and dense areas of multiple categories, while the classification map of our AFA–Mamba is closer to the real ground truth.
Figure 11 illustrates the visualization results of the comparative method in the Terno dataset. This dataset has larger ground objects and fewer categories than the MUUFL Gulfport dataset, so it is easier to classify. We can clearly see that most comparison methods still have the problem of introducing a large amount of noise, while our method still preserves high-precision classification effects. In addition, the classification results of the Augsburg dataset with higher resolution are depicted in Figure 12; our AFA–Mamba still maintains the optimal classification performance in denser scenes with more categories.

4. Conclusions

In this paper, we propose a novel and efficient Adaptive Feature Alignment Network with a Global–Local Mamba (AFA–Mamba) to achieve the efficient fusion of spectral–spatial and elevation features, greatly improving the accuracy of remote sensing classification tasks. Specifically, we propose a novel SSE feature extraction module to explore deep spectral–spatial–elevation semantic information through multi-branch feature extraction. In addition, Global–Local Mamba modules are proposed to enhance ground objects that are sensitive to spectral and elevation information. In order to eliminate the representation differences and spatial misalignment of multi-source features, we propose the SSE Adaptive Alignment and Fusion ( A 2 F) module to effectively learn the discriminative features of heterogeneous data and achieve the adaptive calibration of spatial differences. Extensive experiments demonstrate the advancement and robustness of our method.

Author Contributions

Conceptualization, S.L. and S.H.; methodology, S.L. and S.H.; software, S.L.; formal analysis, S.L.; investigation, S.H.; data curation, S.H.; writing—original draft preparation, S.L.; writing—review and editing, S.H.; visualization, S.L. and S.H.; supervision, S.L. and S.H.; project administration, S.L.; funding acquisition, S.L. and S.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Natural Science Foundation of Shandong Province (Grant number: ZR202103050458); and the start-up foundation for doctoral research of Zaozhuang University (Grant number: 1020714).

Data Availability Statement

All relevant data are within the paper.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Kahraman, S.; Bacher, R. A comprehensive review of hyperspectral data fusion with lidar and sar data. Annu. Rev. Control 2021, 51, 236–253. [Google Scholar] [CrossRef]
  2. Zhou, Z.; Zheng, C.; Liu, X.; Tian, Y.; Chen, X.; Chen, X.; Dong, Z. A dynamic effective class balanced approach for remote sensing imagery semantic segmentation of imbalanced data. Remote Sens. 2023, 15, 1768. [Google Scholar] [CrossRef]
  3. Wang, J.; Hu, J.; Liu, Y.; Hua, Z.; Hao, S.; Yao, Y. El-nas: Efficient lightweight attention cross-domain architecture search for hyperspectral image classification. Remote Sens. 2023, 15, 4688. [Google Scholar] [CrossRef]
  4. Su, Z.; Wan, G.; Zhang, W.; Guo, N.; Wu, Y.; Liu, J.; Cong, D.; Jia, Y.; Wei, Z. An Integrated Detection and Multi-Object Tracking Pipeline for Satellite Video Analysis of Maritime and Aerial Objects. Remote Sens. 2024, 16, 724. [Google Scholar] [CrossRef]
  5. Zhang, G.; Fang, W.; Zheng, Y.; Wang, R. SDBAD-Net: A spatial dual-branch attention dehazing network based on meta-former paradigm. IEEE Trans. Circuits Syst. Video Technol. 2023, 34, 60–70. [Google Scholar] [CrossRef]
  6. Fang, W.; Zhang, G.; Zheng, Y.; Chen, Y. Multi-Task Learning for UAV Aerial Object Detection in Foggy Weather Condition. Remote Sens. 2023, 15, 4617. [Google Scholar] [CrossRef]
  7. Kuras, A.; Brell, M.; Rizzi, J.; Burud, I. Hyperspectral and lidar data applied to the urban land cover machine learning and neural-network-based classification: A review. Remote Sens. 2021, 13, 3393. [Google Scholar] [CrossRef]
  8. Audebert, N.; Le Saux, B.; Lefèvre, S. Deep learning for classification of hyperspectral data: A comparative review. IEEE Geosci. Remote Sens. Mag. 2019, 7, 159–173. [Google Scholar] [CrossRef]
  9. Li, N.; Ho, C.P.; Wang, I.T.; Pitchappa, P.; Fu, Y.H.; Zhu, Y.; Lee, L.Y.T. Spectral imaging and spectral LIDAR systems: Moving toward compact nanophotonics-based sensing. Nanophotonics 2021, 10, 1437–1467. [Google Scholar] [CrossRef]
  10. Khodadadzadeh, M.; Li, J.; Prasad, S.; Plaza, A. Fusion of hyperspectral and LiDAR remote sensing data using multiple feature learning. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2015, 8, 2971–2983. [Google Scholar] [CrossRef]
  11. Ghamisi, P.; Benediktsson, J.A.; Phinn, S. Land-cover classification using both hyperspectral and LiDAR data. Int. J. Image Data Fusion 2015, 6, 189–215. [Google Scholar] [CrossRef]
  12. Murphy, R.J.; Taylor, Z.; Schneider, S.; Nieto, J. Mapping clay minerals in an open-pit mine using hyperspectral and LiDAR data. Eur. J. Remote Sens. 2015, 48, 511–526. [Google Scholar] [CrossRef]
  13. Voss, M.; Sugumaran, R. Seasonal effect on tree species classification in an urban environment using hyperspectral data, LiDAR, and an object-oriented approach. Sensors 2008, 8, 3020–3036. [Google Scholar] [CrossRef]
  14. Liu, L.; Coops, N.C.; Aven, N.W.; Pang, Y. Mapping urban tree species using integrated airborne hyperspectral and LiDAR remote sensing data. Remote Sens. Environ. 2017, 200, 170–182. [Google Scholar] [CrossRef]
  15. Maxwell, A.E.; Warner, T.A.; Fang, F. Implementation of machine-learning classification in remote sensing: An applied review. Int. J. Remote Sens. 2018, 39, 2784–2817. [Google Scholar] [CrossRef]
  16. Gómez-Chova, L.; Tuia, D.; Moser, G.; Camps-Valls, G. Multimodal classification of remote sensing images: A review and future directions. Proc. IEEE 2015, 103, 1560–1584. [Google Scholar] [CrossRef]
  17. Li, S.; Song, W.; Fang, L.; Chen, Y.; Ghamisi, P.; Benediktsson, J.A. Deep learning for hyperspectral image classification: An overview. IEEE Trans. Geosci. Remote Sens. 2019, 57, 6690–6709. [Google Scholar] [CrossRef]
  18. Kumar, B.; Dikshit, O.; Gupta, A.; Singh, M.K. Feature extraction for hyperspectral image classification: A review. Int. J. Remote Sens. 2020, 41, 6248–6287. [Google Scholar] [CrossRef]
  19. Ghamisi, P.; Höfle, B.; Zhu, X.X. Hyperspectral and LiDAR data fusion using extinction profiles and deep convolutional neural network. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2016, 10, 3011–3024. [Google Scholar] [CrossRef]
  20. Ham, J.; Chen, Y.; Crawford, M.M.; Ghosh, J. Investigation of the random forest framework for classification of hyperspectral data. IEEE Trans. Geosci. Remote Sens. 2005, 43, 492–501. [Google Scholar] [CrossRef]
  21. Melgani, F.; Bruzzone, L. Classification of hyperspectral remote sensing images with support vector machines. IEEE Trans. Geosci. Remote Sens. 2004, 42, 1778–1790. [Google Scholar] [CrossRef]
  22. Pal, M.; Mather, P.M. Support vector machines for classification in remote sensing. Int. J. Remote Sens. 2005, 26, 1007–1011. [Google Scholar] [CrossRef]
  23. Belgiu, M.; Drăguţ, L. Random forest in remote sensing: A review of applications and future directions. ISPRS J. Photogramm. Remote Sens. 2016, 114, 24–31. [Google Scholar] [CrossRef]
  24. Huang, F.; Lu, J.; Tao, J.; Li, L.; Tan, X.; Liu, P. Research on optimization methods of ELM classification algorithm for hyperspectral remote sensing images. IEEE Access 2019, 7, 108070–108089. [Google Scholar] [CrossRef]
  25. Hang, R.; Liu, Q.; Song, H.; Sun, Y. Matrix-based discriminant subspace ensemble for hyperspectral image spatial–spectral feature fusion. IEEE Trans. Geosci. Remote Sens. 2015, 54, 783–794. [Google Scholar] [CrossRef]
  26. Wang, Q.; Gu, Y.; Tuia, D. Discriminative multiple kernel learning for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2016, 54, 3912–3927. [Google Scholar] [CrossRef]
  27. Tejasree, G.; Agilandeeswari, L. An extensive review of hyperspectral image classification and prediction: Techniques and challenges. Multimed. Tools Appl. 2024, 83, 80941–81038. [Google Scholar] [CrossRef]
  28. Ahmad, M.; Shabbir, S.; Roy, S.K.; Hong, D.; Wu, X.; Yao, J.; Khan, A.M.; Mazzara, M.; Distefano, S.; Chanussot, J. Hyperspectral image classification—Traditional to deep models: A survey for future prospects. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 15, 968–999. [Google Scholar] [CrossRef]
  29. Roy, S.K.; Krishna, G.; Dubey, S.R.; Chaudhuri, B.B. HybridSN: Exploring 3-D–2-D CNN Feature Hierarchy for Hyperspectral Image Classification. IEEE Geosci. Remote Sens. Lett. 2020, 17, 277–281. [Google Scholar] [CrossRef]
  30. Paoletti, M.E.; Haut, J.M.; Fernandez-Beltran, R.; Plaza, J.; Plaza, A.J.; Pla, F. Deep pyramidal residual networks for spectral–spatial hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2018, 57, 740–754. [Google Scholar] [CrossRef]
  31. Paoletti, M.E.; Haut, J.M.; Plaza, J.; Plaza, A. Deep learning classifiers for hyperspectral imaging: A review. ISPRS J. Photogramm. Remote Sens. 2019, 158, 279–317. [Google Scholar] [CrossRef]
  32. Hong, D.; Han, Z.; Yao, J.; Gao, L.; Zhang, B.; Plaza, A.; Chanussot, J. SpectralFormer: Rethinking hyperspectral image classification with transformers. IEEE Trans. Geosci. Remote Sens. 2021, 60, 1–15. [Google Scholar] [CrossRef]
  33. Zou, J.; He, W.; Zhang, H. Lessformer: Local-enhanced spectral-spatial transformer for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–16. [Google Scholar] [CrossRef]
  34. Yuan, D.; Yu, D.; Qian, Y.; Xu, Y.; Liu, Y. S2Former: Parallel Spectral–Spatial Transformer for Hyperspectral Image Classification. Electronics 2023, 12, 3937. [Google Scholar] [CrossRef]
  35. Mei, S.; Song, C.; Ma, M.; Xu, F. Hyperspectral image classification using group-aware hierarchical transformer. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–14. [Google Scholar] [CrossRef]
  36. Gu, A.; Dao, T. Mamba: Linear-time sequence modeling with selective state spaces. arXiv 2023, arXiv:2312.00752. [Google Scholar]
  37. Li, Y.; Luo, Y.; Zhang, L.; Wang, Z.; Du, B. MambaHSI: Spatial–Spectral Mamba for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2024, 62, 1–16. [Google Scholar] [CrossRef]
  38. Ma, X.; Zhang, X.; Pun, M.O. RS3Mamba: Visual State Space Model for Remote Sensing Image Semantic Segmentation. IEEE Geosci. Remote Sens. Lett. 2024, 21. [Google Scholar] [CrossRef]
  39. Zhao, S.; Chen, H.; Zhang, X.; Xiao, P.; Bai, L.; Ouyang, W. Rs-mamba for large remote sensing image dense prediction. arXiv 2024, arXiv:2404.02668. [Google Scholar] [CrossRef]
  40. Hang, R.; Li, Z.; Ghamisi, P.; Hong, D.; Xia, G.; Liu, Q. Classification of hyperspectral and LiDAR data using coupled CNNs. IEEE Trans. Geosci. Remote Sens. 2020, 58, 4939–4950. [Google Scholar] [CrossRef]
  41. Zhang, M.; Li, W.; Tao, R.; Li, H.; Du, Q. Information fusion for classification of hyperspectral and LiDAR data using IP-CNN. IEEE Trans. Geosci. Remote Sens. 2021, 60, 1–12. [Google Scholar] [CrossRef]
  42. Lu, T.; Ding, K.; Fu, W.; Li, S.; Guo, A. Coupled adversarial learning for fusion classification of hyperspectral and LiDAR data. Inf. Fusion 2023, 93, 118–131. [Google Scholar] [CrossRef]
  43. Zhao, G.; Ye, Q.; Sun, L.; Wu, Z.; Pan, C.; Jeon, B. Joint classification of hyperspectral and LiDAR data using a hierarchical CNN and transformer. IEEE Trans. Geosci. Remote Sens. 2022, 61, 1–16. [Google Scholar] [CrossRef]
  44. Sun, L.; Wang, X.; Zheng, Y.; Wu, Z.; Fu, L. Multiscale 3-D–2-D Mixed CNN and Lightweight Attention-Free Transformer for Hyperspectral and LiDAR Classification. IEEE Trans. Geosci. Remote Sens. 2024, 62, 1–16. [Google Scholar] [CrossRef]
  45. Mäyrä, J.; Keski-Saari, S.; Kivinen, S.; Tanhuanpää, T.; Hurskainen, P.; Kullberg, P.; Poikolainen, L.; Viinikka, A.; Tuominen, S.; Kumpula, T.; et al. Tree species classification from airborne hyperspectral and LiDAR data using 3D convolutional neural networks. Remote Sens. Environ. 2021, 256, 112322. [Google Scholar] [CrossRef]
  46. Mohla, S.; Pande, S.; Banerjee, B.; Chaudhuri, S. Fusatnet: Dual attention based spectrospatial multimodal fusion network for hyperspectral and lidar classification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Seattle, WA, USA, 14–19 June 2020; pp. 92–93. [Google Scholar]
  47. Wang, X.; Feng, Y.; Song, R.; Mu, Z.; Song, C. Multi-attentive hierarchical dense fusion net for fusion classification of hyperspectral and LiDAR data. Inf. Fusion 2022, 82, 1–18. [Google Scholar] [CrossRef]
  48. Dong, W.; Zhang, T.; Qu, J.; Xiao, S.; Zhang, T.; Li, Y. Multibranch feature fusion network with self-and cross-guided attention for hyperspectral and LiDAR classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–12. [Google Scholar] [CrossRef]
  49. Zhao, W.; Du, S. Spectral–spatial feature extraction for hyperspectral image classification: A dimension reduction and deep learning approach. IEEE Trans. Geosci. Remote Sens. 2016, 54, 4544–4554. [Google Scholar] [CrossRef]
  50. Zhou, W.; Kamata, S.I.; Wang, H.; Wong, M.S.; Hou, H.C. Mamba-in-Mamba: Centralized Mamba-Cross-Scan in Tokenized Mamba Model for Hyperspectral Image Classification. Neurocomputing 2024, 613, 128751. [Google Scholar] [CrossRef]
  51. Liao, D.; Wang, Q.; Lai, T.; Huang, H. Joint Classification of Hyperspectral and LiDAR Data Based on Mamba. IEEE Trans. Geosci. Remote Sens. 2024, 62, 1–15. [Google Scholar] [CrossRef]
Figure 1. The architecture of our proposed AFA–Mamba. The proposed method consists of two branches: the upper branch performs fine feature extraction for the spectral–spatial information from HSI, and the lower branch performs correlation feature extraction for the elevation information from LiDAR. Finally, the feature level of the multi-source information is calibrated by the A 2 F module to realize the high-precision classification of the landform categories.
Figure 1. The architecture of our proposed AFA–Mamba. The proposed method consists of two branches: the upper branch performs fine feature extraction for the spectral–spatial information from HSI, and the lower branch performs correlation feature extraction for the elevation information from LiDAR. Finally, the feature level of the multi-source information is calibrated by the A 2 F module to realize the high-precision classification of the landform categories.
Remotesensing 16 04050 g001
Figure 2. Illustration of Global–Local Mamba Encoder. DWConv denotes Depth-Wise Convolution and SiLU denotes the Sigmoid Linear Unit function. Also, A is the state matrix, B is the input matrix, C is the output matrix, and D is the feed-through matrix. We introduce a selective state-space model with the premise of allowing the spatial features of localized objects to be enhanced by adjusting parameters according to the inputs. This enhances selective information processing across sequences.
Figure 2. Illustration of Global–Local Mamba Encoder. DWConv denotes Depth-Wise Convolution and SiLU denotes the Sigmoid Linear Unit function. Also, A is the state matrix, B is the input matrix, C is the output matrix, and D is the feed-through matrix. We introduce a selective state-space model with the premise of allowing the spatial features of localized objects to be enhanced by adjusting parameters according to the inputs. This enhances selective information processing across sequences.
Remotesensing 16 04050 g002
Figure 3. Illustration of SSE A 2 F Module. F in the figure denotes the individual feature representations of the intermediate processes of the network. The module consists of two sets of upper and lower feature inputs: spectral–spatial features and elevation features. The module can adaptively adjust the relative positions between multi-source features, and achieve optimal fusion by constructing a bootstrap subspace to accurately estimate feature-level offsets.
Figure 3. Illustration of SSE A 2 F Module. F in the figure denotes the individual feature representations of the intermediate processes of the network. The module consists of two sets of upper and lower feature inputs: spectral–spatial features and elevation features. The module can adaptively adjust the relative positions between multi-source features, and achieve optimal fusion by constructing a bootstrap subspace to accurately estimate feature-level offsets.
Remotesensing 16 04050 g003
Figure 4. Visual representation of the MUUFL Gulfport dataset. (a) False-color map. (b) Grayscale image for the LiDAR. (c) Ground truth map of various land cover categories.
Figure 4. Visual representation of the MUUFL Gulfport dataset. (a) False-color map. (b) Grayscale image for the LiDAR. (c) Ground truth map of various land cover categories.
Remotesensing 16 04050 g004
Figure 5. Visual representation of the Trento dataset. (a) False-color map. (b) Grayscale image for the LiDAR. (c) Ground truth map of various land cover categories.
Figure 5. Visual representation of the Trento dataset. (a) False-color map. (b) Grayscale image for the LiDAR. (c) Ground truth map of various land cover categories.
Remotesensing 16 04050 g005
Figure 6. Visual representation of the Augsburg dataset. (a) False-color map. (b) Grayscale image for the LiDAR. (c) Ground truth map of various land cover categories.
Figure 6. Visual representation of the Augsburg dataset. (a) False-color map. (b) Grayscale image for the LiDAR. (c) Ground truth map of various land cover categories.
Remotesensing 16 04050 g006
Figure 7. Effect of patch size on the OA, AA, and Kappa coefficient. (a) MUUFL Gulfport. (b) Trento. (c) Augsburg. It can be seen that our method shows different sensitivities to the patch size on different datasets, thus better tuning the hyperparameters of the model to the optimum.
Figure 7. Effect of patch size on the OA, AA, and Kappa coefficient. (a) MUUFL Gulfport. (b) Trento. (c) Augsburg. It can be seen that our method shows different sensitivities to the patch size on different datasets, thus better tuning the hyperparameters of the model to the optimum.
Remotesensing 16 04050 g007
Figure 8. Effect of reducing spectral dimensionality on the OA, AA, and Kappa coefficient. (a) MUUFL Gulfport. (b) Trento. (c) Augsburg. It can be seen that our method shows different sensitivities to reductions on different datasets, thus better tuning the hyperparameters of the model to the optimum.
Figure 8. Effect of reducing spectral dimensionality on the OA, AA, and Kappa coefficient. (a) MUUFL Gulfport. (b) Trento. (c) Augsburg. It can be seen that our method shows different sensitivities to reductions on different datasets, thus better tuning the hyperparameters of the model to the optimum.
Remotesensing 16 04050 g008
Figure 9. Effect of learning rate on the OA, AA, and Kappa coefficient. (a) MUUFL Gulfport. (b) Trento. (c) Augsburg. It can be seen that our method shows different sensitivities to the learning rate on different datasets, thus better tuning the hyperparameters of the model to the optimum.
Figure 9. Effect of learning rate on the OA, AA, and Kappa coefficient. (a) MUUFL Gulfport. (b) Trento. (c) Augsburg. It can be seen that our method shows different sensitivities to the learning rate on different datasets, thus better tuning the hyperparameters of the model to the optimum.
Remotesensing 16 04050 g009
Figure 10. Maps depicting the classification of MUUFL Gulfport using various methods. (a) RF. (b) SVM. (c) 2D-CNN. (d) HybridSN. (e) GAHT. (f) MiM. (g) CoupledCNN. (h) CALC. (i) HCTnet. (j) M2FNet. (k) HLMamba. (l) AFA–Mamba.
Figure 10. Maps depicting the classification of MUUFL Gulfport using various methods. (a) RF. (b) SVM. (c) 2D-CNN. (d) HybridSN. (e) GAHT. (f) MiM. (g) CoupledCNN. (h) CALC. (i) HCTnet. (j) M2FNet. (k) HLMamba. (l) AFA–Mamba.
Remotesensing 16 04050 g010
Figure 11. Maps depicting the classification of Trento using various methods. (a) RF. (b) SVM. (c) 2D-CNN. (d) HybridSN. (e) GAHT. (f) MiM. (g) CoupledCNN. (h) CALC. (i) HCTnet. (j) M2FNet. (k) HLMamba. (l) AFA–Mamba.
Figure 11. Maps depicting the classification of Trento using various methods. (a) RF. (b) SVM. (c) 2D-CNN. (d) HybridSN. (e) GAHT. (f) MiM. (g) CoupledCNN. (h) CALC. (i) HCTnet. (j) M2FNet. (k) HLMamba. (l) AFA–Mamba.
Remotesensing 16 04050 g011
Figure 12. Maps depicting the classification of Augsburg using various methods. (a) RF. (b) SVM. (c) 2D-CNN. (d) HybridSN. (e) GAHT. (f) MiM. (g) CoupledCNN. (h) CALC. (i) HCTnet. (j) M2FNet. (k) HLMamba. (l) AFA–Mamba.
Figure 12. Maps depicting the classification of Augsburg using various methods. (a) RF. (b) SVM. (c) 2D-CNN. (d) HybridSN. (e) GAHT. (f) MiM. (g) CoupledCNN. (h) CALC. (i) HCTnet. (j) M2FNet. (k) HLMamba. (l) AFA–Mamba.
Remotesensing 16 04050 g012
Table 1. Training and test samples in MUUFL Gulfport, Trento, and Augsburg.
Table 1. Training and test samples in MUUFL Gulfport, Trento, and Augsburg.
IDMUUFL GulfportTrentoAugsburg
Land Cover Class Training Test Land Cover Class Training Test Land Cover Class Training Test
C01Trees Mostly116322,083Apple Trees214013Forest67612,831
C02Grass2144056Buildings152888Residential Area151728,812
C03Mixed Ground Surface3456537Ground3476Industrial Area1933658
C04Dirt and Sand921734Woods469077Low Plants134325,514
C05Road3356352Vineyard5310,448Allotment29546
C06Water24442Roads163158Commercial Area831562
C07Buildings Shadow1122121   Water771453
C08Buildings3125928     
C09Sidewalk701315     
C10Yellow Curb10173     
C11Cloth Panels14255     
Total269150,990Total15430,060Total391874,376
Table 2. Ablation experimental results. The results show that each module of our method works and is consistent with the motivation for the experiment.
Table 2. Ablation experimental results. The results show that each module of our method works and is consistent with the motivation for the experiment.
CasesComponentIndicators
PCA SSE-FE HSI-GL-Mamba LiDAR-GL-Mamba A 2 F OA (%) AA (%) k 100
1×89.8779.4187.62
2CNN91.2581.2489.35
3×90.4980.1288.14
4×90.4880.5487.02
5×90.0581.2888.53
6×91.2380.3688.15
7Original Mamba91.5281.6289.32
8Original Mamba90.7380.5488.95
993.1182.8190.90
Table 3. Classification accuracy of comparison experiment in MUUFL Gulfport using various methods. Nos. 1 to 11 are the classification accuracies for each landform category.
Table 3. Classification accuracy of comparison experiment in MUUFL Gulfport using various methods. Nos. 1 to 11 are the classification accuracies for each landform category.
No.Only HSI InputHSI and LiDAR Input
RF [20] SVM [21] 2D-CNN [49] HybridSN [29] GAHT [35] MiM [50] CoupledCNN [40] CALC [42] HCTnet [43] M2FNet [44] HLMamba [51] Ours
196.54 ± 1.1397.64 ± 1.0797.15 ± 0.3297.07 ± 0.396.21 ± 0.7495.75 ± 1.2297.23 ± 0.5297.46 ± 1.0397.13 ± 0.6796.95 ± 0.4197.12 ± 0.5596.94 ± 0.31
279.11 ± 4.3087.39 ± 3.0484.61 ± 1.2881.91 ± 7.9984.15 ± 5.4285.96 ± 1.3282.21 ± 4.0088.41 ± 3.1986.31 ± 1.6086.12 ± 2.5486.40 ± 3.1287.43 ± 4.3
383.76 ± 3.1486.14 ± 4.8386.63 ± 3.3683.61 ± 4.0385.32 ± 3.8788.23 ± 1.287.64 ± 2.7384.82 ± 5.4489.40 ± 1.5188.86 ± 1.8188.39 ± 1.7291.68 ± 1.16
489.36 ± 2.7293.64 ± 2.0194.92 ± 0.8792.39 ± 6.9389.52 ± 4.5591.65 ± 3.0992.90 ± 2.6889.39 ± 3.4890.40 ± 1.5289.85 ± 1.6491.59 ± 1.6291.42 ± 2.76
585.99 ± 4.1190.39 ± 2.3990.78 ± 0.9790.74 ± 1.0386.95 ± 3.3691.16 ± 0.6793.93 ± 1.4593.03 ± 1.9893.62 ± 1.0394.01 ± 0.7092.02 ± 0.6794.45 ± 1.06
689.37 ± 4.7795.42 ± 2.8695.44 ± 3.6184.47 ± 7.0394.40 ± 3.0396.44 ± 0.3194.57 ± 2.8790.16 ± 3.8683.62 ± 5.3493.28 ± 1.7596.24 ± 1.0391.9 ± 3.24
780.31 ± 3.6983.18 ± 2.3081.88 ± 2.1778.41 ± 4.2975.55 ± 7.7581.09 ± 2.1588.77 ± 1.0683.61 ± 4.2386.04 ± 1.9984.89 ± 2.7685.82 ± 0.9386.36 ± 1.99
895.70 ± 2.0996.86 ± 1.1896.39 ± 0.9295.82 ± 1.4593.81 ± 2.8695.8 ± 0.5196.21 ± 1.4396.60 ± 0.9596.68 ± 0.4797.14 ± 0.5395.72 ± 0.3796.58 ± 0.79
928.39 ± 4.1140.75 ± 5.5854.18 ± 5.4448.30 ± 7.1936.06 ± 7.0448.19 ± 5.4745.31 ± 6.1150.78 ± 9.3452.11 ± 4.1657.64 ± 3.0848.55 ± 2.8253.44 ± 4.87
105.53 ± 2.8413.43 ± 7.1823.60 ± 5.9426.63 ± 5.6511.22 ± 5.4120.56 ± 1.9116.18 ± 3.7811.56 ± 8.3820.00 ± 2.0224.57 ± 4.7026.25 ± 4.5827.11 ± 5.54
1170.00 ± 8.3080.47 ± 7.5386.71 ± 4.7492.59 ± 4.1861.84 ± 7.6864.62 ± 6.6288.08 ± 6.9272.39 ± 11.376.20 ± 2.8469.37 ± 5.6961.31 ± 2.393.57 ± 1.77
OA(%)88.92 ± 0.7991.76 ± 0.5191.79 ± 0.4590.27 ± 1.0889.11 ± 0.8891.63 ± 0.2292.09 ± 0.6191.94 ± 0.6592.45 ± 0.2892.53 ± 0.1992.01 ± 0.2793.11 ± 0.39
AA(%)73.09 ± 1.0278.66 ± 1.3381.12 ± 0.8379.19 ± 1.9674.05 ± 1.2978.25 ± 0.8980.28 ± 1.9178.02 ± 1.8779.23 ± 0.8580.24 ± 0.8079.04 ± 0.6682.81 ± 1.31
k × 10085.27 ± 1.0489.05 ± 0.7089.10 ± 0.6187.13 ± 1.4385.59 ± 1.1188.91 ± 0.2989.52 ± 0.8189.30 ± 0.8690.00 ± 0.3690.12 ± 0.2589.42 ± 0.3590.90 ± 0.52
Table 4. Classification accuracy of comparison experiment in Trento using various methods. Nos. 1 to 6 are the classification accuracies for each landform category.
Table 4. Classification accuracy of comparison experiment in Trento using various methods. Nos. 1 to 6 are the classification accuracies for each landform category.
No.Only HSI InputHSI and LiDAR Input
RF [20] SVM [21] 2D-CNN [49] HybridSN [29] GAHT [35] MiM [50] CoupledCNN [40] CALC [42] HCTnet [43] M2FNet [44] HLMamba [51] Ours
194.62 ± 4.9196.28 ± 4.6197.41 ± 1.8796.33 ± 5.0998.69 ± 0.7990.96 ± 4.8497.91 ± 3.4598.56 ± 1.3099.30 ± 0.8199.08 ± 0.8298.95 ± 0.7899.43 ± 0.40
282.93 ± 12.8192.68 ± 3.3385.49 ± 3.7476.90 ± 7.4379.42 ± 6.6948.14 ± 11.1494.17 ± 1.4694.49 ± 2.8098.10 ± 2.1890.54 ± 5.5384.29 ± 4.9994.29 ± 2.20
36.02 ± 10.709.08 ± 5.0739.56 ± 13.6445.35 ± 16.1331.97 ± 7.1654.38 ± 9.2418.30 ± 8.5151.41 ± 23.4863.57 ± 18.9165.8 ± 12.7334.3 ± 10.7358.42 ± 15.5
499.90 ± 0.1099.88 ± 0.1899.85 ± 0.2099.86 ± 0.2699.93 ± 0.0599.82 ± 0.1998.77 ± 0.8099.99 ± 0.02100.00 ± 0.0099.97 ± 0.0499.96 ± 0.0599.99 ± 0.02
599.69 ± 0.63100.00 ± 0.0099.81 ± 0.1994.96 ± 3.2799.99 ± 0.0295.35 ± 2.599.91 ± 0.1099.96 ± 0.0899.76 ± 0.3299.96 ± 0.0799.57 ± 0.2299.99 ± 0.01
678.31 ± 5.5081.83 ± 4.1283.76 ± 2.3071.88 ± 4.8192.73 ± 3.1993.28 ± 3.4598.20 ± 0.5288.22 ± 3.6489.75 ± 2.9894.07 ± 2.2585.66 ± 4.9898.43 ± 0.44
OA(%)93.74 ± 1.0395.41 ± 0.7595.48 ± 0.6591.67 ± 1.7195.98 ± 0.5090.71 ± 1.2597.27 ± 0.5397.25 ± 0.5697.99 ± 0.2797.78 ± 0.6195.64 ± 0.5898.55 ± 0.35
AA(%)76.91 ± 2.7879.96 ± 1.1184.31 ± 2.5580.88 ± 3.2583.79 ± 1.6480.32 ± 2.7084.54 ± 1.2188.77 ± 3.7691.75 ± 3.0791.57 ± 2.2983.79 ± 2.2391.76 ± 2.68
k × 10091.57 ± 1.4193.83 ± 1.0393.94 ± 0.8788.86 ± 2.3194.61 ± 0.6887.61 ± 1.6696.35 ± 0.7296.32 ± 0.7697.31 ± 0.3697.03 ± 0.8294.16 ± 0.7898.06 ± 0.47
Table 5. Classification accuracy of comparison experiment in Augsburg using various methods. Nos. 1 to 7 are the classification accuracies for each landform category.
Table 5. Classification accuracy of comparison experiment in Augsburg using various methods. Nos. 1 to 7 are the classification accuracies for each landform category.
No.Only HSI InputHSI and LiDAR Input
RF [20] SVM [21] 2D-CNN [49] HybridSN [29] GAHT [35] MiM [50] CoupledCNN [40] CALC [42] HCTnet [43] M2FNet [44] HLMamba [51] Ours
198.58 ± 0.4198.89 ± 0.3299.15 ± 0.1898.88 ± 0.6797.76 ± 0.5299.24 ± 0.299.40 ± 0.2198.76 ± 0.3099.09 ± 0.1798.23 ± 0.7299.1 ± 0.2299.16 ± 0.21
296.52 ± 3.1297.86 ± 1.5197.75 ± 0.3697.29 ± 0.7698.13 ± 0.6298.18 ± 0.2599.17 ± 0.2099.34 ± 0.1298.80 ± 0.2698.37 ± 0.3298.22 ± 0.5298.83 ± 0.36
377.82 ± 5.2082.22 ± 3.4480.18 ± 2.8676.87 ± 4.2285.70 ± 3.2180.8 ± 1.4282.25 ± 4.9991.69 ± 3.1691.02 ± 2.6391.93 ± 2.0483.15 ± 4.089.38 ± 2.6
498.77 ± 0.2199.02 ± 0.3599.16 ± 0.1098.94 ± 0.1997.57 ± 0.4899.05 ± 0.0798.98 ± 0.2199.08 ± 0.1798.60 ± 0.2698.43 ± 0.5698.8 ± 0.8898.92 ± 0.26
541.93 ± 14.7761.47 ± 9.8469.41 ± 5.1467.53 ± 4.4465.46 ± 10.0669.87 ± 5.5278.99 ± 7.9685.35 ± 3.1581.92 ± 5.5980.95 ± 5.0178.17 ± 3.8383.86 ± 2.96
622.25 ± 5.6532.39 ± 4.9248.68 ± 5.7040.33 ± 4.7667.74 ± 5.9251.02 ± 5.4755.43 ± 6.8858.67 ± 9.4868.07 ± 4.1171.54 ± 3.2359.42 ± 4.4676.2 ± 4.67
757.71 ± 3.3660.30 ± 2.2765.87 ± 1.2464.21 ± 2.3860.10 ± 4.2067.12 ± 3.3064.84 ± 2.5663.78 ± 1.6968.31 ± 2.6268.29 ± 3.2268.62 ± 1.5166.81 ± 3.53
OA(%)94.01 ± 1.0795.29 ± 0.6395.75 ± 0.1195.07 ± 0.4595.64 ± 0.3096.00 ± 0.1396.57 ± 0.2897.12 ± 0.1597.03 ± 0.2096.77 ± 0.3196.29 ± 0.5897.24 ± 0.14
AA(%)70.51 ± 2.2576.02 ± 1.8380.03 ± 1.1877.72 ± 0.9581.78 ± 2.0280.76 ± 0.6182.72 ± 0.8585.24 ± 1.2986.54 ± 1.2686.82 ± 1.0783.64 ± 0.6487.60 ± 0.93
k × 10091.40 ± 1.4693.24 ± 0.8793.91 ± 0.1692.93 ± 0.6393.75 ± 0.4394.26 ± 0.1895.08 ± 0.4095.87 ± 0.2395.75 ± 0.2995.37 ± 0.4594.69 ± 0.8196.05 ± 0.20
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, S.; Huang, S. AFA–Mamba: Adaptive Feature Alignment with Global–Local Mamba for Hyperspectral and LiDAR Data Classification. Remote Sens. 2024, 16, 4050. https://doi.org/10.3390/rs16214050

AMA Style

Li S, Huang S. AFA–Mamba: Adaptive Feature Alignment with Global–Local Mamba for Hyperspectral and LiDAR Data Classification. Remote Sensing. 2024; 16(21):4050. https://doi.org/10.3390/rs16214050

Chicago/Turabian Style

Li, Sai, and Shuo Huang. 2024. "AFA–Mamba: Adaptive Feature Alignment with Global–Local Mamba for Hyperspectral and LiDAR Data Classification" Remote Sensing 16, no. 21: 4050. https://doi.org/10.3390/rs16214050

APA Style

Li, S., & Huang, S. (2024). AFA–Mamba: Adaptive Feature Alignment with Global–Local Mamba for Hyperspectral and LiDAR Data Classification. Remote Sensing, 16(21), 4050. https://doi.org/10.3390/rs16214050

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop