Next Article in Journal
Intelligent Fault Diagnosis of Hydraulic System Based on Multiscale One-Dimensional Convolutional Neural Networks with Multiattention Mechanism
Previous Article in Journal
Strategy for Accurate Detection of Six Tropane Alkaloids in Honey Using Lateral Flow Immunosensors
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

MFMamba: A Mamba-Based Multi-Modal Fusion Network for Semantic Segmentation of Remote Sensing Images

by
Yan Wang
1,
Li Cao
1,* and
He Deng
2
1
School of Electrical and Electronic Engineering, Wuhan Polytechnic University, Wuhan 430023, China
2
School of Computer Science and Technology, Wuhan University of Science and Technology, Wuhan 430081, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(22), 7266; https://doi.org/10.3390/s24227266
Submission received: 4 October 2024 / Revised: 5 November 2024 / Accepted: 12 November 2024 / Published: 13 November 2024
(This article belongs to the Section Remote Sensors)

Abstract

:
Semantic segmentation of remote sensing images is a fundamental task in computer vision, holding substantial relevance in applications such as land cover surveys, environmental protection, and urban building planning. In recent years, multi-modal fusion-based models have garnered considerable attention, exhibiting superior segmentation performance when compared with traditional single-modal techniques. Nonetheless, the majority of these multi-modal models, which rely on Convolutional Neural Networks (CNNs) or Vision Transformers (ViTs) for feature fusion, face limitations in terms of remote modeling capabilities or computational complexity. This paper presents a novel Mamba-based multi-modal fusion network called MFMamba for semantic segmentation of remote sensing images. Specifically, the network employs a dual-branch encoding structure, consisting of a CNN-based main encoder for extracting local features from high-resolution remote sensing images (HRRSIs) and of a Mamba-based auxiliary encoder for capturing global features on its corresponding digital surface model (DSM). To capitalize on the distinct attributes of the multi-modal remote sensing data from both branches, a feature fusion block (FFB) is designed to synergistically enhance and integrate the features extracted from the dual-branch structure at each stage. Extensive experiments on the Vaihingen and the Potsdam datasets have verified the effectiveness and superiority of MFMamba in semantic segmentation of remote sensing images. Compared with state-of-the-art methods, MFMamba achieves higher overall accuracy (OA) and a higher mean F1 score (mF1) and mean intersection over union (mIoU), while maintaining low computational complexity.

1. Introduction

Semantic segmentation of remote sensing images is a pixel-level classification task that aims to assign a specific land cover class to each pixel [1]. The derived classification information holds significant value across various fields, such as land cover [2], change detection [3], environmental protection [4], and building extraction [5]. The progression in earth observation technology has rendered the acquisition of diverse high-resolution remote sensing data more accessible [6]. This includes, but is not limited to, multi-spectral imaging (MSI), synthetic aperture radar (SAR), and light detection and ranging (LiDAR) technologies, each offering unique insights for detailed analysis and classification. Exploiting the complementary characteristics from different modalities can significantly enhance semantic segmentation performance [7]. Nonetheless, high-resolution remote sensing data are characterized by several distinct attributes, such as complex backgrounds, rich feature details, substantial intra-class variance, and relatively limited inter-class variance. Consequently, the extraction of significant features from different modal images prior to fusion remains a considerable challenge.
Traditional approaches, such as the Support Vector Machine (SVM) [8], Random Forest [9], and Conditional Random Field (CRF) [10], have been found wanting in addressing the requirements of modern high-resolution remote sensing image (HRRSI) applications due to their limitations stemming from a lack of robust abstraction and semantic feature extraction capabilities. In recent years, the predominant methods for semantic segmentation of HRRSIs have shifted towards deep learning techniques. The prevailing models based on deep learning can be broadly categorized into two main types: Convolutional Neural Networks (CNNs) and Vision Transformers (ViTs). These methods employ data-driven automation technology, allowing for autonomous learning from datasets and they generally outperform traditional approaches due to their advanced feature extraction capabilities. Nevertheless, CNNs face limitations in capturing long-range dependencies due to the inherent constraints of convolutional operations. On the other hand, ViTs excel in extracting global contextual information, though their effectiveness is contingent upon access to substantial memory and computational resources. Recently, Mamba [11], a novel state space model (SSM) originating from the natural language processing (NLP) field, has been introduced as a potential alternative for establishing long-range dependency relationships while maintaining linear computational complexity [12]. Although these deep learning-based methods have shown promise in feature extraction from single-modal images, relying solely on single-modal images for semantic segmentation limits the full utilization of available multi-modal images [13]. Compared with conventional single-modal segmentation approaches, models based on multi-modal fusion can harness target features from multiple perspectives, thereby enhancing segmentation performance [7]. However, current multi-modal fusion models based on CNNs or ViTs face challenges in capturing remote dependencies or reducing computational complexity, as well as in dealing with the incompatibility of multi-modal data.
To overcome the aforementioned challenges in semantic segmentation of remote sensing images, this paper proposes a multi-modal fusion network based on Mamba, namely MFMamba. Specifically, our MFMamba employs a dual-branch encoding structure, where the main branch extracts the features of HRRSIs utilizing a CNN-based encoder, while the auxiliary branch extracts the features of the corresponding digital surface model (DSM) based on a Mamba-based encoder. The features extracted from each stage of the two encoders are fused by an innovative feature fusion block (FFB). Subsequently, the fused features are input into the next stage of the main encoder and fed to the corresponding decoding layer as well. Finally, the segmentation result is obtained by a Transformer-based decoder.
Overall, this paper makes the following contributions:
  • A novel multi-modal fusion network based on Mamba (named as MFMamba) is proposed for semantic segmentation of remote sensing images. In this novel network, a Mamba-based auxiliary encoder is utilized to effectively capture global information from the DSM, while ensuring the network maintains low computational complexity.
  • An innovative feature fusion block (FFB) is designed to effectively fuse the features extracted from HRRSI and its corresponding DSM data, where the multi-convolutional kernel attention (MCKA) unit can further capture local details, while the efficient additive attention (EAA) unit can effectively capture long-range dependencies.
  • Extensive comparison experiments conducted on the Vaihingen dataset and the Potsdam dataset demonstrate that our proposed MFMamba has superior semantic segmentation performance and low computational complexity compared with seven state-of-the-art methods.

2. Related Work

2.1. Single-Modal Semantic Segmentation

For single-modal remote sensing images, the Fully Convolutional Network (FCN) was pioneering in its use of full convolution for pixel-level prediction [14], effectively creating a CNN architecture that addressed the semantic segmentation problem in an end-to-end manner and significantly advanced the development of semantic segmentation models. However, the sampling operation of the FCN on resolution recovery is straightforward, which can negate the impact of global contextual information. This method also encounters challenges, such as edge blurring and imprecise segmentation. To mitigate these issues, Unet was proposed as a network comprising a pyramid encoder and a symmetric decoder [15]. The encoder facilitates the extraction of multi-scale features via a progressive downsampling process, while the decoder recovers spatial resolution and contextualizes the semantic information. Following this innovation, the encoder–decoder framework has become the norm for remote sensing image segmentation networks [16]. Nevertheless, the size of the convolution kernel is limited by the convolution operation itself, preventing it from being too large. Each convolutional kernel primarily focuses on local information within its receptive field. As a result, CNN-based methods are not effective in capturing global semantic information and long-range dependencies within input images [17,18].
In order to address these challenges, the ViT was the first to introduce Transformers into computer vision tasks [19], leveraging its ability to capture long-range dependencies and yielding promising results in image classification. SegFormer further refined the architecture of the ViT to make it more suitable for semantic segmentation [20]. TransUNet employed a Transformer-based encoder and a UNet decoder for medical image segmentation [21]. The ViT excels in sequence-to-sequence modeling and significantly outperforms CNN-based models in extracting global contextual information [7]. However, the self-attention mechanism of Transformer-based models also results in a quadratic complexity with respect to input size [11], and a considerable computational load [22].
Recently, Mamba has garnered significant attention for its capability of establishing long-range dependencies while maintaining linear computational complexity [11] and has shown great potential in language understanding and vision-related applications [23], including language recognition [11], medical image segmentation [24,25], image classification [26], and 3D scene understanding [27]. Following these successes, Mamba has also demonstrated promising advancements in the domain of remote sensing imagery. One example is the novel remote sensing image classification system proposed in RSMamba [26]. In addition, RSCaMa incorporated SSM into remote sensing image change captioning (RSICC) [28], utilizing multiple CaMa layers for iterative spatial change sensing and temporal interactions. Pan-Mamba explored the application of Mamba in the field of pan-sharpening and proposed a new pan-sharpening network [29]. Samba introduced a new semantic segmentation framework for HRRSIs based on Mamba, utilizing an encoder–decoder architecture [30]. Despite the development of Mamba-based methods in single-modal remote sensing image processing, there has been no exploration into applying Mamba for semantic segmentation of multi-modal remote sensing data.

2.2. Multi-Modal Semantic Segmentation

In contrast to images employed in traditional computer vision tasks, HRRSIs encompass extensive geographical areas. The limited spectral characteristics of HRRSIs make low-level inter-class differences diverse, which are further complicated by the presence of shadows, noise, obstacles, geometric distortions, and building height variations [13]. Consequently, semantic segmentation relying solely on single-modal HRRSIs frequently encounters significant challenges. As the DSM encapsulates the elevation data of surface features, it can provide crucial information for the identification of highly consistent categories within an image, as well as for the clear delineation of boundaries between different categories based on elevation data. Therefore, introducing the features of the DSM can significantly improve overall segmentation accuracy [31].
To address the multi-modal fusion problem in remote sensing images, three different fusion strategies have been explored: early fusion, middle fusion, and late fusion [32]. As an early or data-level fusion approach, ResUNet-a incorporated RGB and DSM data with structural information as input into a multi-modal network [33]. Generally, early fusion requires the precise alignment of multi-modal images, yet this process frequently results in the generation of erroneous or uncorrelated features during network training. Consequently, the potential for effectively leveraging the complementary attributes of different modalities is significantly diminished [13]. In contrast, as a late or decision-level fusion method, VFuseNet employed a two-branch network to fuse RGB and DSM data after the decoding stage [34]. Similarly, the boundary detection-based semantic segmentation method proposed by Marmanis et al. also performed fusion at the final prediction stage [18]. Compared with early fusion, late fusion offers greater scalability and flexibility. However, it may suffer from insufficient cross-modal correlation [13]. Middle or feature-level fusion methods, such as FuseNet [35], employed a simple two-branch network to fuse RGB and DSM data just before the decoder stage, and cross-modal features were combined through element-wise summation and merged at various scales. However, middle fusion is only in the decoder part of the fusion, resulting in a deficiency of information interaction with the decoder. Audebert et al. extended the application of FuseNet for cross-modal fusion by incorporating residual correction in late fusion [34]. HAFNet introduced a feature-level fusion network based on a hybrid attention perception mechanism [36], while CMGFNet developed a gated fusion module to adaptively learn discriminative features and remove irrelevant information [13]. Nonetheless, these methods often fall short in effectively extracting global semantics, as they ignore long-range spatial relations [7].
Recently, Transformers have been used for multi-modal fusion [37,38] and for fusing different modalities in semantic segmentation due to their proficiency in extracting global contextual information [39]. CMFNet [31] introduced a multi-modal multi-scale fusion network based on the improved skip connection within Transformers. MFTransNet [40] presented a multi-modal semantic segmentation structure that combines CNNs with Transformers. FTransUNet [7] introduced a multi-modal semantic segmentation network that fuses shallow and deep features in a multi-level way, thereby capturing both local details and global semantics. Although multi-modal Transformers are able to capture long-range dependencies and achieve multi-scale feature fusion, their high computational complexity poses significant challenges in terms of considering model efficiency and memory footprint [12].
As Mamba has shown strong potential for establishing remote dependencies, some recent works directly use the SSM within Mamba as a module, without in-depth design for specific tasks [41]. Furthermore, the application of Mamba to multi-modal tasks has not been thoroughly investigated. Even though MambaReID [42] introduced a multi-modal fusion network based on Mamba for multi-modal object re-identification (ReID) and Sigma presented a semantic segmentation network that incorporates an attention-based Mamba fusion mechanism along with a channel-aware Mamba decoder [41], there is no Mamba-based method dedicated to semantic segmentation of multi-modal remote sensing data. Our proposed MFMamba, which leverages CNNs to extract local features and Mamba to capture global contextual information and which has feature fusion facilitated by the fusion blocks (FFBs), represents a pioneering endeavor in semantic segmentation of multi-modal remote sensing data.

3. Method

In this section, we provide a comprehensive overview of MFMamba, followed by detailed descriptions of its main constituent modules and, finally, a brief introduction of the used loss function is given.

3.1. Framework of MFMamba

The framework of MFMamba is illustrated in Figure 1, which mainly consists of four parts: a Mamba-based auxiliary encoder, a CNN-based main encoder, feature fusion blocks (FFBs), and a Transformer-based decoder. Specifically, the CNN-based main encoder is employed to extract local features from the HRRSIs, while the Mamba-based auxiliary encoder is employed to extract global features from the corresponding DSM data.
Following the feature extraction at stage i, i 1 , 2 , 3 , 4 , an FFB is used to fuse the features extracted from both encoder branches, generating multi-scale fusion features. At each stage, these fused features are then skip-connected to the features of the corresponding layer in the Transformer-based decoder. In particular, these fused features, along with the features generated by the global–local transformer block (GLTB) from the previous layer in the decoder [43], are aggregated together to contribute to segmentation accuracy. This aggregation is selectively weighted by the weighted sum (WS) operation, aimed at learning more generalized fusion features. The formulation of the WS operation can be described as
F d i = β F f i + 1 β F g i + 1 ,
where F d i represents the aggregated features at stage i in the decoder, F f i denotes the fused features produced by the FFB at stage i, F g i + 1 indicates the features generated by the GLTB at stage i + 1, i 1 , 2 , 3 , and β is the learnable weight. Finally, the aggregated features at stage 1 are processed by the feature refinement head (FRH) to generate the final prediction result.

3.2. Mamba-Based Auxiliary Encoder

As shown in Figure 1, the Mamba-based auxiliary encoder is structured into four successive stages. The first stage contains a patch embedding layer followed by a visual state space (VSS) block, whereas each of the remaining stages consists of a patch merging layer and a VSS block [12]. The auxiliary encoder receives the DSM data as input, which is denoted as Y H × W × 1 , where H and W respectively represent the height and width of the input. In the first stage, Y is divided into non-overlapping patches of size 2 × 2 by the patch embedding layer. The embedded image is then processed by the first VSS block. In contrast to the first stage, the remaining three stages commence with a patch merging operation that reduces the height and width of the input features while expanding the number of channels.
The VSS block, derived from VMamba [44], is the core module of the auxiliary encoder, as illustrated in Figure 2a. The input feature is processed through a series of operations, including the layer normalizations (Layer Norm), the linear projections (Linear), the depth-wise convolution (DWConv) utilized in the original Mamba, and the selective scan 2D (SS2D) unit used to model long-range spatial information from the feature. Within the SS2D unit, given an input feature map y, the output feature map y ¯ of the SS2D can be expressed as
y s = S 6 s c a n y , s ,
y ¯ = m e r g e y 1 , y 2 , y 3 , y 4 ,
where s { 1 , 2 , 3 , 4 } denotes four different scanning directions (left to right, right to left, top to bottom, and bottom to top) and where scan (·), S6 [·], and merge (·) denote the cross scan, the selective scan of the S6 block, and the scan merging operation, respectively. Figure 2b shows the visualization of the SS2D mechanism applied to the DSM data. The SS2D unit first unfolds the input patches into sequences along four distinct traversal paths. Each patch sequence is then independently processed through a dedicated S6 block [44], and its resulting sequence is subsequently reshaped and merged to form the output feature map. The detailed structure of the S6 block can be found in [44].

3.3. CNN-Based Main Encoder

The ResNet18 is utilized as the CNN-based main encoder for processing HRRSIs. Assuming that the input is represented as X H × W × 3 , H and W respectively represent the height and width of the input. As illustrated in Figure 1, the ResNet18 encompasses four sequential ResBlocks, each of which performs downsampling on the feature map with a scale factor of two. In order to enhance the utilization of multi-modal feature information, with the exception of the initial ResBlock, the inputs for the subsequent three ResBlocks are the fused features that have been processed through the FFBs. Moreover, skip connections are utilized by directly feeding the outputs of the FFBs into the corresponding decoder layers, which are designed to recover local details and contextual information.

3.4. Feature Fusion Block

The overall architecture of an FFB is illustrated in Figure 3a. At each stage, the left input F L i n i originates from the output of the VSS block in the auxiliary encoder branch, while the right input F R i n i is derived from the output of the ResBlock in the main encoder branch.
Prior to the fusion of the two input features, given that the left input derived from the VSS block is with long-range properties, while the right input derived through convolution operations primarily encapsulates local attributes with a limited incorporation of global information, it is necessary to employ the multi-convolutional kernel attention (MCKA) unit on the left branch in order to further capture local details and to employ the efficient additive attention (EAA) unit [45] on the right branch in order to effectively capture long-range dependencies. The first fusion of input features from both encoder branches is accomplished through the element-wise summation operation performed on the outputs from the MCKA unit and the EAA unit. In order to further extract features efficiently, the fused features are then subjected to deep separable convolution processing, where a convolution kernel of 8 × 8 is employed in the depth-wise convolution to extract spatial features, and a convolution kernel of 1 × 1 is subsequently utilized in the point-wise convolution to extract channel features. This approach enables a reduction in computation while simultaneously enhancing computation efficiency, without compromising the overall performance of the network. Considering that the DSM may exhibit varying degrees of information loss and noise, whereas the HRRSI contains crucial information that requires refinement, the input feature F R i n i extracted from the HRRSI is once more employed for fusion with the processed features through the element-wise summation operation, thus further enhancing the detailed features. Prior to being processed by a multilayer perceptron (MLP) unit, the features—after secondary fusion—are subjected to a process of layer normalization, which makes the training process converge faster. The final fusion features are obtained by the residual connection between the outputs of the MLP unit and the features after secondary fusion.
The detailed structure of an MCKA unit is depicted in Figure 3b, which consists of a 3 × 3 convolutional layer to capture local information, followed by two parallel convolutional layers with kernel sizes of 5 and 1, respectively, enabling the capture of contextual information across various scales [46]. Finally, the local and contextual features are combined through a 1 × 1 convolutional layer to capture a wide range of contextual information without compromising the integrity of local texture features.
The detailed structure of an EAA unit is shown in Figure 3c. The conventional attention mechanism in natural language processing (NLP) encodes a relevance score for the contextual information of the input sequence based on the interaction among three attention components: query, key, and value. In contrast, the EAA mechanism eliminates key–value interactions while maintaining performance and focuses on effectively encoding query–key interactions through the incorporation of a linear projection layer. The input embedding matrix e is transformed into Q u e r y and K e y , where Q u e r y , K e y n × d , n is the token length, and d is the dimensions of the embedding vector. Next, the Q u e r y is multiplied by the learnable parameter vector G d to learn the attention weights of the query, producing the global attention query vector α n as follows:
α = Query G d .
Then, the Q u e r y is pooled based on the learned attention weights, resulting in a single global query vector q d as follows:
q = i = 1 n α i Query i .
Here, denotes the broadcasted element-wise multiplication operation. Finally, the output of the EAA unit E can be described as:
E = Q ^ + T ( Key q ) ,
where Q ^ denotes to the normalized Q u e r y , T denotes to the linear transformation.

3.5. Transformer-Based Decoder

The challenge of multi-scale issues in remote sensing images complicates target localization and recognition. Conventional decoders often struggle to accurately recover image details due to the absence of global semantic information [43]. The decoder in UNetFormer effectively addresses this by capturing global and local contextual information across multiple scales [43]. As depicted in Figure 1, the Transformer-based decoder mainly consists of three GLTBs and an FRH. The detailed structure of a GLTB is shown in Figure 4a, which contains two parallel branches to extract global and local context, respectively. The local branch adopts two parallel convolutional layers with kernel sizes of 3 and 1 to extract local context, while the global branch relies on the multi-head self-attention of the window to capture global context. Because the shallow features generated from the FFB at stage 1 retain rich spatial details yet fall short in semantic content and because the features processed by the GLTB at stage 2 provide precise semantic information but with low spatial resolution, the FRH is used at stage 1 to narrow the semantic gap between the two features, thereby further improving overall accuracy. As illustrated in Figure 4b, the detailed structure of an FRH features two parallel pathways that are designed to enhance the channel-wise and spatial-wise feature representations, and the attentional features generated by the two pathways are further fused by summation operations. Subsequent to this fusion, convolution and upsampling operations are employed to yield the final segmentation map. Detailed descriptions of the GLTB and the FRH can be found in UNetFormer [43].

3.6. Loss Function

Based on the multi-head design in the Transformer-based decoder, the loss function applied in this paper is a combination of the Cross-entropy loss and the Dice loss.
The Cross-entropy loss is expressed as follows:
L ce = 1 N n = 1 N k = 1 K y k n log y ^ k n ,
where N is the number of samples, K represents the number of categories, y k n represents the one-hot encoding of the true semantic labels, and y ^ k n is the confidence of sample n belonging to the category k.
The Dice loss is expressed as follows:
L dice = 1 2 N n = 1 N k = 1 K y k n y ^ k n y k n + y ^ k n .
The total loss function is expressed as:
L total = L ce + L dice + γ × L GLTB ,
where L GLTB is the loss of operations on three GLTB blocks. The detailed description of L GLTB can be found in UNetFormer [43]. The symbol γ is the cofactor and is set to 0.4 by default.

4. Experiments and Results

4.1. Datasets

The Vaihingen dataset consists of 16 very high-resolution true orthophotos (TOPs), each with an average size of 2500 × 2000 pixels. Every orthophoto has three channels, namely near-infrared, red, and green (NIRRG), along with a DSM with a 9 cm ground sampling distance (GSD). The dataset consists of five foreground classes, namely Impervious surface (Imp.), Building (Bui.), Low vegetation (Low.), Tree (Tre.), and Car, as well as one background class (Clutter). In our experiments, we utilized TOP image tiles and complete images. The 16 orthophoto images and their corresponding DSM data were divided into a training set containing 12 patches and a test set containing 4 patches. The training set comprises images identified by the indices 1, 3, 23, 26, 7, 11, 13, 28, 17, 32, 34, and 37. The test set consists of images identified by the indices 5, 21, 15, and 30.
The Potsdam dataset is composed of 24 very high-resolution TOPs, each with a size of 6000 × 6000 pixels, involving the same category information as the Vaihingen dataset. Distinct from the Vaihingen dataset, it provides four multi-spectral channels, including infrared, red, green, and blue (IRRGB), along with a normalized DSM at 5 cm GSD. In our experiments, we selected to use their RGB composites and the corresponding DSM data. The 24 orthophoto images and their corresponding DSM data were divided into a training set containing 18 patches and a test set containing 6 patches. The training set includes images identified by the following indices: 6_10, 7_10, 2_12, 3_11, 2_10, 7_8, 5_10, 3_12, 5_12, 7_11, 7_9, 6_9, 7_7, 4_12, 6_8, 6_12, 6_7, and 4_11. The test set is composed of images with the indices 2_11, 3_10, 4_10, 5_11, 6_11, and 7_12. Figure 5 presents a selection of data samples from the Vaihingen and the Potsdam datasets.

4.2. Evaluation Metrics and Experimental Setup

In order to quantitatively evaluate segmentation performance, we utilized overall accuracy (OA) and the mean F1 score (mF1) and mean intersection over union (mIoU) as our evaluation metrics. Based on the accumulated confusion matrix, the calculations for OA and the mF1 and mIoU can be expressed as follows:
OA = k = 1 N T P k + T N k k = 1 N T P k + F P k + T N k + F N k ,
Q p = 1 N k = 1 N T P k T P k + F P k ,
Q r = 1 N k = 1 N T P k T P k + F N k ,
F 1 = 2 × Q p × Q r Q p + Q r ,
mIoU = 1 N k = 1 N T P k T P k + F P k + F N k ,
where T P k , F P k , T N k , and F N k denote true positives, true negatives, false positives, and false negatives, respectively, for objects indexed as class k . Specifically, we incorporated Clutter into the evaluation of OA and calculated the mF1 and mIoU values for five foreground classes.
All the experiments were conducted within the PyTorch framework on an NVIDIA A40 GPU with 48 GB RAM. During the training process, the images were randomly cropped into 256 × 256 patches, and data augmentation techniques were adopted such as random vertical flip, random horizontal flip, and random rotations. The training epoch was set as 50. The models were trained using the stochastic gradient descent (SGD) algorithm with the following parameters: a learning rate of 0.01, a momentum of 0.9, and a decay coefficient of 0.0005. In order to determine the optimal settings of hyperparameters for our proposed MFMamba, a series of experiments were conducted to adjust and optimize various hyperparameter settings. The quantitative results of our model with different batch sizes and learning rate schedulers are illustrated in Table 1. It can be seen that the best results are obtained by setting the batch size to 32 with a learning rate scheduler set to decay at the 10th, 20th, and 30th training epochs.

4.3. Experimental Results

We have adopted seven state-of-the-art methods for performance comparison, which include two CNN-based approaches named as ABCNet [47] and MAResU-Net [48], four Transformer-based techniques named as CMFNet [31], TransUNet [21], UNetFormer [43], and CMTFNet [49], as well as one Mamba-based method named as RS3Mamba [12].

4.3.1. Comparison Results on the Vaihingen Dataset

As can be seen from Table 2, our proposed MFMamba attains the highest scores for OA and the mF1 and mIoU on the Vaihingen dataset. Notably, our MFMamba exhibits significant enhancements in terms of OA and the mIoU when compared with the baseline RS3Mamba, with increments of 0.51% and 0.57%, respectively. This validates the efficacy of our proposed dual-branch encoding architecture for extracting multi-modal features. In comparison with the existing state-of-the-art methods, MFMamba outperforms in segmenting four classes: Impervious surface, Building, Tree, and Car. In particular, the F1 score for Impervious surface is increased by 3.11% and for Building by 2.66%, outperforming the CMFNet.
Figure 6 illustrates a visualization of the results achieved by all eight methods, which also serves to highlight the effectiveness of our MFMamba. It can be observed that HRRSIs possess more intricate backgrounds and a greater abundance of intricate details in comparison with natural images. As illustrated in Figure 6, our proposed MFMamba is capable of more accurately classifying Building and Impervious surface. In addition, the classification of Tree and Car is more closely aligned with the Ground Truth. In order to highlight the segmentation results, two purple boxes are added to each subfigure in Figure 6. We can see that in the lower left box, our MFMamba splits Building arranged in it relatively completely. In the upper box, our MFMamba effectively identifies Tree around Building, providing a more tidy and complete segmentation of Tree.

4.3.2. Comparison Results on the Potsdam Dataset

The experiments conducted on the Potsdam dataset also yielded results similar to those obtained from the Vaihingen dataset. As illustrated in Table 3, the segmentation performance for Impervious surface, Building, Low vegetation, and Tree were 93.31%/87.46%, 97.81%/95.75%, 86.76%/76.62%, and 87.19%/77.28%, respectively, representing an improvement of 0.11%/0.19%, 0.49%/0.96%, 0.69%/1.08%, and 0.59%/0.91% in F1 and IoU, respectively, compared with the baseline RS3Mamba. When compared with existing state-of-the-art methods, MFMamba outperforms in segmenting three classes: Impervious surface, Building, and Low vegetation.
Figure 7 shows a visualization example from the Potsdam dataset for all eight methods under consideration. It can be observed by the highlighted two purple boxes in each subfigure that, in the upper left box, our proposed MFMamba more accurately segments the Low vegetation in Tree region and that, in the lower box, our MFMamba divides Building more completely and contains less Clutter. Clearly, our MFMamba is capable of effectively detecting intricate edges and generating smoother results. Our approach produces more comprehensive and cohesive results, reducing the number of independent points.

4.3.3. Analysis of Computational Complexity

We employed floating-point operations (FLOPs) and the number of model parameters as our evaluation criteria to evaluate the computational complexity of our proposed MFMamba. FLOPs can be used to assess the temporal complexity of a deep learning-based model, whereas the number of model parameters serves to quantify the size of the model. An ideal model would exhibit a reduction in the numbers of floating-point operations and model parameters while preserving a high standard of processing performance.
The first two columns in Table 4 present the numbers of floating-point operations and model parameters of all methods considered in this paper, and the last column shows their corresponding processing performance on the Vaihingen dataset. Table 4 shows that, although UNetFormer has the fewest FLOPs and model parameters, its mIoU score is much lower than our model. Our MFMamba significantly reduces FLOPs and requires fewer model parameters while maintaining a higher mIoU score, in comparison with the multi-modal fusion methods like TransUNet and CMFNet. This efficiency stems from using Mamba as an auxiliary branch in the encoder, which is a less resource-intensive method than using a Transformer. Compared with single-modal segmentation methods, the computational complexity of our model is slightly higher due to the introduction of multi-modal data, but our method has better segmentation performance. Furthermore, our MFMamba provides a clear improvement in segmentation performance with a slight increase in the number of model parameters compared with the baseline RS3Mamba.

4.4. Ablation Study

To verify the effectiveness of introducing the DSM data, we conducted ablation experiments on the Vaihingen dataset and the Potsdam dataset by configuring the inputs as HRRSIs only and as HRRSIs combined with the DSM data, respectively. As shown in Table 5, the proposed MFMamba can improve the accuracy of most categories on both datasets by effectively utilizing the additional DSM data. In particular, there is a significant improvement for segmenting Building, Tree, and Impervious surface, since ground objects in these three categories typically have distinct height characteristics. In addition, as cars are usually located on roads, it is also helpful to identify the car–road boundary by the height of the road. Because the elevation characteristics of Low vegetation are very similar to Impervious surface and Tree, and Low vegetation is often occluded due to its location adjacent to Tree, our proposed MFMamba encounters challenges in improving the accuracy of Low vegetation.
The effectiveness of the proposed FFB was confirmed via ablation studies on the Vaihingen dataset. These studies preserved the FFB’s architecture but altered it to incorporate just a single attention, as presented in Table 6. The results in Table 6 indicate that both the EAA and MCKA units are essential components for the FFB of the proposed MFMamba to achieve better segmentation performance.

5. Conclusions

In this paper, we proposed a novel Mamba-based multi-modal fusion network called MFMamba, which is the first bold attempt of Mamba applied in semantic segmentation of multi-modal remote sensing data. MFMamba integrates a dual-branch encoder for feature extraction from multi-modal remote sensing data, where the CNN-based main encoder is utilized to extract the features of HRRSIs and the Mamba-based auxiliary encoder is utilized to extract the DSM features, respectively. In order to better fuse the global and local features, a new feature fusion block (FFB) is designed to fuse the features extracted from the dual-branch encoder at each stage and further generate multi-scale fusion features. Compared with the methods that use CNNs or ViTs, MFMamba provides a new idea for semantic segmentation of multi-modal remote sensing data. Comprehensive experiments conducted on two public datasets, the Vaihingen and the Potsdam, demonstrate that our proposed MFMamba outperforms other seven state-of the-art methods in terms of semantic segmentation performance with low computational complexity. Further research on real-world noise data and model robustness will be carried out to continuously improve the performance of our proposed model.

Author Contributions

Conceptualization, Y.W., L.C. and H.D.; methodology, Y.W.; validation, Y.W.; data curation, Y.W.; writing—original draft preparation, Y.W.; writing—review and editing, L.C.; supervision, L.C. All authors have read and agreed to the published version of the manuscript.

Funding

This work was funded by Research Funding of Wuhan Polytechnic University No. 2023RZ036.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The code can be obtained from https://github.com/YanWang-WHPU/MFMamba (accessed on 1 October 2024). The Vaihingen and the Potsdam datasets can be obtained from https://www.isprs.org/education/benchmarks/UrbanSemLab/default.aspx (accessed on 1 October 2024).

Acknowledgments

The authors would like to acknowledge the provision of the datasets (Vaihingen and Potsdam) by the International Society for Photogrammetry and Remote Sensing.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Gao, Y.; Zhang, S.; Zuo, D.; Yan, W.; Pan, X. TMNet: A Two-Branch Multi-Scale Semantic Segmentation Network for Remote Sensing Images. Sensors 2023, 23, 5909. [Google Scholar] [CrossRef] [PubMed]
  2. Li, Y.; Zhou, Y.; Zhang, Y.; Zhong, L.; Wang, J.; Chen, J. DKDFN: Domain knowledge-guided deep collaborative fusion network for multimodal unitemporal remote sensing land cover classification. ISPRS J. Photogramm. Remote Sens. 2022, 186, 170–189. [Google Scholar] [CrossRef]
  3. Xing, J.; Sieber, R.; Caelli, T. A scale-invariant change detection method for land use/cover change research. ISPRS J. Photogramm. Remote Sens. 2018, 141, 252–264. [Google Scholar] [CrossRef]
  4. Samie, A.; Abbas, A.; Azeem, M.M.; Hamid, S.; Iqbal, M.A.; Hasan, S.S.; Deng, X. Examining the impacts of future land use/land cover changes on climate in Punjab province, Pakistan: Implications for environmental sustainability and economic growth. Environ. Sci. Pollut. Res. 2020, 27, 25415–25433. [Google Scholar] [CrossRef] [PubMed]
  5. Griffiths, D.; Boehm, J. Improving public data for building segmentation from Convolutional Neural Networks (CNNs) for fused airborne lidar and image data using active contours. ISPRS J. Photogramm. Remote Sens. 2019, 154, 70–83. [Google Scholar] [CrossRef]
  6. Salach, A.; Bakuła, K.; Pilarska, M.; Ostrowski, W.; Górski, K.; Kurczyński, Z. Accuracy assessment of point clouds from LiDAR and dense image matching acquired using the UAV platform for DTM creation. ISPRS Int. J. Geo-Inf. 2018, 7, 342. [Google Scholar] [CrossRef]
  7. Ma, X.P.; Zhang, X.K.; Pun, M.O.; Liu, M. A Multilevel Multimodal Fusion Transformer for Remote Sensing Semantic Segmentation. Ieee Trans. Geosci. Remote Sens. 2024, 62, 1–15. [Google Scholar] [CrossRef]
  8. Gao, L.; Li, J.; Khodadadzadeh, M.; Plaza, A.; Zhang, B.; He, Z.; Yan, H. Subspace-based support vector machines for hyperspectral image classification. IEEE Geosci. Remote Sens. Lett. 2014, 12, 349–353. [Google Scholar]
  9. Gislason, P.O.; Benediktsson, J.A.; Sveinsson, J.R. Random forests for land cover classification. Pattern Recognit. Lett. 2006, 27, 294–300. [Google Scholar] [CrossRef]
  10. Krähenbühl, P.; Koltun, V. Efficient inference in fully connected crfs with gaussian edge potentials. In Proceedings of the 24th International Conference on Neural Information Processing Systems, Granada, Spain, 12–15 December 2011. [Google Scholar]
  11. Gu, A.; Dao, T. Mamba: Linear-time sequence modeling with selective state spaces. arXiv 2023, arXiv:2312.00752. [Google Scholar]
  12. Ma, X.P.; Zhang, X.K.; Pun, M.O. RS3Mamba: Visual State Space Model for Remote Sensing Image Semantic Segmentation. Ieee Geosci. Remote Sens. Lett. 2024, 21, 3414293. [Google Scholar] [CrossRef]
  13. Hosseinpour, H.; Samadzadegan, F.; Javan, F.D. CMGFNet: A deep cross-modal gated fusion network for building extraction from very high-resolution remote sensing images. ISPRS J. Photogramm. Remote Sens. 2022, 184, 96–115. [Google Scholar] [CrossRef]
  14. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  15. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015: 18th International Conference, Munich, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar]
  16. Badrinarayanan, V.; Kendall, A.; Cipolla, R. Segnet: A deep convolutional encoder-decoder architecture for image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 2481–2495. [Google Scholar] [CrossRef] [PubMed]
  17. Peng, Z.; Guo, Z.; Huang, W.; Wang, Y.; Xie, L.; Jiao, J.; Tian, Q.; Conformer, Q.Y. Local features coupling global representations for recognition and detection. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 2, 9454–9468. [Google Scholar] [CrossRef] [PubMed]
  18. Marmanis, D.; Schindler, K.; Wegner, J.D.; Galliani, S.; Datcu, M.; Stilla, U. Classification with an edge: Improving semantic image segmentation with boundary detection. ISPRS J. Photogramm. Remote Sens. 2018, 135, 158–172. [Google Scholar] [CrossRef]
  19. Vaswani, A. Attention is all you need. In Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  20. Xie, E.; Wang, W.; Yu, Z.; Anandkumar, A.; Alvarez, J.M.; Luo, P. SegFormer: Simple and efficient design for semantic segmentation with transformers. Adv. Neural Inf. Process. Syst. 2021, 34, 12077–12090. [Google Scholar]
  21. Chen, J.; Lu, Y.; Yu, Q.; Luo, X.; Adeli, E.; Wang, Y.; Lu, L.; Yuille, A.L.; Zhou, Y. Transunet: Transformers make strong encoders for medical image segmentation. arXiv 2021, arXiv:2102.04306. [Google Scholar]
  22. Liu, Z.; Mao, H.; Wu, C.-Y.; Feichtenhofer, C.; Darrell, T.; Xie, S. A convnet for the 2020s. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, New Orleans, LA, USA, 18-24 June 2022; pp. 11976–11986. [Google Scholar]
  23. Wang, X.; Wang, S.; Ding, Y.; Li, Y.; Wu, W.; Rong, Y.; Kong, W.; Huang, J.; Li, S.; Yang, H. State space model for new-generation network alternative to transformers: A survey. arXiv 2024, arXiv:2404.09516. [Google Scholar]
  24. Ruan, J.; Xiang, S. Vm-unet: Vision mamba unet for medical image segmentation. arXiv 2024, arXiv:2402.02491. [Google Scholar]
  25. Ma, J.; Li, F.; Wang, B. U-mamba: Enhancing long-range dependency for biomedical image segmentation. arXiv 2024, arXiv:2401.04722. [Google Scholar]
  26. Chen, K.; Chen, B.; Liu, C.; Li, W.; Zou, Z.; Shi, Z. Rsmamba: Remote sensing image classification with state space model. IEEE Geosci. Remote Sens. Lett. 2024, 21, 8002605. [Google Scholar] [CrossRef]
  27. Liang, D.; Zhou, X.; Wang, X.; Zhu, X.; Xu, W.; Zou, Z.; Ye, X.; Bai, X. Pointmamba: A simple state space model for point cloud analysis. arXiv 2024, arXiv:2402.10739. [Google Scholar]
  28. Liu, C.; Chen, K.; Chen, B.; Zhang, H.; Zou, Z.; Shi, Z. Rscama: Remote sensing image change captioning with state space model. IEEE Geosci. Remote Sens. Lett. 2024, 21, 6010405. [Google Scholar] [CrossRef]
  29. He, X.; Cao, K.; Yan, K.; Li, R.; Xie, C.; Zhang, J.; Zhou, M. Pan-Mamba: Effective pan-sharpening with State Space Model. arXiv 2024, arXiv:2402.12192. [Google Scholar] [CrossRef]
  30. Zhu, Q.; Cai, Y.; Fang, Y.; Yang, Y.; Chen, C.; Fan, L.; Nguyen, A. Samba: Semantic segmentation of remotely sensed images with state space model. arXiv 2024, arXiv:2404.01705. [Google Scholar] [CrossRef]
  31. Ma, X.P.; Zhang, X.K.; Pun, M.O. A Crossmodal Multiscale Fusion Network for Semantic Segmentation of Remote Sensing Data. Ieee J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 3463–3474. [Google Scholar] [CrossRef]
  32. Zhang, Y.; Sidibé, D.; Morel, O.; Mériaudeau, F. Deep multimodal fusion for semantic image segmentation: A survey. Image Vis. Comput. 2021, 105, 104042. [Google Scholar] [CrossRef]
  33. Diakogiannis, F.I.; Waldner, F.; Caccetta, P.; Wu, C. ResUNet-a: A deep learning framework for semantic segmentation of remotely sensed data. ISPRS J. Photogramm. Remote Sens. 2020, 162, 94–114. [Google Scholar] [CrossRef]
  34. Audebert, N.; Le Saux, B.; Lefèvre, S. Beyond RGB: Very high resolution urban remote sensing with multimodal deep networks. ISPRS J. Photogramm. Remote Sens. 2018, 140, 20–32. [Google Scholar] [CrossRef]
  35. Hazirbas, C.; Ma, L.; Domokos, C.; Cremers, D. Fusenet: Incorporating depth into semantic segmentation via fusion-based cnn architecture. In Proceedings of the Computer Vision–ACCV 2016: 13th Asian Conference on Computer Vision, Revised Selected Papers, Part I 13, 2017. Taipei, Taiwan, 20-24 November 2016; pp. 213–228. [Google Scholar]
  36. Zhang, P.; Du, P.; Lin, C.; Wang, X.; Li, E.; Xue, Z.; Bai, X. A hybrid attention-aware fusion network (HAFNet) for building extraction from high-resolution imagery and LiDAR data. Remote Sens. 2020, 12, 3764. [Google Scholar] [CrossRef]
  37. He, X.; Zhou, Y.; Zhao, J.; Zhang, D.; Yao, R.; Xue, Y. Swin transformer embedding UNet for remote sensing image semantic segmentation. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–15. [Google Scholar] [CrossRef]
  38. Dosovitskiy, A. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  39. Prakash, A.; Chitta, K.; Geiger, A. Multi-modal fusion transformer for end-to-end autonomous driving. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, Nashville, TN, USA, 20–25 June 2021; pp. 7077–7087. [Google Scholar]
  40. He, S.; Yang, H.; Zhang, X.; Li, X. MFTransNet: A multi-modal fusion with CNN-transformer network for semantic segmentation of HSR remote sensing images. Mathematics 2023, 11, 722. [Google Scholar] [CrossRef]
  41. Wan, Z.; Wang, Y.; Yong, S.; Zhang, P.; Stepputtis, S.; Sycara, K.; Xie, Y. Sigma: Siamese mamba network for multi-modal semantic segmentation. arXiv 2024, arXiv:2404.04256. [Google Scholar]
  42. Zhang, R.; Xu, L.; Yang, S.; Wang, L. MambaReID: Exploiting Vision Mamba for Multi-Modal Object Re-Identification. Sensors 2024, 24, 4639. [Google Scholar] [CrossRef]
  43. Wang, L.B.; Li, R.; Zhang, C.; Fang, S.H.; Duan, C.X.; Meng, X.L.; Atkinson, P.M. UNetFormer: A UNet-like transformer for efficient semantic segmentation of remote sensing urban scene imagery. Isprs J. Photogramm. Remote Sens. 2022, 190, 196–214. [Google Scholar] [CrossRef]
  44. Liu, Y.; Tian, Y.; Zhao, Y.; Yu, H.; Xie, L.; Wang, Y.; Ye, Q.; Liu, Y. VMamba: Visual State Space Model. arXiv 2024, arXiv:2401.10166. [Google Scholar]
  45. Shaker, A.; Maaz, M.; Rasheed, H.; Khan, S.; Yang, M.-H.; Khan, F.S. Swiftformer: Efficient additive attention for transformer-based real-time mobile vision applications. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 1–6 October 2023; pp. 17425–17436. [Google Scholar]
  46. Cai, X.; Lai, Q.; Wang, Y.; Wang, W.; Sun, Z.; Yao, Y. Poly kernel inception network for remote sensing detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 16–22 June 2024; pp. 27706–27716. [Google Scholar]
  47. Li, R.; Zheng, S.Y.; Zhang, C.; Duan, C.X.; Wang, L.B.; Atkinson, P.M. ABCNet: Attentive bilateral contextual network for efficient semantic segmentation of Fine-Resolution remotely sensed imagery. Isprs J. Photogramm. Remote Sens. 2021, 181, 84–98. [Google Scholar] [CrossRef]
  48. Li, R.; Zheng, S.Y.; Duan, C.X.; Su, J.L.; Zhang, C. Multistage Attention ResU-Net for Semantic Segmentation of Fine-Resolution Remote Sensing Images. Ieee Geosci. Remote Sens. Lett. 2022, 19, 3063381. [Google Scholar] [CrossRef]
  49. Wu, H.L.; Huang, P.; Zhang, M.; Tang, W.L.; Yu, X.Y. CMTFNet: CNN and Multiscale Transformer Fusion Network for Remote-Sensing Image Semantic Segmentation. Ieee Trans. Geosci. Remote Sens. 2023, 61, 3314641. [Google Scholar] [CrossRef]
Figure 1. The overall architecture of our proposed MFMamba.
Figure 1. The overall architecture of our proposed MFMamba.
Sensors 24 07266 g001
Figure 2. (a) The detailed architecture of a VSS block. (b) The visualization of an SS2D unit.
Figure 2. (a) The detailed architecture of a VSS block. (b) The visualization of an SS2D unit.
Sensors 24 07266 g002
Figure 3. (a) The overall architecture of an FFB. (b) The structure of an MCKA unit. (c) The structure of an EAA unit.
Figure 3. (a) The overall architecture of an FFB. (b) The structure of an MCKA unit. (c) The structure of an EAA unit.
Sensors 24 07266 g003
Figure 4. (a) The structure of a GLTB. (b) The structure of an FRH.
Figure 4. (a) The structure of a GLTB. (b) The structure of an FRH.
Sensors 24 07266 g004
Figure 5. Samples (a,b) are 256 × 256 from Vaihingen and (c,d) are 256 × 256 from Potsdam. The first row shows the orthophotos with three channels (NIRRG for Vaihingen and RGB for Potsdam). The second and third rows show the corresponding depth information and semantic labels in pixel-wise mapping.
Figure 5. Samples (a,b) are 256 × 256 from Vaihingen and (c,d) are 256 × 256 from Potsdam. The first row shows the orthophotos with three channels (NIRRG for Vaihingen and RGB for Potsdam). The second and third rows show the corresponding depth information and semantic labels in pixel-wise mapping.
Sensors 24 07266 g005
Figure 6. Visualization of the segmentation results from different methods on the Vaihingen dataset. (a) NIRRG images, (b) DSM, (c) Ground Truth, (d) CMFNet, (e) ABCNet, (f) TransUNet, (g) UNetFormer, (h) MAResU-Net, (i) CMTFNet, (j) RS3Mamba, and (k) the proposed MFMamba. Two purple boxes are added to each subfigure to highlight the differences.
Figure 6. Visualization of the segmentation results from different methods on the Vaihingen dataset. (a) NIRRG images, (b) DSM, (c) Ground Truth, (d) CMFNet, (e) ABCNet, (f) TransUNet, (g) UNetFormer, (h) MAResU-Net, (i) CMTFNet, (j) RS3Mamba, and (k) the proposed MFMamba. Two purple boxes are added to each subfigure to highlight the differences.
Sensors 24 07266 g006
Figure 7. Visualization of the segmentation results from different methods on the Potsdam dataset. (a) RGB images, (b) DSM, (c) Ground Truth, (d) CMFNet, (e) ABCNet, (f) TransUNet, (g) UNetFormer, (h) MAResU-Net, (i) CMTFNet, (j) RS3Mamba, and (k) the proposed MFMamba. Two purple boxes are added to each subfigure to highlight the differences.
Figure 7. Visualization of the segmentation results from different methods on the Potsdam dataset. (a) RGB images, (b) DSM, (c) Ground Truth, (d) CMFNet, (e) ABCNet, (f) TransUNet, (g) UNetFormer, (h) MAResU-Net, (i) CMTFNet, (j) RS3Mamba, and (k) the proposed MFMamba. Two purple boxes are added to each subfigure to highlight the differences.
Sensors 24 07266 g007
Table 1. Quantitative results of our proposed MFMamba with different batch sizes and learning rate schedulers on the Vaihingen dataset (%). The best results are highlighted in red.
Table 1. Quantitative results of our proposed MFMamba with different batch sizes and learning rate schedulers on the Vaihingen dataset (%). The best results are highlighted in red.
Batch SizeLearning Rate SchedulerOAmF1mIoU
16[25, 35, 45]91.6090.3082.74
24[25, 35, 45]91.7490.3782.90
28[25, 35, 45]91.5090.3682.85
30[25, 35, 45]91.6790.3882.91
36[25, 35, 45]91.3890.2282.58
48[25, 35, 45]91.4790.3582.83
32[25, 35, 45]91.6490.3882.93
32[15, 25, 35]91.7190.4583.01
32[10, 20, 30]91.8190.5283.13
Table 2. Results of comparison with other methods on the Vaihingen dataset (%). The best and second-best results are respectively marked in red and blue colors.
Table 2. Results of comparison with other methods on the Vaihingen dataset (%). The best and second-best results are respectively marked in red and blue colors.
MethodBackboneF1/IoUOAmF1mIoU
Imp.Bui.Low.Tre.Car
CMFNet [31]VGG-1690.11/81.9994.51/89.6077.72/63.5690.09/81.9786.52/76.2489.3887.7978.67
ABCNet [47]ResNet-1892.08/85.3295.96/92.2479.87/66.4990.38/82.4585.61/74.8490.7988.7880.27
TransUNet [21]R50-ViT-B92.21/85.5496.10/92.4880.79/67.7790.87/83.2789.60/81.1691.2189.9182.04
UNetFormer [43]ResNet-1892.23/85.5896.34/92.9380.74/67.7091.04/83.5590.37/82.4391.2990.1482.44
MAResU-Net [48]ResNet-3492.66/86.3396.84/93.8780.57/67.4790.84/83.2289.93/81.7191.5090.1782.51
CMTFNet [49]ResNet-5092.68/86.3796.71/93.6380.47/67.3390.78/83.1190.22/82.1891.4290.1782.52
RS3Mamba [12]R18-Mamba-T92.69/86.3896.67/93.5580.54/67.4290.59/82.7990.49/82.6491.3090.2082.56
MFMamba (Ours)R18-Mamba-T93.22/87.3197.17/94.5080.63/67.5491.05/83.5890.53/82.7091.8190.5283.13
Table 3. Results of quantitative comparison on the Potsdam dataset (%). The best and second-best results are respectively marked in red and blue colors.
Table 3. Results of quantitative comparison on the Potsdam dataset (%). The best and second-best results are respectively marked in red and blue colors.
MethodBackboneF1/IoUOAmF1mIoU
Imp.Bui.Low.Tre.Car
CMFNet [31]VGG-1693.09/87.0796.90/93.9985.88/75.2686.52/76.2596.17/92.6290.7291.7185.04
ABCNet [47]ResNet-1892.90/86.7496.99/94.1686.11/75.6287.02/77.0296.31/92.8890.8291.8785.28
TransUNet [21]R50-ViT-B93.08/87.0696.88/93.9486.74/76.5987.66/78.0396.40/93.0591.0392.1585.73
UNetFormer [43]ResNet-1893.02/86.9597.14/94.4386.21/75.7686.93/76.8896.79/93.7890.8992.0285.56
MAResU-Net [48]ResNet-3493.15/87.1797.21/94.5786.73/76.5787.14/77.2196.67/93.5691.0592.1885.82
CMTFNet [49]ResNet-5093.08/87.0697.30/94.7386.32/75.9487.13/77.2096.89/93.9790.9792.1585.78
RS3Mamba [12]R18-Mamba-T93.20/87.2797.32/94.7986.07/75.5486.60/76.3796.74/93.6890.9291.9985.53
MFMamba (Ours)R18-Mamba-T93.31/87.4697.81/95.7586.76/76.6287.19/77.2896.63/93.4891.3892.3486.12
Table 4. Comparison results of computational complexity and the mIoU on the Vaihingen dataset. The best and second-best results are respectively marked in red and blue colors.
Table 4. Comparison results of computational complexity and the mIoU on the Vaihingen dataset. The best and second-best results are respectively marked in red and blue colors.
MethodFLOPs (G)Parameter (M)mIoU (%)
CMFNet [31]255.28104.0778.67
ABCNet [47]12.5813.6780.27
TransUNet [21]123.49105.3282.04
UNetFormer [43]9.4511.6982.44
MAResU-Net [48]23.0826.2882.51
CMTFNet [49]28.6830.0782.52
RS3Mamba [12]31.6543.3282.56
MFMamba (Ours)30.5962.4383.13
Table 5. Ablation study of introducing DSM data on the Vaihingen and the Potsdam datasets.
Table 5. Ablation study of introducing DSM data on the Vaihingen and the Potsdam datasets.
DatasetBandsClass OA(%)
Imp.Bui.Low.Tre.Car
VaihingenNIRRG91.5897.0180.0891.2487.86
NIRRG + DSM92.18 (+0.60)97.91 (+0.90)79.67 (−0.41)92.08 (+0.84)89.43 (+1.57)
PotsdamRGB92.6597.9488.6086.4096.69
RGB + DSM93.06 (+0.41)98.29 (+0.35)88.64 (+0.04)87.19 (+0.79)96.31 (−0.38)
Table 6. Ablation study of the proposed FFBs on the Vaihingen dataset.
Table 6. Ablation study of the proposed FFBs on the Vaihingen dataset.
MCKAEAAOA (%)mF1 (%)mIoU (%)
91.6890.4582.98
91.5090.2282.60
91.8190.5283.13
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, Y.; Cao, L.; Deng, H. MFMamba: A Mamba-Based Multi-Modal Fusion Network for Semantic Segmentation of Remote Sensing Images. Sensors 2024, 24, 7266. https://doi.org/10.3390/s24227266

AMA Style

Wang Y, Cao L, Deng H. MFMamba: A Mamba-Based Multi-Modal Fusion Network for Semantic Segmentation of Remote Sensing Images. Sensors. 2024; 24(22):7266. https://doi.org/10.3390/s24227266

Chicago/Turabian Style

Wang, Yan, Li Cao, and He Deng. 2024. "MFMamba: A Mamba-Based Multi-Modal Fusion Network for Semantic Segmentation of Remote Sensing Images" Sensors 24, no. 22: 7266. https://doi.org/10.3390/s24227266

APA Style

Wang, Y., Cao, L., & Deng, H. (2024). MFMamba: A Mamba-Based Multi-Modal Fusion Network for Semantic Segmentation of Remote Sensing Images. Sensors, 24(22), 7266. https://doi.org/10.3390/s24227266

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop