Next Article in Journal
Video-Based Human Activity Recognition Using Deep Learning Approaches
Next Article in Special Issue
Liquid Content Detection In Transparent Containers: A Benchmark
Previous Article in Journal
Monitoring the Air Quality in an HVAC System via an Energy Harvesting Device
Previous Article in Special Issue
Multi-Modality Adaptive Feature Fusion Graph Convolutional Network for Skeleton-Based Action Recognition
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

MFAFNet: A Lightweight and Efficient Network with Multi-Level Feature Adaptive Fusion for Real-Time Semantic Segmentation

1
School of Cyberspace Security (School of Cryptology), Hainan University, Haikou 570228, China
2
Department of Public Safety Technology, Hainan Vocational College of Politics and Law, Haikou 571100, China
3
School of Computer Science and Technology, Hainan University, Haikou 570228, China
4
Faculty of Engineering, The University of Sydney, Camperdown, NSW 2006, Australia
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(14), 6382; https://doi.org/10.3390/s23146382
Submission received: 5 June 2023 / Revised: 4 July 2023 / Accepted: 11 July 2023 / Published: 13 July 2023
(This article belongs to the Special Issue Artificial Intelligence in Computer Vision: Methods and Applications)

Abstract

:
Currently, real-time semantic segmentation networks are intensely demanded in resource-constrained practical applications, such as mobile devices, drones and autonomous driving systems. However, most of the current popular approaches have difficulty in obtaining sufficiently large receptive fields, and they sacrifice low-level details to improve inference speed, leading to decreased segmentation accuracy. In this paper, a lightweight and efficient multi-level feature adaptive fusion network (MFAFNet) is proposed to address this problem. Specifically, we design a separable asymmetric reinforcement non-bottleneck module, which designs a parallel structure to extract short- and long-range contextual information and use optimized convolution to increase the inference speed. In addition, we propose a feature adaptive fusion module that effectively balances feature maps with multiple resolutions to reduce the loss of spatial detail information. We evaluate our model with state-of-the-art real-time semantic segmentation methods on the Cityscapes and Camvid datasets. Without any pre-training and post-processing, our MFAFNet has only 1.27 M parameters, while achieving accuracies of 75.9% and 69.9% mean IoU with speeds of 60.1 and 82.6 FPS on the Cityscapes and Camvid test sets, respectively. The experimental results demonstrate that the proposed method achieves an excellent trade-off between inference speed, segmentation accuracy and model size.

1. Introduction

Semantic segmentation is a basic but challenging task in computer vision that aims to predict the class label of each pixel point in an image. It plays a crucial role in the deep understanding of various objects in urban street scenes and is widely used in areas such as autonomous driving [1], robotics perception [2], and medical image analysis [3]. Robustness and accuracy are crucial for vision tasks [4]. Traditional image segmentation methods, such as conditional random field and k-mean clustering, have poor robustness and low accuracy, and it is difficult to extract semantic and spatial features of objects stably. Deep learning techniques can address this problem better. In recent years, since deep learning methods are rapidly developing and have achieved remarkable results in the field of engineering [5,6,7,8,9], semantic segmentation models based on deep learning have gradually become mainstream. For applications such as self-driving cars and mobile devices that have limited computational resources and require efficient interactions, strict inference speed and segmentation accuracy are usually required to achieve fast actions. While complex semantic segmentation models have excellent performance in terms of segmentation accuracy, they suffer from the problems of high computational power and slow inference speed, which cannot meet the needs of low-performance IoT devices for real-time tasks. Therefore, it is necessary to design an efficient and lightweight real-time semantic segmentation method to achieve a better speed–accuracy trade-off.
To fulfill the demand of real-time interaction, real-time semantic segmentation methods have emerged in recent years and achieved remarkable results. These methods can be broadly classified into two categories: (1) Module optimization first extracts features based on a lightweight real-time backbone network, and later builds an efficient context information aggregation module to improve segmentation accuracy. For example, FPANet [10] uses ResNet18 as the feature encoder, and then builds the feature pyramid fusion module (FPFM) in the decoding stage to fully fuse the output features at different levels. DFFNet [11] and MIFNet [12] are all based on the lightweight MobileNetV2 network as feature encoder, and in order to capture multi-scale contextual information to improve the segmentation accuracy of the network, propose the lightweight semantic pyramid module and lightweight information fusion module, respectively. (2) Architecture design uses optimized convolution to design bottleneck module to reduce the computational cost of the network and improve the segmentation accuracy. For example, ESNet [13], LRDNet [14], and LAANet [15] build network architectures using asymmetric convolution, dilated convolution, and depth-wise convolution to extract context information with fewer network parameters and computational cost. Although the above works build lightweight models that effectively reduce the computational overhead, for the former, the modules provide limited performance improvements and the number of network parameters is difficult to reduce further. For the latter, the optimized convolution reduces the feature extraction capability of the network, which leads to a significant reduction in segmentation accuracy. Therefore, achieving a balance between model size, inference speed, and segmentation accuracy is still an open research problem.
To solve the problem of the balance between inference speed, model size and segmentation accuracy, we propose a lightweight and efficient network with multi-level feature adaptive fusion for real-time semantic segmentation, called MFAFNet. The MFAFNet is an asymmetric encoder–decoder structure. In the encoder, we design an efficient separable asymmetric reinforcement non-bottleneck (SAR-nbt) module, which first captures short- and long-range features through a parallel structure, and then fuses the multi-scale features and enhances the representations. The parallel structure takes advantage of asymmetric convolution, depth-wise convolution and dilated convolution to reduce the number of parameters and computational overhead of the network. In the decoder, we propose a feature adaptive fusion module (FAFM) that fuses low-level, mid-level, and high-level features to reduce the loss of spatial detail information caused by multiple down-sampling. To reduce the semantic gaps existing between multi-level features, the channel attention and spatial attention representation maps of the mid-level features are constructed, which are applied to the low-level and high-level features to draw closer the relationship between the three branches. Extensive quantitative and qualitative evaluations demonstrate that the proposed approach achieves competitive results on the Cityscapes test set compared to state-of-the-art real-time semantic segmentation methods as shown in Figure 1.
The main contributions of this paper are summarized as follows:
  • We propose a separable asymmetric reinforcement non-bottleneck module. This module makes full use of the optimized convolution to reduce the number of parameters and increase the inference speed, and constructs a parallel structure to extract short- and long-range information to obtain multi-scale features. Then, the channel attention weights of the multi-scale feature maps are computed for feature re-calibration.
  • We propose a feature adaptive fusion module that effectively balances the multi-level feature maps to reduce the loss of spatial detail information. It also combines spatial attention and channel attention that adaptively enhances multi-level features to reduce the semantic gaps existing in multiple resolution feature maps.
  • We design a real-time semantic segmentation network to achieve competitive segmentation accuracy with lower computational overhead and fewer network parameters, achieving a trade-off between inference speed, segmentation accuracy and model size.
The remainder of this paper is organized as follows. Section 2 reviews the related work. Section 3 presents the design details of our proposed MFAFNet and related components. Section 4 conducts the experiments and analyzes the results. Section 5 concludes the work.

2. Related Work

2.1. Real-Time Semantic Segmentation

Real-time semantic segmentation aims to achieve high-quality segmentation results with low computational overhead and is often used in IoT devices that lack computational resources. In recent years, there have been many deep learning works devoted to real-time semantic segmentation research. ESNet [13], ERFNet [16], and LEANet [17] constructed encoding and decoding architectures with asymmetric convolution and dilated convolution to further reduce the number of network parameters to maintain a balance of speed and accuracy. Fast-SCNN [18] and BiSeNetV2 [19] proposed a two-branch or multi-branch network that captures semantic information using deep semantic branches, reduces the loss of spatial details using information branches, and then fully integrates high-level semantic and spatial detail information to improve the segmentation accuracy with lower computational overhead. STDC [20] modified and optimized the backbone network of BiSeNet by designing a more efficient short-term dense connectivity module, which reduces the computational effort in the network while extracting rich feature information. Although these methods can achieve high accuracy and real-time inference speed, there is still room for further improvement on the real-time semantic segmentation task.

2.2. Multi-Scale Feature Fusion

Different layers in the segmentation model contain different semantic information. The shallow network contains rich spatial detail information, and the deep network has a strong ability to characterize semantic information. Fusing the features of deep and shallow layers can effectively improve the accuracy of the network [21]. In addition, global semantic information can provide clues to segment the category distribution, and fusing global and local features can improve the robustness of the model.
The current multi-scale feature fusion methods mainly include two types: parallel multi-branching structures and jump-connected structures. The former is mainly used to fuse local and global features, for example, PSPNet [22] proposed a pyramid pooling module (PPM), which extracted multi-scale features by pooling operations at different scales to reduce the loss of contextual information characterizing the relationships between different sub-regions. DeepLabv3 [23] designed an atrous spatial pyramid pooling (ASPP) module that captured multi-scale contexts with different dilated convolution expansion rates to further integrate multi-scale features. MIFNet [12] and DFFNet [11] are optimized for the ASPP module, utilizing operations such as dense join, depth-wise convolution, and dilated convolution to efficiently integrate multi-scale information. The latter is mainly used to fuse shallow and deep features, e.g., LAANet [15] proposed the attention-guided feature fusion upsampling (AFFU) module, which combined the spatial attention map generated by shallow features and the channel attention map generated by deep features to efficiently fuse the output features of different layers in the encoding stage. EFRNet [24], FPANet [10] and JPANet [25] reduced the loss of deep feature spatial detail information by constructing efficient feature fusion modules to aggregate outputs of different depths in the network with strong semantic information.

3. Proposed Method

In this section, we first overview the architecture of the proposed network. Then, the design details of the main components of the network, the separate asymmetric reinforcement non-bottleneck (SAR-nbt) module and the feature adaptive fusion module (FAFM) are presented.

3.1. Network Overview

The overall network architecture of MFAFNet is shown in Figure 2, which is an asymmetric encoder–decoder structure. The encoder contains four down-sampling operations, decreasing the resolution to 1/16 to reduce the computational overhead of the feature extraction process. The decoder contains a feature adaptive fusion module (FAFM), which fuses the output features of multi-levels in the encoding stage to reduce the loss of spatial detail information and improve the segmentation accuracy.
The design details of the encoder are shown in Figure 3, which consists of four stages. Stage 1 uses a 3 × 3 standard convolution with stride of 2 to achieve the first down-sampling, which quickly reduces the resolution of the input features and reduces the computational complexity. The 2nd, 3rd and 4th stages are composed of SAR-nbt modules, containing 2, 3, 9 SAR-nbt modules, respectively, and down-sampling only happens in the first SAR-nbt module of each stage. To avoid weakening the feature-extraction capability of the network due to the use of depth-wise convolution, the channel expansion is added to the SAR-nbt module, and the expansion factor is 4 only for the last one in stage 3 and stage 4, and 2 for the others so as to avoid a too-wide channel dimension that would substantially increase the computational cost. In addition, in order to expand the receptive field to extract context information at a distance and to avoid losing semantic information by introducing dilated convolution prematurely, dilated convolution is introduced in the middle layer of the network. The SAR-nbt module in stage 4 introduces dilated convolution to gradually expand the receptive field, and the dilation rates are set to 2, 4, 6, 8, 10, 12, 12, 12, and 12, respectively.
For faster reasoning speed, many networks directly use bilinear interpolation methods to recover deep semantic features to the original image resolution, and although this operation improves efficiency, it leads to the loss of spatial detail information and loss of accuracy. While the decoder structure of the progressive layer-by-layer fusion of shallow features performs better in terms of segmentation accuracy, the negative effects, such as slower inference and increased computational cost, are also evident. Therefore, we design a feature adaptive fusion module in the decoder, as shown in Figure 5, to efficiently fuse multi-level features to compensate for the loss of spatial information caused by the high abstraction of deep features and multiple down-sampling, and to significantly improve the segmentation accuracy with only a small computational overhead. Finally, we use 1 × 1 convolution and bilinear interpolation to recover the features to the original input size to predict the final segmentation result.

3.2. Separable Asymmetric Reinforcement Non-Bottleneck Module

In recent years, many lightweight real-time semantic segmentation networks [14] have adopted the residual structure when designing bottleneck blocks and used optimized convolution to replace the standard convolution of the original residual blocks to reduce the number of network parameters and computational overhead.
Optimized convolution includes two categories: factorized convolution and dilated convolution. Among them, factorized convolution is mainly used to reduce the number of parameters and the computational overhead, such as group convolution [26], asymmetric convolution [16], and depth-wise separable convolution [27]. Dilated convolution [23] can effectively expand the receptive field of the convolution kernel to capture more contextual information without increasing the number of parameters. In addition, there are many works that combine multiple optimized convolutions to further improve the performance of convolution, such as dilated depth-wise separable convolution [11], asymmetric depth-wise separable convolution [14], and asymmetric dilated depth-wise separable convolution [15]. Assuming that the size of the input feature map is H × W × C i n , the size of the convolution kernel is k × k , the size of the output feature map is H × W × C o u t , and the dilation rate is d, regardless of bias. The comparison of the parameters and receptive fields between several common optimized convolutions and standard convolutions is shown in Table 1. Obviously, the optimized convolution can effectively reduce the number of parameters of the residual blocks, and obtain a larger receptive field. Inspired by the success of aforementioned works, we take advantage of asymmetric dilated depth-wise separable convolution to design a separable asymmetric reinforcement non-bottleneck (SAR-nbt) module as shown in Figure 4.
Figure 4. The structure of the proposed separable asymmetric reinforcement non-bottleneck (SAR-nbt) module. “DWConv” indicates depth-wise convolution. “DDWConv” indicates dilated depth-wise convolution.
Figure 4. The structure of the proposed separable asymmetric reinforcement non-bottleneck (SAR-nbt) module. “DWConv” indicates depth-wise convolution. “DDWConv” indicates dilated depth-wise convolution.
Sensors 23 06382 g004
The SAR-nbt module is mainly implemented in six steps. Each convolutional layer in the SAR-nbt module is followed by one batch normalization layer and one PReLU activation layer (the PReLU activation function has been demonstrated in many works to perform excellent in lightweight networks). First, to avoid the problem of feature encoding capability degradation due to the use of depth-wise convolution in the network, the SAR-nbt module first uses 1 × 1 convolution to expand the number of channels, expanding the channels to k times the original channels to improve the feature extraction capability. Second, in order to reduce the computational complexity of the network, a channel-splitting operation is used to divide the boosted feature maps into two equal groups along the channels: G 0 and G 1 . Third, multi-scale features are extracted by two parallel groups. Many works demonstrated that fusing multi-scale features can effectively improve segmentation accuracy, so different scale features are extracted from two groups separately. G 0 uses an asymmetric depth-wise convolution to capture short-range context information. Specifically, a standard 3 × 3 depth-wise convolution is decomposed into a 1 × 3 convolution and a 3 × 1 convolution, an operation that reduces the number of network parameters without reducing the receptive field. G 1 further incorporates dilated convolution to effectively expand the receptive field and extract long-range context information without increasing the computational overhead. Specially, when the stride of the SAR-nbt module is set to 2, it indicates that the stride of the 1 × 3 convolution in G 0 and G 1 is 2, which achieves down-sampling of the input feature maps. After that, the features of the two groups are fused using the channel concatenate operation to obtain rich multi-scale features and adapt to the recognition of objects of different sizes.
Fourth, the multi-scale feature maps are enhanced by using channel attention SENet. Specifically, the channel-wise attention vector is obtained by using the SEWeight module to extract the attention of the feature maps with different scales, and the attention vector is re-calibrated by using the Softmax. Then, the re-calibrated weights are multiplied with the corresponding feature maps to achieve the enhancement of key features and obtain the refined feature maps. Fifth, the number of channels of the refined feature maps is compressed using a 1 × 1 point-wise convolution to keep the same number of channels as the input feature maps. Finally, a residual branch is constructed to fuse the initial input with the output to solve the problem of gradient disappearance and explosion during the training process. Specifically, when the stride of the SAR-nbt module is 1, the residual branch represents the input feature map. When the stride = 2, the residual branch contains an average pooling layer, a 1 × 1 convolutional layer, and a batch normalization layer to act on the input to keep the resolution of the residual branch consistent with the output feature map.

3.3. Feature Adaptive Fusion Module

High-level features contain rich semantic information and have high abstract generalization; low-level features have rich spatial detail information and can guide object position features. Making full use of the deep semantic information and the shallow spatial detail information can effectively improve the segmentation accuracy. Most existing real-time semantic segmentation methods use element addition or channel concatenation operations when performing feature fusion, ignoring the existence of semantic gaps in features at different levels. To effectively fuse information from multi-levels, we propose a feature adaptive fusion module (FAFM), which adaptively enhances features at different levels through an attention mechanism to reduce the semantic gap existing between low-level and high-level features.
The FAFM module is shown in Figure 5. H and W represent the height and width of the input RGB image. The output features of stage 2, stage 3, and stage 4 are used in the fusion process, corresponding to the low-level feature F l , the mid-level feature F m , and the high-level feature F h , and the feature map sizes are ( W / 4 ) × ( H / 4 ) × 48 , ( W / 8 ) × ( H / 8 ) × 64 , and ( W / 16 ) × ( H / 16 ) × 160 , respectively. To avoid excessive computational overhead, an element addition operation is used to fuse multi-level features. And the number of channels and resolution of each level feature maps are different; therefore, the consistency processing of each level feature is needed before feature fusion. Considering that the mid-level features have better semantic information and also retain spatial information, the number of channels and resolution of the middle-level features are chosen as the benchmark. The low-level features are down-sampled using the averaging pooling operation, and the number of channels is expanded using 1 × 1 point-wise convolution, while the high-level features are up-sampled using the bilinear interpolation method. The number of channels is compressed using 1 × 1 point-wise convolution, and the processed low-level feature F l and high-level feature F h are obtained in turn, both with feature map size ( W / 8 ) × ( H / 8 ) × 64 . The above operations can be expressed as follows:
F l = f 1 × 1 ( D o w n ( F l ) )
F h = f 1 × 1 ( U p ( F h ) )
where D o w n and U p indicate the down-sampling and up-sampling operations, respectively. f k × k refers to the standard convolution with a kernel size of k, with the same meaning for subsequent references.
Figure 5. Design details of feature adaptive fusion module.
Figure 5. Design details of feature adaptive fusion module.
Sensors 23 06382 g005
Then, to reduce the semantic gaps existing in different level features, the channel attention weights W c and spatial attention weights W s of mid-level features are obtained to enhance the representations of other level features. The attention weights W c and W s are calculated as follows:
W c = σ ( f 1 × 1 ( R ( f 1 × 1 ( P m a x s ( F m ) + P a v g s ( F m ) ) ) ) )
W s = σ ( f 3 × 3 ( C [ P m a x c ( F m ) , P a v g c ( F m ) ] ) )
where P m a x s and P a v g s , and P m a x c and P a v g c represent the two sets of max-pooling and average-pooling operations along the spatial and channel axes, respectively. C [ . , . ] represents the channel concatenation operation. R and σ represent the ReLU activation function and the sigmoid activation function, respectively. The ReLU and sigmoid activation functions are simple to implement and easy to optimize, and are still used by most attention mechanisms to compute attention weights.
Then, both channel attention weights W c and spatial attention weights W s are multiplied with low-level features and high-level features to achieve the adaptive adjustment of features at different levels. Finally, the multi-level features are element-wise added and fused, and the 3 × 3 convolutional layer is used to make the feature fusion more fully, and the final output of the FAFM module is obtained. The final features F o u t can be obtained according to the following formula:
F o u t = f 3 × 3 ( F l × W c × W s + F m + F h × W c × W s )
Obviously, the proposed FAFM module uses an attention mechanism to reduce the semantic gaps that exist between multi-levels, effectively balances feature maps with multiple resolutions, and improves segmentation accuracy.

4. Results

4.1. Datasets and Evaluation Metrics

4.1.1. Datasets

Cityscapes dataset [28]. The Cityscapes dataset is a large-scale image dataset on the semantic understanding of urban street scenes. It is mainly derived from a set of video sequences recorded in 50 different urban driving scenes and consists of 5000 high-quality finely annotated images and 20,000 coarsely annotated images. The dataset contains semantic, instance and dense pixel annotations for 30 categories, 19 of which are used for training and evaluation. We use only 5000 finely annotated images for experiments, of which 2975 are used for training, 1525 for testing, and 500 for validation.
Camvid dataset [29]. Camvid is another dataset for road and driving scene understanding. It contains a total of 701 images, which can be divided into three groups, the training set, validation set, and test set, each containing 367, 101, and 233 images. The dataset contains 11 different classes, and the resolution is 960 × 720.

4.1.2. Evaluation Metrics

Accuracy. The accuracy of semantic segmentation is mainly measured by the mean intersection over union (MIoU), where MIoU refers to the mean of the ratio between the intersection and the union of the predicted segmentation maps and the true labels for all categories. Assuming that k is the number of segmentation categories, MIoU can be defined as follows:
M I o U = 1 k i = 0 k P i i j = 0 k P i j + j = 0 k P j i P i i
where P i j denotes the number of pixels inferred from category i to category j, and P i i denotes the true examples. P i j and P j i refer to false-positive and false-negative examples, respectively.
Inference speed. FPS is an image domain definition of the number of frames per second of the transmitted images, which is often used to indicate the real-time performance of a model.
Network parameters. The total number of weights and biases for each layer in the network.
Computational complexity. It is used to measure the number of floating-point operations (FLOPs) in the model to evaluate the speed and efficiency of the model.

4.2. Implementation Details

We used Pytorch 1.9 and NVIDIA 3090 GPU for all experiments. We followed the strategies proposed by RegSeg [30]. We utilized the “poly” learning rate policy, in which the initial rate is multiplied by ( 1 i t e r m a x i t e r ) p o w e r . The power was set to 0.9, the initial learning rate was set as 0.05, and the weight decay was 0.0005. In addition, a linear warm-up with a learning rate of 0.1 to 1 time was used for the first 3000 iterations during training. The data-augmentation method included random horizontal flip, random scaling to [400, 1600] and random cropping to [768, 768]. It also used a set of reduced random augmentation operations (auto contrast, equalization, rotation, color, contrast, brightness, and sharpness), and two random operations of the 0.2 level were applied to each image. To accelerate the training speed without losing accuracy, mixed precision training was used. The loss function used in the training process is the online hard example mining loss (OHEM) [31], the batch size is 8, and a total of 1000 epochs were trained. When submitting to the test server, we trained on the trainval set.
When training on the Camvid dataset, the training strategy was similar to that on the Cityscapes dataset. Since the model is not pre-trained on other datasets, and the Camvid dataset is too small, it may not be possible to fit the network. Therefore, the training results on the Cityscapes dataset were used as pre-training weights for the Camvid dataset. The data were augmented using random level flipping, random scaling to [288, 1152], with a batch size of 12. A total of 200 epochs were trained without using the random augmentation strategy.

4.3. Ablation Study

In this section, we design a series of experiments to demonstrate the effectiveness of our method. All ablation experiments were conducted on the Cityscapes dataset.

4.3.1. Ablation for SAR-nbt Module

To study the effect of the attention mechanism on the feature extraction ability of the network, we first conducted the SE attention ablation study on the backbone network, and the results are shown in Table 2. The SAR-nbt module that used the SE attention mechanism improved the accuracy by 1.51% with only 0.28 M additional parameters, and achieved a segmentation accuracy of 77.08% on the Cityscapes validation set, which indicated that the simple attention mechanism can significantly improve the ability to extract contextual information with a small amount of computational cost.
In addition, to explore the effect of dilated convolution and to select the appropriate dilation rates to better capture long-range semantic information and improve model performance, we designed dilation rate ablation experiments for the SAR-nbt module. We only used the dilated convolution in the SAR-nbt module of stage 4, and the left branch captured the short-range features by default, so only the convolution block of the right branch was filled with the dilation rate. Table 2 shows the comparison results under different dilation rates. It can be seen that the dilated convolution did not increase the number of network parameters, and the segmentation accuracy was highest for the settings with the dilation rate list of 2, 4, 6, 8, 10, 12, 12, 12, and 12. The segmentation accuracy was the lowest for the settings with dilation rates of 2, 6, 6, 12, 12, 18, 18, 24, 24. Due to the large dilation rates, too much noise information may be introduced, which leads to the poor information capture ability of the convolutional kernel and reduces the segmentation accuracy. Finally, the proposed MFAFNet used the highest segmentation accuracy with the dilation rate list setting, and the ablation of subsequent modules was performed based on this setting.

4.3.2. Ablation for FAFM Module

To verify the efficiency of the proposed FAFM module, ablation experiments with different branch fusion strategies were performed. In order to reduce the training time, the ablation experiments of the FAFM module trained 500 epochs. Table 3 shows the comparison results of different fusion methods, where x16, x8, and x4 indicate the down-sampling multiplier of the feature map of the branch compared to the original input image size, corresponding to the output of stage 4, stage 3, and stage 2, respectively. As shown in Table 3, the strategy with branch fusion showed a significant improvement in segmentation accuracy compared to predicting x16 (without the fusion module) directly, which proves that the shallow features can effectively compensate for the loss of spatial detail information of the deep features. And the FAFM module had the highest accuracy, compared with the direct prediction of x16 branch features; only 0.05 M number of parameters and 0.36 G floating point operations were increased, and the MIoU was improved by 4.41% .
Figure 6 displays the visualization results of the FAFM module on the Cityscapes validation set, with the red dashed boxes marking the parts of the same image where the segmentation accuracy contrast is more obvious. As can be seen, for the results without the FAFM module, such as the sky in the first image, the car in the third image, and the traffic sign in the fourth image, after a series of encoding and decoding operations, the result prediction directly by the high-level feature maps lost the original spatial detail information of the object. The method that uses the FAFM module may also suffer from the problem of inconspicuous object edge segmentation, but it can basically retain the original spatial information of the objects and achieve better segmentation results. In summary, the proposed FAFM module can efficiently fuse the multi-level output feature maps in the encoder to reduce false predictions and substantially improve the accuracy of the object segmentation with less computational overhead.
Finally, we conducted comparative ablation experiments of the modules to verify the effectiveness of each module in the model structure. We considered the backbone without using the dilated convolution (dilation rate of 1 in all SAR-nbt modules) and without the FAFM module as the baseline. As shown in Table 4, compared with the baseline method, the SAR-nbt (with dilated convolution ) and FAFM module increased only 0.05 M parameters and 0.36 GFLOPs, which was able to achieve a 9.78% improvement in MIoU. In addition, to clearly illustrate the training process, Figure 7 intuitively shows the validation loss curve and accuracy curve of MFAFNet on the Cityscapes validation set. The loss curve dropped smoothly and converged eventually, indicating that our MFAFNet can be well trained. In general, qualitative and quantitative analyses proved the effectiveness of SAR-nbt and the FAFM module. The proposed MFAFNet can achieve better efficiency and an accurate balance.

4.4. Comparison with Other State-of-the-Art Methods

4.4.1. Results on Cityscapes

This section reports some performance results of MFAFNet on the Cityscapes dataset, such as model size, accuracy, inference speed, etc. It is then compared and analyzed with other state-of-the-art real-time semantic segmentation methods to illustrate the efficiency of MFAFNet. In this section, GTX 1080Ti GPU is used uniformly for testing the inference speed.
Table 5 shows the comparison results; our proposed MFAFNet was not pre-trained with any other dataset and contained only 1.27 M number of parameters, achieving 75.9% MIoU on the Cityscapes test set. Moreover, for input images with a resolution of 512 × 1024 , MFAFNet achieved an inference speed of 60.1 FPS on a single 1080Ti GPU. Compared with the excellent real-time semantic segmentation methods in recent years, our MFAFNet achieved a more superior performance. Specifically, compared with MIFNet [12], DSANet [32], LightSeg [33], and CSRNet-light [34], the proposed MFAFNet not only has a more lightweight network structure and faster inference speed but also has a higher segmentation accuracy. Our method was also 5.8% and 2.3% higher than LRDNet [14] and LAANet [15] in MIOU. Compared with SwiftNet [35] and FPANet [10], which also have high accuracy, our MFAFNet has the highest accuracy with at least 9 times fewer parameters and is not pre-trained on other large datasets. In conclusion, compared with other segmentation methods, our proposed MFAFNet has superior performance, and achieved real-time inference speed while significantly improving the segmentation accuracy with a smaller number of parameters, and achieved a better trade-off between model size, inference speed and segmentation accuracy.
We visually compared the prediction results of different methods on the Cityscapes validation set as shown in Figure 8. To facilitate comparison, we use red boxes in Figure 8 to mark areas where segmentation errors are more obvious among all methods. It can be seen that our algorithm is very competitive among the current excellent algorithms.
In Table 6, we further evaluated the IoU (%) performance for each category on the Cityscapes test set. Our method achieved better segmentation results on most of the categories, with the highest IoU scores in 16 out of 19 categories, such as roads, buildings, cars, skies, and traffic lights. Our method has better recognition ability for objects of different sizes with an average accuracy of 75.9% for the categories, which fully illustrates the efficiency of the MFAFNet.

4.4.2. Results on Camvid

In this subsection, to verify the robustness and effectiveness of the model, we evaluate the performance of our method on the Camvid test set. And then, it is compared with several state-of-the-art real-time semantic segmentation models. The comparison results are shown in Table 7, and it can be seen that our method can still achieve better segmentation results on the Camvid dataset. Our method is 4.4% higher than CGNet [41] in MIOU. In addition, the proposed method has the highest accuracy and is equally lightweight compared to recent real-time segmentation methods, such as MIFNet [12], JPANet [25], and LEANet [17]. Therefore, our method still maintains good accuracy (69.9% MIoU), a lightweight model size (less than 1.3 M) and fast inference speed (82.6 FPS on a 640 × 360 resolution image) compared to most existing semantic segmentation methods.

5. Conclusions

In this paper, a lightweight and efficient multi-level feature adaptive fusion network for real-time semantic segmentation is proposed, named MFAFNet. The MFAFNet is an asymmetric encoding and decoding structure, and in the encoder, we design a separable asymmetric enhancement non-bottleneck (SAR-nbt) module that makes full use of the optimized convolution to increase the inference speed, and constructs a parallel structure to extract short- and long-range information to obtain multi-scale features. In the decoder, we propose a feature adaptive fusion module that effectively balances the multi-level feature maps to reduce the loss of spatial detail information. It also combines spatial attention and channel attention to adaptively enhance multi-level feature maps and reduce the semantic gaps that exist at multi-levels during feature fusion. To verify the effectiveness of the proposed MFAFNet, we evaluate the performance on two challenging datasets with state-of-the-art real-time semantic segmentation methods. Experimental results demonstrate that our method achieves a trade-off between inference speed, segmentation accuracy, and model size.
Although the proposed MFAFNet achieves excellent segmentation accuracy in the field of real-time semantic segmentation and can basically restore the spatial location of the object, there is still room for further improvement in the segmentation of the object edges. Our method only extracts the spatial domain features of the image but ignores the effect of the frequency domain information. The high-frequency component of the frequency domain information contains rich edge outline information, which can be used reasonably to effectively improve the segmentation accuracy. In the future, our research will focus on how to combine frequency domain learning techniques in the network to further improve the accuracy of object edge segmentation.

Author Contributions

Conceptualization, K.L. and J.C.; methodology, K.L. and H.L.; validation, K.L. and H.L.; writing—original draft preparation, K.L., H.L. and T.O.; writing—review and editing, K.L. and H.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by Hainan Provincial Natural Science Foundation of China (Grant No. 620MS021, 621QN211, 621RC1082, 622RC720), the Key Research and Development Program of Hainan Province (Grant No. ZDYF2021GXJS003, ZDYF2020040), the Major science and technology project of Hainan Province (Grant No. ZDKJ2020012), National Natural Science Foundation of China (NSFC) (Grant No. 62162022, 62162024), Science and Technology Development Center of the Ministry of Education Industry-university-Research Innovation Fund (2021JQR017).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Xiao, X.; Zhao, Y.; Zhang, F.; Luo, B.; Yu, L.; Chen, B.; Yang, C. BASeg: Boundary aware semantic segmentation for autonomous driving. Neural Netw. 2023, 157, 460–470. [Google Scholar] [CrossRef] [PubMed]
  2. Peng, H.; Zhong, J.; Liu, H.; Li, J.; Yao, M.; Zhang, X. ResDense-focal-DeepLabV3+ enabled litchi branch semantic segmentation for robotic harvesting. Comput. Electron. Agric. 2023, 206, 107691. [Google Scholar] [CrossRef]
  3. Song, E.; Zhan, B.; Liu, H.; Cetinkaya, C.; Hung, C. NMNet: Learning multi-level semantic information from scale extension domain for improved medical image segmentation. Biomed. Signal Process. Control 2023, 83, 104651. [Google Scholar] [CrossRef]
  4. Zhang, J.; Jin, X.; Sun, J.; Wang, J.; Sangaiah, A. Spatial and semantic convolutional features for robust visual object tracking. Multimed. Tools Appl. 2020, 79, 15095–15115. [Google Scholar] [CrossRef]
  5. Zhao, K.; Hu, J.; Shao, H.; Hu, J. Federated multi-source domain adversarial adaptation framework for machinery fault diagnosis with data privacy. Reliab. Eng. Syst. Saf. 2023, 236, 109246. [Google Scholar] [CrossRef]
  6. Zhao, K.; Jia, F.; Shao, H. A novel conditional weighting transfer Wasserstein auto-encoder for rolling bearing fault diagnosis with multi-source domains. Knowl. Based Syst. 2023, 262, 110203. [Google Scholar] [CrossRef]
  7. Jin, B.; Cruz, L.; Gonçalves, N. Deep facial diagnosis: Deep transfer learning from face recognition to facial diagnosis. IEEE Access 2020, 8, 123649–123661. [Google Scholar] [CrossRef]
  8. Liu, Y.; Wang, K.; Liu, L.; Lan, H.; Lin, L. TCGL: Temporal contrastive graph for self-supervised video representation learning. IEEE Trans. Image Process. 2022, 31, 1978–1993. [Google Scholar] [CrossRef]
  9. Deng, Y.; Lv, J.; Huang, D.; Du, S. Combining the theoretical bound and deep adversarial network for machinery open-set diagnosis transfer. Neurocomputing 2023, 548, 126391. [Google Scholar] [CrossRef]
  10. Wu, Y.; Jiang, J.; Huang, Z.; Tian, Y. FPANet: Feature pyramid aggregation network for real-time semantic segmentation. Appl. Intell. 2022, 52, 3319–3336. [Google Scholar] [CrossRef]
  11. Tang, X.; Tu, W.; Li, K.; Cheng, J. DFFNet: An IoT-perceptive dual feature fusion network for general real-time semantic segmentation. Inf. Sci. 2021, 565, 326–343. [Google Scholar] [CrossRef]
  12. Cheng, J.; Peng, X.; Tang, X.; Tu, W.; Xu, W. MIFNet: A lightweight multiscale information fusion network. Int. J. Intell. Syst. 2022, 37, 5617–5642. [Google Scholar] [CrossRef]
  13. Wang, Y.; Zhou, Q.; Xiong, J.; Wu, X.; Jin, X. ESNet: An efficient symmetric network for real-time Semantic segmentation. In Proceedings of the Pattern Recognition and Computer Vision—Second Chinese Conference (PRCV), Xi’an, China, 8–11 November 2019; pp. 41–52. [Google Scholar]
  14. Zhuang, M.; Zhong, X.; Gu, D.; Feng, L.; Zhong, X.; Hu, H. LRDNet: A lightweight and efficient network with refined dual attention decorder for real-time semantic segmentation. Neurocomputing 2021, 459, 349–360. [Google Scholar] [CrossRef]
  15. Zhang, X.; Du, B.; Wu, Z.; Wan, T. LAANet: Lightweight attention-guided asymmetric network for real-time semantic segmentation. Neural Comput. Appl. 2022, 34, 3573–3587. [Google Scholar] [CrossRef]
  16. Romera, E.; Alvarez, J.M.; Bergasa, L.; Arroyo, R. Erfnet: Efficient residual factorized convnet for real-time semantic segmentation. IEEE Trans. Intell. Transp. Syst. 2018, 19, 263–272. [Google Scholar] [CrossRef]
  17. Zhang, X.; Du, B.C.; Luo, Z.; Ma, K. Lightweight and efficient asymmetric network design for real-time semantic segmentation. Appl. Intell. 2022, 52, 564–579. [Google Scholar] [CrossRef]
  18. Poudel, R.P.K.; Liwicki, S.; Cipolla, R. Fast-SCNN: Fast semantic segmentation network. In Proceedings of the 30th British Machine Vision Conference (BMVC), Cardiff, UK, 9–12 September 2019; p. 289. [Google Scholar]
  19. Yu, C.; Gao, C.; Wang, J.; Yu, G.; Shen, C.; Sang, N. BiSeNet V2: Bilateral network with guided aggregation for real-time semantic segmentation. Int. J. Comput. Vis. 2021, 129, 3051–3068. [Google Scholar] [CrossRef]
  20. Fan, M.; Lai, S.; Huang, J.; Wei, X.; Chai, Z.; Luo, J.; Wei, X. Rethinking biSeNet for real-time semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Virtual, 19–25 June 2021; pp. 9716–9725. [Google Scholar]
  21. Cheng, J.; Li, H.; Li, D.; Hua, S.; Victor, S.S. A survey on image semantic segmentation using deep learning techniques. Comput. Mater. Contin. 2023, 74, 1941–1957. [Google Scholar] [CrossRef]
  22. Zhao, H.; Shi, J.; Qi, X. Pyramid scene parsing network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 6230–6239. [Google Scholar]
  23. Cheng, L.C.; Papandreou, G.; Schroff, F.; Adam, H. Rethinking atrous convolution for semantic image segmentation. arXiv, 2017; arXiv:1706.05587. [Google Scholar]
  24. Li, Y.; Li, M.; Li, Z.; Xiao, C.; Li, H. Efrnet: Efficient feature reuse network for real-time semantic segmentation. Neural Process. Lett. 2022, 54, 4647–4659. [Google Scholar] [CrossRef]
  25. Hu, X.; Jing, L.; Sehar, U. Joint pyramid attention network for real-time semantic segmentation of urban scenes. Appl. Intell. 2022, 52, 580–594. [Google Scholar] [CrossRef]
  26. Ma, N.; Zhang, X.; Zheng, H.; Sun, J. ShuffleNet V2: Practical guidelines for efficient CNN architecture design. In Proceedings of the Computer Vision 15th European Conference, Munich, Germany, 8–14 September 2018; pp. 122–138. [Google Scholar]
  27. Sandler, M.; Howard, A.G.; Zhu, M.; Zhmoginov, A.; Chen, L. MobileNetV2: Inverted residuals and linear bottlenecks. In Proceedings of the 2018 IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 4510–4520. [Google Scholar]
  28. Cordts, M.; Omran, M.; Ramos, S.; Rehfeld, T.; Enzweiler, M.; Benenson, R.; Franke, U.; Roth, S.; Schiele, B. The cityscapes dataset for semantic urban scene understanding. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 27–30 June 2016; pp. 3213–3223. [Google Scholar]
  29. Brostow, G.J.; Fauqueur, J.; Cipolla, R. Semantic object classes in video: A high-definition ground truth database. Pattern Recognit. Lett. 2009, 30, 88–97. [Google Scholar] [CrossRef]
  30. Gao, R. Rethink Dilated Convolution for Real-time Semantic Segmentation. arXiv, 2021; arXiv:2111.09957. [Google Scholar]
  31. Shrivastava, A.; Gupta, A.; Girshick, R. Training region-based object detectors with online hard example mining. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 27–30 June 2016; pp. 761–769. [Google Scholar]
  32. Elhassan, M.A.M.; Huang, C.; Yang, C.; Munea, T.L. DSANet: Dilated spatial attention for real-time semantic segmentation in urban street scenes. Expert Syst. Appl. 2021, 183, 115090. [Google Scholar] [CrossRef]
  33. Liu, C.; Gao, H.; Chen, A. A real-time semantic segmentation algorithm based on improved lightweight network. In Proceedings of the 2020 International Symposium on Autonomous Systems (ISAS), Guangzhou, China, 6–8 December 2020. [Google Scholar]
  34. Xiong, J.; Po, L.; Yu, W.; Zhou, C.; Xian, P.; Qu, W. CSRNet: Cascaded selective resolution network for real-time semantic segmentation. Expert Syst. Appl. 2023, 211, 118537. [Google Scholar] [CrossRef]
  35. Orsic, M.; Kreso, I.; Bevandic, P.; Segvic, S. In defense of pre-trained imagenet architectures for real-time semantic segmentation of road-driving images. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 16–20 June 2019; pp. 12607–12616. [Google Scholar]
  36. Mehta, S.; Rastegari, M.; Shapiro, L.G.; Hajishirzi, H. ESPNetv2: A light-weight, power efficient, and general purpose convolutional neural network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 16–20 June 2019; pp. 9190–9200. [Google Scholar]
  37. Zhao, H.; Qi, X.; Shen, X.; Shi, J.; Jia, J. Icnet for real-time semantic segmentation on high-resolution images. In Proceedings of the Computer Vision—ECCV 2018—15th European Conference, Munich, Germany, 8–14 September 2018; pp. 418–434. [Google Scholar]
  38. Zhou, Q.; Wang, Y.; Fan, Y.; Wu, X.; Zhang, S.; Kang, B.; Latecki, L. AGLNet: Towards real-time semantic segmentation of self-driving images via attention-guided lightweight network. Appl. Soft Comput. 2020, 96, 106682. [Google Scholar] [CrossRef]
  39. Gao, G.; Xu, G.; Yu, Y.; Xie, J.; Yang, J.; Yue, D. MSCFNet: A lightweight network with multi-scale context fusion for real-time semantic segmentation. IEEE Trans. Intell. Transp. Syst. 2022, 23, 25489–25499. [Google Scholar] [CrossRef]
  40. Lv, Q.; Sun, X.; Dong, J.; Zhou, H. Parallel complement network for real-Time semantic segmentation of road scenes. IEEE Trans. Intell. Transp. Syst. 2022, 23, 4432–4444. [Google Scholar] [CrossRef]
  41. Wu, T.; Tang, S.; Zhang, R.; Cao, J.; Zhang, Y. CGNet: A light-weight context guided network for semantic segmentation. IEEE Trans. Image Process. 2021, 30, 1169–1179. [Google Scholar] [CrossRef]
Figure 1. Performance comparison on the Cityscapes test set. (a) Speed–accuracy, (b) model size–accuracy. Red dot indicates our method, while blue dots mean other methods. The red line represents the real-time speed. “*” indicates that the corresponding method is pre-trained on other datasets. Our method achieves the most advanced trade-off between inference speed, segmentation accuracy, and model size.
Figure 1. Performance comparison on the Cityscapes test set. (a) Speed–accuracy, (b) model size–accuracy. Red dot indicates our method, while blue dots mean other methods. The red line represents the real-time speed. “*” indicates that the corresponding method is pre-trained on other datasets. Our method achieves the most advanced trade-off between inference speed, segmentation accuracy, and model size.
Sensors 23 06382 g001
Figure 2. Overview architecture of the proposed MFAFNet.
Figure 2. Overview architecture of the proposed MFAFNet.
Sensors 23 06382 g002
Figure 3. The structure of the encoder. The three digits besides the SAR-nbt indicate the stride, the channel expansion factor, and the dilatation rate of DDW, respectively. The detailed information of the SAR-nbt is shown in Figure 4.
Figure 3. The structure of the encoder. The three digits besides the SAR-nbt indicate the stride, the channel expansion factor, and the dilatation rate of DDW, respectively. The detailed information of the SAR-nbt is shown in Figure 4.
Sensors 23 06382 g003
Figure 6. Visualization results of the FAFM module on Cityscapes validation set.
Figure 6. Visualization results of the FAFM module on Cityscapes validation set.
Sensors 23 06382 g006
Figure 7. The validation loss curve and accuracy curve of MFAFNet during 500 training epochs on Cityscapes validation set.
Figure 7. The validation loss curve and accuracy curve of MFAFNet during 500 training epochs on Cityscapes validation set.
Sensors 23 06382 g007
Figure 8. Some visual comparisons on the Cityscapes validation set. From left to right are input image, ground truth, predicted results from BiseNet2, ERFNet, Fast-SCNN, ICNet, and our MFAFNet.
Figure 8. Some visual comparisons on the Cityscapes validation set. From left to right are input image, ground truth, predicted results from BiseNet2, ERFNet, Fast-SCNN, ICNet, and our MFAFNet.
Sensors 23 06382 g008
Table 1. Comparison of parameters and receptive field between common factorized convolutions and standard convolution.
Table 1. Comparison of parameters and receptive field between common factorized convolutions and standard convolution.
Convolution TypeParametersReceptive Field
Standard convolution k 2 C i n C o u t k × k
Group convolution k 2 C i n C o u t / g k × k
Asymmetric convolution 2 k C i n C o u t k × k
Depth-wise separable convolution k 2 C i n + C i n C o u t k × k
Dilated convolution k 2 C i n C o u t ( ( k 1 ) d + 1 ) 2
Dilated depth-wise separable convolution k 2 C i n + C i n C o u t ( ( k 1 ) d + 1 ) 2
Asymmetric depth-wise separable convolution 2 k C i n + C i n C o u t k × k
Asymmetric dilated depth-wise separable convolution 2 k C i n + C i n C o u t ( ( k 1 ) d + 1 ) 2
k” denotes the size of the convolution kernel. “ C i n ” and “ C o u t ” indicate the number of input and output channels, respectively. “g” represents the number of groups.
Table 2. Ablation study of SAR-nbt module on Cityscapes validation set.
Table 2. Ablation study of SAR-nbt module on Cityscapes validation set.
Ablation StudyTypeParamsMIoU (%)
Attention-0.9975.57
SE Weight1.2777.08
Dilation rates2, 4, 6, 8, 10, 12, 12, 12, 121.2777.08
2, 6, 6, 6, 6, 12, 12, 12, 121.2776.67
2, 4, 6, 8, 10, 12, 14, 16, 181.2776.82
2, 4, 4, 4, 4, 14, 14, 14, 141.2776.36
2, 6, 6, 12, 12, 18, 18, 24, 241.2775.16
Table 3. Ablation study of FAFM module on the Cityscapes validation set.
Table 3. Ablation study of FAFM module on the Cityscapes validation set.
TypeParams (M)FLOPs (G)MIoU (%)
x161.223.7771.62
x16 + x81.264.1074.95
x16 + x41.253.9875.28
FAFM1.274.1376.03
Table 4. Ablation study of SAR-nbt and FAFM components on the Cityscapes validation dataset.
Table 4. Ablation study of SAR-nbt and FAFM components on the Cityscapes validation dataset.
BaselineSAR-nbtFAFMParams (M)FLOPs (G)MIoU (%)
1.223.7766.25
1.223.7771.62
1.274.1371.43
1.274.1376.03
Table 5. Performance comparison with state-of-the-art methods on Cityscapes test set.
Table 5. Performance comparison with state-of-the-art methods on Cityscapes test set.
NetworkResolutionGPUParams (M)FPSMIoU (%)
ESPNet V2 [36] 512 × 1024 TitanX1.2567.066.2
ERFNet [16] 512 × 1024 TitanX2.141.768.0
ESNet [13] 512 × 1024 1080Ti1.6663.069.1
ICNet [37] 1024 × 2048 TitanX26.530.369.5
AGLNet [38] 512 × 1024 1080Ti1.1252.070.1
LRDNet [14] 512 × 1024 1080Ti0.6677.070.1
DFFNet [11] 512 × 1024 1080Ti1.962.571.0
DSANet [32] 512 × 1024 1080Ti3.4734.0871.4
JPANet [25] 512 × 1024 1080Ti3.49109.971.6
LEANet [17] 512 × 1024 1080Ti0.7477.371.9
MSCFNet [39] 512 × 1024 TitanX1.155071.9
MIFNet [12] 512 × 1024 1080Ti2.035.772.2
PCNet [40] 1024 × 2048 2080Ti1.579.172.9
LAANet [15] 512 × 1024 1080Ti0.6795.873.6
LightSeg [33] 512 × 1024 2080Ti8.649.573.8
CSRNet-light [34] 512 × 1024 1080Ti-56.074.0
BiSeNetV2-L [19] 512 × 1024 1080Ti-47.375.3
SwiftNet [35] 1024 × 2048 1080Ti11.841.075.5
EFRNet [24] 512 × 1024 2080Ti2.43118.075.6
FPANet C [10] 1024 × 1024 2080Ti15.4531.075.9
Ours 512 × 1024 1080Ti1.2760.175.9
“-” denotes that the model does not provide the corresponding result. “” indicates that the corresponding method is pre-trained on ImageNet or Scratch dataset. The computational efficiency of GPU is TitanX < 1080Ti < 2080Ti.
Table 6. Per-class IoU (%) performance on the Cityscapes test set.
Table 6. Per-class IoU (%) performance on the Cityscapes test set.
MethodMIoURoadSidewalkBuildingWallFencePoleTraffic LightTraffic SignVegetationTerrainSkyPersonRiderCarTruckBusTrainMotorcycleBicycle
ESPNet V2 [36]66.297.378.688.843.542.149.352.660.090.566.893.372.953.191.853.065.953.244.259.9
FastSCNN [18]68.097.981.689.746.448.648.353.060.090.767.294.374.054.693.057.465.558.250.061.2
ESNet [13]69.197.178.590.446.548.160.160.470.991.159.993.274.351.892.361.072.351.043.370.2
LRDNet [14]70.197.981.790.443.049.158.963.766.892.169.594.677.658.493.353.262.559.354.565.9
JPANet [25]71.698.182.790.647.148.253.160.266.191.669.294.578.660.594.064.275.966.563.765.9
LEANet [17]71.998.182.791.051.053.258.865.970.392.570.594.381.659.994.152.368.257.255.569.8
LAANet [15]73.697.982.991.047.551.559.366.070.392.369.994.781.861.494.258.679.575.154.369.4
Ours75.998.484.792.353.657.660.969.373.093.070.895.283.365.895.166.278.971.561.072.0
Bold values indicate that the corresponding indicator is superior to other methods.
Table 7. Comparison with other methods on the CamVid test set.
Table 7. Comparison with other methods on the CamVid test set.
ModelParams (M)MIoU (%)
ERFNet [16]2.0663.7
DFFNet [11]1.964.7
CGNet [41]0.565.5
MIFNet [12]2.065.7
SwiftNet [35]12.965.7
EFRNet [24]2.4366.2
PCNet [40]1.667.0
JPANet-G [25]3.4967.5
LEANet [17]0.7467.5
AGLNet [38]1.1269.4
Ours1.2769.9
Bold values indicate that the corresponding indicator is superior to other methods.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lu, K.; Cheng, J.; Li, H.; Ouyang, T. MFAFNet: A Lightweight and Efficient Network with Multi-Level Feature Adaptive Fusion for Real-Time Semantic Segmentation. Sensors 2023, 23, 6382. https://doi.org/10.3390/s23146382

AMA Style

Lu K, Cheng J, Li H, Ouyang T. MFAFNet: A Lightweight and Efficient Network with Multi-Level Feature Adaptive Fusion for Real-Time Semantic Segmentation. Sensors. 2023; 23(14):6382. https://doi.org/10.3390/s23146382

Chicago/Turabian Style

Lu, Kai, Jieren Cheng, Hua Li, and Tianyu Ouyang. 2023. "MFAFNet: A Lightweight and Efficient Network with Multi-Level Feature Adaptive Fusion for Real-Time Semantic Segmentation" Sensors 23, no. 14: 6382. https://doi.org/10.3390/s23146382

APA Style

Lu, K., Cheng, J., Li, H., & Ouyang, T. (2023). MFAFNet: A Lightweight and Efficient Network with Multi-Level Feature Adaptive Fusion for Real-Time Semantic Segmentation. Sensors, 23(14), 6382. https://doi.org/10.3390/s23146382

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop