LCMorph: Exploiting Frequency Cues and Morphological Perception for Low-Contrast Road Extraction in Remote Sensing Images
Abstract
:1. Introduction
- (1)
- Expanding the receptive field and capturing long-range context. D-LinkNet [20] enhances LinkNet by integrating dilated convolution operations and skip connections, which expand the receptive field while preserving detailed information. SIINet [21] adopts a spatial information inference structure to learn both local visual features of roads and global information, alleviating the occlusion problem. To capture long-range context, NL-LinkNet [22] incorporates nonlocal operations into LinkNet’s encoder. Moreover, the Transformer architecture excels at modeling long-range dependencies, leading to the development of Transformer-based road extraction methods. Luo et al. [23] proposed BDTNet, which employs a Bi-direction Transformer Module (BDTM) to capture contextual information of roads. To extract roads precisely, UMiT-Net [24] was developed. It consists of four mix-Transformer blocks for global feature extraction and a Dilated Attention Module (DAM) for semantic feature fusion. Inspired by the sparse target pixels in remote sensing images, Chen et al. [25] proposed the Sparse Token Transformer (STT) to learn sparse feature representations. STT not only reduces computational complexity but also enhances extraction accuracy.
- (2)
- Emphasizing the geometric attributes of roads. Roads exhibit distinctive geometric attributes, such as direction, connectivity, shape, and topology. Ding et al. [26] proposed the Direction-Aware Residual Network (DiResNet), incorporating direction supervision during training. Besides road extraction, CoANet [27] employs a connectivity attention module to explore the relationship between neighboring pixels. Consequently, road connectivity is well preserved. RSANet [28] is proposed to address the challenges of complex road shapes. It uses the Efficient Strip Transformer Module (ESTM) to model the long-distance dependencies required by long roads. And Road Edge Focal loss (REF loss) is introduced to alleviate sample imbalance caused by thin roads. Considering the topology of road networks, SDUNet [29] was designed to learn multi-level features and global prior information of road networks. From the perspective of constraints on model learning, Hu et al. [30] proposed PolyRoad. It uses a polyline matching cost and additional losses for improved road topology.
- (3)
- Reducing the labeled data needed for training. Constructing large-scale labeled datasets is both costly and time-consuming. To address this issue, a semi-supervised network, SemiRoadExNet [31], was proposed to leverage pseudo-label information. Road extraction methods based on unsupervised learning do not rely on labeled datasets. To tackle the domain shift challenge, Zhang et al. [32] designed RoadDA, a two-stage unsupervised domain adaptation network for road extraction. Besides these standard models, researchers have explored large models for road extraction. For example, Chen et al. [33] proposed RSPrompter to learn the generation of appropriate prompts for the Segment Anything Model (SAM) [34]. RSPrompter enables the SAM to produce semantically discernible segmentation results for remote sensing images. Moreover, Hetang et al. [35] improved the SAM by designing SAM-Road to extract road networks.
2. Materials and Methods
2.1. Datasets
2.1.1. DeepGlobe Dataset
2.1.2. Massachusetts Roads Dataset
2.1.3. LC-Roads Dataset
2.2. Proposed Method
2.2.1. Frequency-Enhanced Localization
2.2.2. Morphology-Enhanced Extraction
2.2.3. Loss Function
3. Results
3.1. Implementation Details and Evaluation Metrics
3.2. Comparison Experiments on LC-Roads Dataset
- (1)
- UNet [16]. U-Net is a widely used semantic segmentation model featuring a symmetric encoder–decoder architecture with skip connections, enabling precise segmentation and context capture.
- (2)
- SegNet [17]. SegNet utilizes an encoder–decoder architecture, where the decoder employs pooling indices from the encoder for up-sampling, which preserves spatial details.
- (3)
- LinkNet [18]. LinkNet is an efficient semantic segmentation model designed for real-time applications. It combines an encoder–decoder architecture with residual connections to maintain high accuracy with fewer parameters.
- (4)
- DeepLabV3+ [19]. DeepLabV3+ employs an encoder–decoder architecture with atrous convolution. The encoder captures multi-scale contextual information, and the decoder is simple yet effective.
- (5)
- PSPNet [48]. To capture global contextual information, PSPNet introduces pyramid pooling modules, enhancing the model’s ability to understand various object scales.
- (6)
- D-LinkNet [20]. D-LinkNet is a classical road extraction model. Based on LinkNet, D-LinkNet contains dilated convolution layers to expand the receptive field. It won first place in the CVPR DeepGlobe 2018 Road Extraction Challenge.
- (7)
- SIINet [21]. SIINet enhances road extraction by facilitating multidirectional message passing between pixels. It effectively captures both local and global spatial information.
- (8)
- CoANet [27]. CoANet is a road extraction model which integrates strip convolution operations with a connectivity attention module. It addresses occlusions and achieves good results.
- (9)
- NL-LinkNet [22]. NL-LinkNet is the first road extraction model to use nonlocal operations. The nonlocal block enables the model to capture long-range dependencies and distant information.
- (10)
- SDUNet [29]. SDUNet is a spatially enhanced and densely connected UNet. It aggregates multi-level features and preserves road structure.
- (11)
- DSCNet [39]. DSCNet is a tubular structure segmentation model applied to both vessel segmentation and road extraction. Snake convolution is proposed by DSCNet.
- (12)
- RoadExNet. RoadExNet is the generator of SemiRoadExNet [31]. For fair comparison, we trained RoadExNet in a fully supervised manner.
- (13)
- OARENet [49]. OARENet is a road extraction model designed to address dense occlusions. It proposes an occlusion-aware decoder, achieving excellent performance in complex scenes.
3.3. Comparison Experiments on Public Datasets
4. Discussion
4.1. Effect of Different Encoders
4.2. Effectiveness of Each Module in LCMorph
4.3. Effect of Different Frequency Components
4.4. Computational Efficiency
5. Conclusions
- The Frequency-Aware Module (FAM) is introduced to enhance the distinction between low-contrast roads and the background. With its help, LCMorph effectively identifies overlooked low-contrast roads.
- To handle elongated and curved roads, we propose the Morphological Perception Blocks (MPBlocks). These blocks adaptively adjust the receptive field to the road morphology, achieving accurate road extraction.
- LCMorph achieves state-of-the-art performance in terms of F1 score and IoU on the LC-Roads, DeepGlobe, and Massachusetts Roads datasets. And the effectiveness of the FAM and MPBlock is validated through adequate ablation experiments.
Author Contributions
Funding
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- Son, T.H.; Weedon, Z.; Yigitcanlar, T.; Sanchez, T.; Corchado, J.M.; Mehmood, R. Algorithmic urban planning for smart and sustainable development: Systematic review of the literature. Sustain. Cities Soc. 2023, 94, 104562. [Google Scholar] [CrossRef]
- Żochowska, R.; Pamuła, T. Impact of Traffic Flow Rate on the Accuracy of Short-Term Prediction of Origin-Destination Matrix in Urban Transportation Networks. Remote Sens. 2024, 16, 1202. [Google Scholar] [CrossRef]
- Chen, P.; Wu, J.; Li, N. A personalized navigation route recommendation strategy based on differential perceptron tracking user’s driving preference. Comput. Intell. Neurosci. 2023, 2023, 8978398. [Google Scholar] [CrossRef] [PubMed]
- Stewart, C.; Lazzarini, M.; Luna, A.; Albani, S. Deep learning with open data for desert road mapping. Remote Sens. 2020, 12, 2274. [Google Scholar] [CrossRef]
- Lian, R.; Wang, W.; Mustafa, N.; Huang, L. Road extraction methods in high-resolution remote sensing images: A comprehensive review. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 13, 5489–5507. [Google Scholar] [CrossRef]
- Stoica, R.; Descombes, X.; Zerubia, J. A Gibbs point process for road extraction from remotely sensed images. Int. J. Comput. Vis. 2004, 57, 121–136. [Google Scholar] [CrossRef]
- Sghaier, M.O.; Lepage, R. Road extraction from very high resolution remote sensing optical images based on texture analysis and beamlet transform. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2015, 9, 1946–1958. [Google Scholar] [CrossRef]
- Mohammadzadeh, A.; Tavakoli, A.; Valadan Zoej, M.J. Road extraction based on fuzzy logic and mathematical morphology from pan-sharpened IKONOS images. Photogramm. Rec. 2006, 21, 44–60. [Google Scholar] [CrossRef]
- Maurya, R.; Gupta, P.; Shukla, A.S. Road extraction using k-means clustering and morphological operations. In Proceedings of the 2011 International Conference on Image Information Processing, Shimla, India, 3–5 November 2011; pp. 1–6. [Google Scholar]
- Yager, N.; Sowmya, A. Support vector machines for road extraction from remotely sensed images. In Proceedings of the International Conference on Computer Analysis of Images and Patterns, Groningen, The Netherlands, 25–27 August 2003; pp. 285–292. [Google Scholar]
- Gamba, P.; Dell’Acqua, F.; Lisini, G. Improving urban road extraction in high-resolution images exploiting directional filtering, perceptual grouping, and simple topological concepts. IEEE Geosci. Remote Sens. Lett. 2006, 3, 387–391. [Google Scholar] [CrossRef]
- Shi, W.; Miao, Z.; Debayle, J. An integrated method for urban main-road centerline extraction from optical remotely sensed imagery. IEEE Trans. Geosci. Remote Sens. 2013, 52, 3359–3372. [Google Scholar] [CrossRef]
- Nie, J.; Wang, Z.; Liang, X.; Yang, C.; Zheng, C.; Wei, Z. Semantic Category Balance-Aware Involved Anti-Interference Network for Remote Sensing Semantic Segmentation. IEEE Trans. Geosci. Remote Sens. 2023, 61, 4409712. [Google Scholar] [CrossRef]
- Li, X.; Wang, Z.; Chen, C.; Tao, C.; Qiu, Y.; Liu, J.; Sun, B. SemID: Blind Image Inpainting with Semantic Inconsistency Detection. Tsinghua Sci. Technol. 2024, 29, 1053–1068. [Google Scholar] [CrossRef]
- Liu, R.; Wu, J.; Lu, W.; Miao, Q.; Zhang, H.; Liu, X.; Lu, Z.; Li, L. A Review of Deep Learning-Based Methods for Road Extraction from High-Resolution Remote Sensing Images. Remote Sens. 2024, 16, 2056. [Google Scholar] [CrossRef]
- Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention—MICCAI 2015: 18th International Conference, Munich, Germany, 5–9 October 2015; Proceedings, Part III 18. pp. 234–241. [Google Scholar]
- Badrinarayanan, V.; Kendall, A.; Cipolla, R. Segnet: A deep convolutional encoder-decoder architecture for image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 2481–2495. [Google Scholar] [CrossRef] [PubMed]
- Chaurasia, A.; Culurciello, E. Linknet: Exploiting encoder representations for efficient semantic segmentation. In Proceedings of the 2017 IEEE Visual Communications and Image Processing (VCIP), St. Petersburg, FL, USA, 10–13 December 2017; pp. 1–4. [Google Scholar]
- Chen, L.C.; Zhu, Y.; Papandreou, G.; Schroff, F.; Adam, H. Encoder-decoder with atrous separable convolution for semantic image segmentation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 801–818. [Google Scholar]
- Zhou, L.; Zhang, C.; Wu, M. D-LinkNet: LinkNet with pretrained encoder and dilated convolution for high resolution satellite imagery road extraction. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Salt Lake City, UT, USA, 18–23 June 2018; pp. 182–186. [Google Scholar]
- Tao, C.; Qi, J.; Li, Y.; Wang, H.; Li, H. Spatial information inference net: Road extraction using road-specific contextual information. ISPRS J. Photogramm. Remote Sens. 2019, 158, 155–166. [Google Scholar] [CrossRef]
- Wang, Y.; Seo, J.; Jeon, T. NL-LinkNet: Toward lighter but more accurate road extraction with nonlocal operations. IEEE Geosci. Remote Sens. Lett. 2022, 19, 3000105. [Google Scholar] [CrossRef]
- Luo, L.; Wang, J.X.; Chen, S.B.; Tang, J.; Luo, B. BDTNet: Road extraction by bi-direction transformer from remote sensing images. IEEE Geosci. Remote Sens. Lett. 2022, 19, 2505605. [Google Scholar] [CrossRef]
- Deng, F.; Luo, W.; Ni, Y.; Wang, X.; Wang, Y.; Zhang, G. UMiT-Net: A U-shaped mix-transformer network for extracting precise roads using remote sensing images. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5801513. [Google Scholar] [CrossRef]
- Chen, K.; Zou, Z.; Shi, Z. Building extraction from remote sensing images with sparse token transformers. Remote Sens. 2021, 13, 4441. [Google Scholar] [CrossRef]
- Ding, L.; Bruzzone, L. DiResNet: Direction-aware residual network for road extraction in VHR remote sensing images. IEEE Trans. Geosci. Remote Sens. 2020, 59, 10243–10254. [Google Scholar] [CrossRef]
- Mei, J.; Li, R.J.; Gao, W.; Cheng, M.M. CoANet: Connectivity attention network for road extraction from satellite imagery. IEEE Trans. Image Process. 2021, 30, 8540–8552. [Google Scholar] [CrossRef]
- Wang, C.; Xu, R.; Xu, S.; Meng, W.; Wang, R.; Zhang, J.; Zhang, X. Toward accurate and efficient road extraction by leveraging the characteristics of road shapes. IEEE Trans. Geosci. Remote Sens. 2023, 61, 4404616. [Google Scholar] [CrossRef]
- Yang, M.; Yuan, Y.; Liu, G. SDUNet: Road extraction via spatial enhanced and densely connected UNet. Pattern Recognit. 2022, 126, 108549. [Google Scholar] [CrossRef]
- Hu, Y.; Wang, Z.; Huang, Z.; Liu, Y. PolyRoad: Polyline Transformer for Topological Road-Boundary Detection. IEEE Trans. Geosci. Remote Sens. 2023, 62, 5602112. [Google Scholar] [CrossRef]
- Chen, H.; Li, Z.; Wu, J.; Xiong, W.; Du, C. SemiRoadExNet: A semi-supervised network for road extraction from remote sensing imagery via adversarial learning. ISPRS J. Photogramm. Remote Sens. 2023, 198, 169–183. [Google Scholar] [CrossRef]
- Zhang, L.; Lan, M.; Zhang, J.; Tao, D. Stagewise unsupervised domain adaptation with adversarial self-training for road segmentation of remote-sensing images. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5609413. [Google Scholar] [CrossRef]
- Chen, K.; Liu, C.; Chen, H.; Zhang, H.; Li, W.; Zou, Z.; Shi, Z. RSPrompter: Learning to prompt for remote sensing instance segmentation based on visual foundation model. IEEE Trans. Geosci. Remote Sens. 2024, 62, 4701117. [Google Scholar] [CrossRef]
- Kirillov, A.; Mintun, E.; Ravi, N.; Mao, H.; Rolland, C.; Gustafson, L.; Xiao, T.; Whitehead, S.; Berg, A.C.; Lo, W.Y.; et al. Segment anything. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 1–6 October 2023; pp. 4015–4026. [Google Scholar]
- Hetang, C.; Xue, H.; Le, C.; Yue, T.; Wang, W.; He, Y. Segment Anything Model for Road Network Graph Extraction. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 16–22 June 2024; pp. 2556–2566. [Google Scholar]
- Zhong, Y.; Li, B.; Tang, L.; Kuang, S.; Wu, S.; Ding, S. Detecting camouflaged object in frequency domain. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 4504–4513. [Google Scholar]
- Cong, R.; Sun, M.; Zhang, S.; Zhou, X.; Zhang, W.; Zhao, Y. Frequency perception network for camouflaged object detection. In Proceedings of the 31st ACM International Conference on Multimedia, Ottawa, ON, Canada, 29 October–3 November 2023; pp. 1179–1189. [Google Scholar]
- Xie, C.; Xia, C.; Yu, T.; Li, J. Frequency representation integration for camouflaged object detection. In Proceedings of the 31st ACM International Conference on Multimedia, Ottawa, ON, Canada, 29 October–3 November 2023; pp. 1789–1797. [Google Scholar]
- Qi, Y.; He, Y.; Qi, X.; Zhang, Y.; Yang, G. Dynamic snake convolution based on topological geometric constraints for tubular structure segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 1–6 October 2023; pp. 6070–6079. [Google Scholar]
- Demir, I.; Koperski, K.; Lindenbaum, D.; Pang, G.; Huang, J.; Basu, S.; Hughes, F.; Tuia, D.; Raskar, R. Deepglobe 2018: A challenge to parse the earth through satellite images. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Salt Lake City, UT, USA, 18–23 June 2018; pp. 172–181. [Google Scholar]
- Mnih, V. Machine Learning for Aerial Image Labeling; University of Toronto: Toronto, ON, Canada, 2013. [Google Scholar]
- He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
- Winograd, S. On computing the discrete Fourier transform. Math. Comput. 1978, 32, 175–199. [Google Scholar] [CrossRef]
- Ahmed, N.; Natarajan, T.; Rao, K.R. Discrete cosine transform. IEEE Trans. Comput. 1974, 100, 90–93. [Google Scholar] [CrossRef]
- Shensa, M.J. The discrete wavelet transform: Wedding the a trous and Mallat algorithms. IEEE Trans. Signal Process. 1992, 40, 2464–2482. [Google Scholar] [CrossRef]
- Chen, Y.; Fan, H.; Xu, B.; Yan, Z.; Kalantidis, Y.; Rohrbach, M.; Yan, S.; Feng, J. Drop an octave: Reducing spatial redundancy in convolutional neural networks with octave convolution. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 3435–3444. [Google Scholar]
- Tong, Z.; Li, Y.; Zhang, J.; He, L.; Gong, Y. MSFANet: Multiscale fusion attention network for road segmentation of multispectral remote sensing data. Remote Sens. 2023, 15, 1978. [Google Scholar] [CrossRef]
- Zhao, H.; Shi, J.; Qi, X.; Wang, X.; Jia, J. Pyramid scene parsing network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2881–2890. [Google Scholar]
- Yang, R.; Zhong, Y.; Liu, Y.; Lu, X.; Zhang, L. Occlusion-aware road extraction network for high-resolution remote sensing imagery. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5619316. [Google Scholar] [CrossRef]
- Liu, S.; Huang, D. Receptive field block net for accurate and fast object detection. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 385–400. [Google Scholar]
Method | Source | P (%) | R (%) | F1 (%) | IoU (%) |
---|---|---|---|---|---|
Semantic Segmentation Models | |||||
UNet | MICCAI’15 | 70.89 | 73.90 | 72.36 | 56.70 |
PSPNet | CVPR’17 | 62.96 | 77.60 | 69.52 | 53.27 |
LinkNet | VCIP’17 | 71.16 | 75.71 | 73.37 | 57.94 |
SegNet | TPAMI’17 | 72.32 | 76.19 | 74.20 | 58.99 |
DeepLabV3+ | ECCV’18 | 66.63 | 77.87 | 71.81 | 56.02 |
Road Extraction Models | |||||
D-LinkNet | CVPRW’18 | 71.87 | 79.65 | 75.56 | 60.72 |
SIINet | ISPRS’19 | 71.16 | 72.72 | 71.93 | 56.23 |
CoANet | TIP’21 | 72.76 | 78.99 | 75.74 | 60.95 |
NL-LinkNet | GRSL’22 | 71.48 | 75.63 | 73.50 | 58.10 |
SDUNet | PR’22 | 72.03 | 73.04 | 72.53 | 56.95 |
DSCNet | ICCV’23 | 71.09 | 75.65 | 73.30 | 57.91 |
RoadExNet | ISPRS’23 | 72.82 | 75.84 | 74.33 | 59.14 |
OARENet | TGRS’24 | 72.14 | 72.51 | 72.21 | 56.51 |
LCMorph | Ours | 72.32 | 81.99 | 76.85 | 62.40 |
Method | Source | DeepGlobe Dataset | Massachusetts Roads Dataset | ||||||
---|---|---|---|---|---|---|---|---|---|
P (%) | R (%) | F1 (%) | IoU (%) | P (%) | R (%) | F1 (%) | IoU (%) | ||
Semantic Segmentation Models | |||||||||
UNet | MICCAI’15 | 73.20 | 70.10 | 71.62 | 59.06 | 77.29 | 72.13 | 73.19 | 59.46 |
PSPNet | CVPR’17 | 75.35 | 78.71 | 76.99 | 62.95 | 76.53 | 69.47 | 72.83 | 57.27 |
LinkNet | VCIP’17 | 71.33 | 79.81 | 75.33 | 61.34 | 79.18 | 73.71 | 75.52 | 61.74 |
SegNet | TPAMI’17 | 79.84 | 75.92 | 77.83 | 63.79 | 72.79 | 77.41 | 74.26 | 60.11 |
DeepLabV3+ | ECCV’18 | 78.20 | 76.24 | 75.69 | 62.33 | 75.47 | 77.97 | 76.70 | 62.25 |
Road Extraction Models | |||||||||
D-LinkNet | CVPRW’18 | 73.50 | 81.38 | 77.24 | 63.36 | 74.57 | 78.85 | 75.58 | 61.75 |
SIINet | ISPRS’19 | 75.42 | 83.15 | 79.09 | 64.35 | 73.47 | 70.89 | 72.16 | 56.85 |
CoANet | TIP’21 | 74.02 | 85.31 | 79.27 | 65.65 | 75.15 | 77.89 | 76.48 | 61.94 |
NL-LinkNet | GRSL’22 | 74.99 | 77.50 | 76.23 | 62.65 | 79.14 | 74.17 | 76.57 | 62.19 |
SDUNet | PR’22 | 78.40 | 80.43 | 79.40 | 65.91 | 77.56 | 74.57 | 75.23 | 61.34 |
DSCNet | ICCV’23 | 77.03 | 75.91 | 76.47 | 62.76 | 75.83 | 77.47 | 76.64 | 62.22 |
RoadExNet | ISPRS’23 | 77.76 | 77.14 | 77.45 | 63.51 | 82.46 | 72.89 | 77.38 | 63.10 |
OARENet | TGRS’24 | 79.88 | 76.70 | 78.26 | 64.04 | 77.79 | 75.23 | 76.49 | 61.96 |
LCMorph | Ours | 78.51 | 81.02 | 79.74 | 66.30 | 76.93 | 78.51 | 77.71 | 63.55 |
Method | Encoder | P (%) | R (%) | F1 (%) | IoU (%) | Params (M) | FLOPs (G) |
---|---|---|---|---|---|---|---|
LCMorph-light | ResNet50 | 72.64 | 77.26 | 74.88 | 59.85 | 52.90 | 216.55 |
LCMorph | ResNet101 | 72.32 | 81.99 | 76.85 | 62.40 | 71.90 | 294.75 |
LCMorph-heavy | ResNet152 | 73.85 | 80.27 | 76.93 | 62.73 | 87.54 | 358.62 |
No. | FAM | MPBlock | P (%) | R (%) | F1 (%) | IoU (%) |
---|---|---|---|---|---|---|
1 | 69.82 | 81.66 | 75.28 | 60.35 | ||
2 | ✓ | 69.77 | 82.36 | 75.54 | 60.70 | |
3 | ✓ | 72.18 | 81.31 | 76.47 | 61.90 | |
4 | ✓ | ✓ | 72.32 | 81.99 | 76.85 | 62.40 |
Approach | P (%) | R (%) | F1 (%) | IoU (%) |
---|---|---|---|---|
Low frequency | 68.15 | 72.02 | 70.03 | 53.81 |
High frequency | 78.28 | 72.18 | 75.11 | 60.14 |
Low and high frequency | 76.95 | 76.25 | 76.85 | 62.40 |
Method | Source | Params (M) | FLOPs (G) | (%) | (%) | (%) |
---|---|---|---|---|---|---|
Semantic Segmentation Models | ||||||
UNet | MICCAI’15 | 26.36 | 223.88 | 59.06 | 59.46 | 56.70 |
PSPNet | CVPR’17 | 86.06 | 327.02 | 62.95 | 57.27 | 53.27 |
LinkNet | VCIP’17 | 11.53 | 12.09 | 61.34 | 61.74 | 57.94 |
SegNet | TPAMI’17 | 29.48 | 170.45 | 63.79 | 60.11 | 58.99 |
DeepLabV3+ | ECCV’18 | 54.70 | 83.24 | 62.33 | 62.25 | 56.02 |
Road Extraction Models | ||||||
D-LinkNet | CVPRW’18 | 31.10 | 33.60 | 63.36 | 61.75 | 60.72 |
SIINet | ISPRS’19 | 7.36 | 36.10 | 64.35 | 56.85 | 56.23 |
CoANet | TIP’21 | 59.15 | 277.58 | 65.65 | 61.94 | 60.95 |
NL-LinkNet | GRSL’22 | 21.82 | 32.07 | 62.65 | 62.19 | 58.10 |
SDUNet | PR’22 | 80.24 | 353.26 | 65.91 | 61.34 | 56.95 |
DSCNet | ICCV’23 | 4.52 | 40.38 | 62.76 | 62.22 | 57.91 |
RoadExNet | ISPRS’23 | 31.13 | 33.84 | 63.51 | 63.10 | 59.14 |
OARENet | TGRS’24 | 71.30 | 99.90 | 64.04 | 61.96 | 56.51 |
LCMorph | Ours | 71.90 | 294.75 | 66.30 | 63.55 | 62.40 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Li, X.; Yang, S.; Meng, F.; Li, W.; Yang, Z.; Wei, R. LCMorph: Exploiting Frequency Cues and Morphological Perception for Low-Contrast Road Extraction in Remote Sensing Images. Remote Sens. 2025, 17, 257. https://doi.org/10.3390/rs17020257
Li X, Yang S, Meng F, Li W, Yang Z, Wei R. LCMorph: Exploiting Frequency Cues and Morphological Perception for Low-Contrast Road Extraction in Remote Sensing Images. Remote Sensing. 2025; 17(2):257. https://doi.org/10.3390/rs17020257
Chicago/Turabian StyleLi, Xin, Shumin Yang, Fan Meng, Wenlong Li, Zongchi Yang, and Ruoyu Wei. 2025. "LCMorph: Exploiting Frequency Cues and Morphological Perception for Low-Contrast Road Extraction in Remote Sensing Images" Remote Sensing 17, no. 2: 257. https://doi.org/10.3390/rs17020257
APA StyleLi, X., Yang, S., Meng, F., Li, W., Yang, Z., & Wei, R. (2025). LCMorph: Exploiting Frequency Cues and Morphological Perception for Low-Contrast Road Extraction in Remote Sensing Images. Remote Sensing, 17(2), 257. https://doi.org/10.3390/rs17020257