SwinLabNet: Jujube Orchard Drivable Area Segmentation Based on Lightweight CNN-Transformer Architecture
Abstract
:1. Introduction
- (1)
- A lightweight CNN-Transformer hybrid architecture, SwinLabNet, is proposed, which includes both encoding and decoding structures.
- (2)
- The SwinASPP feature extraction module is introduced to enhance the fine-grained segmentation of drivable areas in jujube belts by expanding the receptive field and capturing more contextual semantic information, effectively adapting to the complex orchard environment.
- (3)
- The improved model effectively extracts drivable areas in jujube belts and demonstrates good generalization performance on vegetable datasets.
2. Materials and Methods
2.1. Experimental Equipment and Parameter ConFigureuration
2.2. Semantic Segmentation of Drivable Areas Based on Neural Network
2.2.1. Dataset Acquisition
2.2.2. Dataset Augmentation
2.3. Construction of a Model for Identifying Drivable Areas between Rows in Unstructured Jujube Orchards
2.3.1. Lightweight Backbone Network MobileNetV3-ECA Module
2.3.2. SwinASPP
2.3.3. Loss Function Design
3. Experiment and Result Analysis
3.1. Evaluation Metrucs
3.2. Influence of Different Loss Functions on Experimental Results in an Improved Model
3.3. Comparison of Lightweight Backbone Network Performance
3.4. Ablation Study
3.5. Comparative Analysis of Different Model Performances
3.6. Analysis of Visualization Results
3.7. Model Generalization
4. Conclusions
- (1)
- First, MobileNetV3-ECA was used in the feature extraction stage, significantly reducing the model’s parameters. Second, the Swin Transformer was introduced to enhance the model’s ability to capture contextual semantic information, addressing the issue of weak correlations between long-distance features. Finally, a mixed loss function was employed to handle the class imbalance problem, enabling the efficient extraction of abundant semantic information with a simple training method and fewer parameters.
- (2)
- Regarding accuracy, the experimental results show that the improved model achieved an MIoU of 95.73%, a precision of 97.24%, and a recall of 98.36%. Compared to the original DeepLabV3+ network, these metrics improved by 5.22%, 3.62%, and 2.04%, respectively. When handling the jujube belt dataset, characterized by long and blurred boundaries, complex information, and discrete distribution, the proposed method demonstrated superior segmentation performance compared to other mainstream models. It also shows strong robustness and stability on vegetable datasets.
- (3)
- Regarding lightweight design, this model uses MobileNetV3-ECA as the backbone network, with the number of parameters reduced to less than one-tenth of the original model. This provides better adaptability for deployment on edge devices.
Author Contributions
Funding
Institutional Review Board Statement
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- Dou, H.; Chen, Z.; Zhai, C.; Zou, W.; Song, J.; Feng, F.; Zhang, Y. Research Progress on Autonomous Navigation Technology for Intelligent Orchard Operation Equipment. Trans. Chin. Soc. Agric. Mach. 2024, 55, 891–898. [Google Scholar]
- Meng, Z.; Wang, H.; Fu, W.; Liu, M.; Yin, Y.; Zhao, C. Research Status and Prospects of Agricultural Machinery Autonomous Driving. Trans. Chin. Soc. Agric. 2023, 54, 1–24. [Google Scholar]
- Han, L.; He, X.; Wang, C.; Liu, Y.; Song, J.; Qi, P.; Liu, L.; Li, T.; Zheng, Y.; Lin, G.; et al. Key Technologies and Equipment for Smart Orchard Construction and Prospects. Smart Agric. 2022, 4, 1–11. [Google Scholar]
- Zhou, H.; Wang, X.; Au, W.; Kang, H.; Chen, C. Intelligent robots for fruit harvesting: Recent developments and future challenges. Precis. Agric. 2022, 23, 1856–1907. [Google Scholar] [CrossRef]
- Xie, B.; Jin, Y.; Faheem, M.; Gao, W.; Liu, J.; Jiang, H.; Cai, L.; Li, Y. Research progress of autonomous navigation technology for multi-agricultural scenes. Comput. Electron. Agric. 2023, 211, 107963. [Google Scholar] [CrossRef]
- Gao, X.; Li, J.; Fan, L.; Zhou, Q.; Yin, K.; Wang, J.; Song, C.; Huang, L.; Wang, Z. Review of Wheeled Mobile Robots’ Navigation Problems and Application Prospects in Agriculture. IEEE Access 2018, 6, 49248–49268. [Google Scholar] [CrossRef]
- Bai, Y.; Zhang, B.; Xu, N.; Zhou, J.; Shi, J.; Diao, Z. Vision-based navigation and guidance for agricultural autonomous vehicles and robots: A review. Comput. Electron. Agric. 2023, 205, 107584. [Google Scholar] [CrossRef]
- Jin, Y.; Liu, J.; Xu, Z.; Yan, S.; Li, P.; Wang, J. Development status and trend of agricultural robot technology. Int. J. Agric. Biol. Eng. 2021, 14, 1–19. [Google Scholar] [CrossRef]
- Shi, J.; Bai, Y.; Diao, Z.; Zhou, J.; Yao, X.; Zhang, B. Row detection BASED navigation and guidance for agricultural robots and autonomous vehicles in row-crop fields: Methods and applications. Agronomy 2023, 13, 1780. [Google Scholar] [CrossRef]
- Pham, T. Semantic Road Segmentation using Deep Learning. In Proceedings of the 2020 Applying New Technology in Green Buildings (ATiGB), Da Nang, Vietnam, 12–13 March 2021; pp. 45–48. [Google Scholar]
- Wang, L.; Chen, X.; Hu, L.; Li, H. Overview of Image Semantic Segmentation Technology. In Proceedings of the 2020 IEEE 9th Joint International Information Technology and Artificial Intelligence Conference (ITAIC), Chongqing, China, 11–13 December 2020; pp. 19–26. [Google Scholar]
- Kheradmandi, N.; Mehranfar, V. A critical review and comparative study on image segmentation-based techniques for pavement crack detection. Constr. Build. Mater. 2022, 321, 126162. [Google Scholar] [CrossRef]
- Jing, J.; Liu, S.; Wang, G.; Zhang, W.; Sun, C. Recent advances on image edge detection: A comprehensive review. Neurocomputing 2022, 503, 259–271. [Google Scholar] [CrossRef]
- Barhate, D.; Nemade, V. Comprehensive Study on Automated Image Detection by Robotics for Agriculture Applications. In Proceedings of the 2019 3rd International conference on Electronics, Communication and Aerospace Technology (ICECA), Coimbatore, India, 12–14 June 2019; pp. 637–641. [Google Scholar]
- Kamilaris, A.; Prenafeta-Boldú, F.X. A review of the use of convolutional neural networks in agriculture. J. Agric. Sci. 2018, 156, 312–322. [Google Scholar] [CrossRef]
- Saleem, M.H.; Potgieter, J.; Arif, K.M. Automation in agriculture by machine and deep learning techniques: A review of recent developments. Precis. Agric. 2021, 22, 2053–2091. [Google Scholar] [CrossRef]
- Pally, R.J.; Samadi, S. Application of image processing and convolutional neural networks for flood image classification and semantic segmentation. Environ. Model. Softw. 2022, 148, 105285. [Google Scholar] [CrossRef]
- Moazzam, S.I.; Khan, U.S.; Tiwana, M.I.; Iqbal, J.; Qureshi, W.S.; Shah, S.I. A Review of Application of Deep Learning for Weeds and Crops Classification in Agriculture. In Proceedings of the 2019 International Conference on Robotics and Automation in Industry (ICRAI), Rawalpindi, Pakistan, 21–22 October 2019; pp. 1–6. [Google Scholar]
- Wang, J.-J.; Liu, Y.-F.; Nie, X.; Mo, Y.L. Deep convolutional neural networks for semantic segmentation of cracks. Struct. Control Health Monit. 2022, 29, e2850. [Google Scholar] [CrossRef]
- Turay, T.; Vladimirova, T. Toward performing image classification and object detection with convolutional neural networks in autonomous driving systems: A survey. IEEE Access 2022, 10, 14076–14119. [Google Scholar] [CrossRef]
- Kar, M.K.; Nath, M.K.; Neog, D.R. A review on progress in semantic image segmentation and its application to medical images. S.N. Comput. Sci. 2021, 2, 397. [Google Scholar] [CrossRef]
- Mo, Y.; Wu, Y.; Yang, X.; Liu, F.; Liao, Y. Review the state-of-the-art technologies of semantic segmentation based on deep learning. Neurocomputing 2022, 493, 626–646. [Google Scholar] [CrossRef]
- Thisanke, H.; Deshan, C.; Chamith, K.; Senviratne, S.; Vidanaarachchi, R. Semantic segmentation using Vision Transformers: A survey. Eng. Appl. Artif. Intell. 2023, 126, 106669. [Google Scholar] [CrossRef]
- Yu, J.; Zhang, J.; Shu, A.; Chen, Y.; Chen, J.; Yang, Y.; Tang, W.; Zhang, Y. Study of convolutional neural network-based semantic segmentation methods on edge intelligence devices for field agricultural robot navigation line extraction. Comput. Electron. Agric. 2023, 209, 107811. [Google Scholar] [CrossRef]
- Zhang, L.; Li, M.; Zhu, X.; Chen, Y.; Huang, J.; Wang, Z.; Hu, T.; Wang, Z.; Fang, K. Navigation path recognition between rows of fruit trees based on semantic segmentation. Comput. Electron. Agric. 2024, 216, 108511. [Google Scholar] [CrossRef]
- Zheng, Z.; Hu, Y.; Li, X.; Huang, Y. Autonomous navigation method of jujube catch-and-shake harvesting robot based on convolutional neural networks. Comput. Electron. Agric. 2023, 215, 108469. [Google Scholar] [CrossRef]
- Yang, Z.; Ouyang, L.; Zhang, Z.; Duan, J.; Yu, J.; Wang, H. Visual navigation path extraction of orchard hard pavement based on scanning method and neural network. Comput. Electron. Agric. 2022, 197, 106964. [Google Scholar] [CrossRef]
- Li, X.; Su, J.; Yue, Z.; Duan, F. Adaptive multi-ROI agricultural robot navigation line extraction based on image semantic segmentation. Sensors 2022, 22, 7707. [Google Scholar] [CrossRef]
- Cao, M.; Tang, F.; Ji, P.; Ma, F. Improved real-time semantic segmentation network model for crop vision navigation line detection. Front. Plant Sci. 2022, 13, 898131. [Google Scholar] [CrossRef]
- Zhang, X.; Yang, Y.; Li, Z.; Ning, X.; Qin, Y.; Cai, W. An improved encoder-decoder network based on strip pool method applied to segmentation of farmland vacancy field. Entropy 2021, 23, 435. [Google Scholar] [CrossRef]
- Baheti, B.; Innani, S.; Gajre, S.; Talbar, S. Semantic scene segmentation in unstructured environment with modified DeepLabV3+. Pattern Recognit. Lett. 2020, 138, 223–229. [Google Scholar] [CrossRef]
- Bai, H.; Cheng, J.; Huang, X.; Liu, S.; Deng, C. HCANet: A Hierarchical Context Aggregation Network for Semantic Segmentation of High-Resolution Remote Sensing Images. IEEE Geosci. Remote Sens. Lett. 2022, 19, 6002105. [Google Scholar] [CrossRef]
- Dais, D.; Bal, I.E.; Smyrou, E.; Sarhosis, V. Automatic crack classification and segmentation on masonry surfaces using convolutional neural networks and transfer learning. Autom. Constr. 2021, 125, 103606. [Google Scholar] [CrossRef]
- Deng, J.; Lu, Y.; Lee, V.C.-S. Concrete crack detection with handwriting script interferences using faster region-based convolutional neural network. Comput. Aided Civ. Infrastruct. Eng. 2020, 35, 373–388. [Google Scholar] [CrossRef]
- Zhou, Z.; Zhang, J.; Gong, C. Hybrid semantic segmentation for tunnel lining cracks based on Swin Transformer and convolutional neural network. Comput.-Aided Civ. Infrastruct. Eng. 2023, 38, 2491–2510. [Google Scholar] [CrossRef]
- Lin, T.-Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal Loss for Dense Object Detection. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 42, 318–327. [Google Scholar] [CrossRef] [PubMed]
- Qu, Z.; Mei, J.; Liu, L.; Zhou, D.-Y. Crack Detection of Concrete Pavement With Cross-Entropy Loss Function and Improved VGG16 Network Model. IEEE Access 2020, 8, 54564–54573. [Google Scholar] [CrossRef]
- Pratik, V.; Vedhapriyavadhana, R.; Chidambaranathan, S. Polyp Segmentation Using UNet and Enet. In Proceedings of the 2023 6th International Conference on Recent Trends in Advance Computing (ICRTAC), Chennai, India, 14–15 December 2023; pp. 516–522. [Google Scholar]
- Hu, X.; Ren, H. A Road Scene Semantic Segmentation Algorithm Based on Improved BiSeNetV2. In Proceedings of the 5th International Seminar on Artificial Intelligence, Networking and Information Technology (AINIT 2024), Nanjing, China, 29–31 March 2024; pp. 649–652. [Google Scholar]
- Sola, D.; Scott, K.A. Efficient Shallow Network for River Ice Segmentation. Remote Sens. 2022, 14, 2378. [Google Scholar] [CrossRef]
- Lavrynenko, R.; Ryabova, N. Transforming Semantic Segmentation into Instance Segmentation with a Guided U-Net. In Proceedings of the 2023 IEEE 18th International Conference on Computer Science and Information Technologies (CSIT), Lviv, Ukraine, 19–21 October 2023; pp. 1–4. [Google Scholar]
- Zhang, C.; Zhao, J.; Feng, Y. Research on Semantic Segmentation Based on Improved PSPNet. In Proceedings of the 2023 International Conference on Intelligent Perception and Computer Vision (CIPCV), Xi’an, China, 19–21 May 2023; pp. 1–6. [Google Scholar]
- Farhangfar, S.; Rezaeian, M. Semantic Segmentation of Aerial Images using FCN-based Network. In Proceedings of the 2019 27th Iranian Conference on Electrical Engineering (ICEE), Yazd, Iran, 30 April–2 May 2019; pp. 1864–1868. [Google Scholar]
- Chen, L.-C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 40, 834–848. [Google Scholar] [CrossRef]
- Inbaraj, X.; Villavicencio, C.; Macrohon, J.; Jeng, J.-H.; Hsieh, J.-G. Object identification and localization using Grad-CAM++ with mask regional convolution neural network. Electronics 2021, 10, 1541. [Google Scholar] [CrossRef]
Method | MIoU/% | Pr/% | Re/% | Size/M |
---|---|---|---|---|
Xception | 81.51 | 83.62 | 86.32 | 209 |
GhostNet | 84.27 | 86.47 | 84.37 | 24 |
MobileNetV3 | 86.36 | 90.33 | 87.81 | 31.2 |
Our | 89.66 | 90.22 | 90.03 | 31.2 |
Structure | MIoU/% |
---|---|
Xception + ASPP | 84.67 |
Mobilenetv3-ECA + ASPP | 90.66 |
Xception + SwinASPP | 92.34 |
Mobilenetv3-ECA + SwinASPP | 95.73 |
Model | MIoU/% |
---|---|
Enet | 84.90 |
Bisenetv2 | 84.36 |
IRASPP | 88.84 |
U-Net | 92.58 |
PSPNet | 90.83 |
FCN | 70.3 |
DeepLabV3+ | 90.57 |
Our | 95.73 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Liang, M.; Ding, L.; Chen, J.; Xu, L.; Wang, X.; Li, J.; Yang, H. SwinLabNet: Jujube Orchard Drivable Area Segmentation Based on Lightweight CNN-Transformer Architecture. Agriculture 2024, 14, 1760. https://doi.org/10.3390/agriculture14101760
Liang M, Ding L, Chen J, Xu L, Wang X, Li J, Yang H. SwinLabNet: Jujube Orchard Drivable Area Segmentation Based on Lightweight CNN-Transformer Architecture. Agriculture. 2024; 14(10):1760. https://doi.org/10.3390/agriculture14101760
Chicago/Turabian StyleLiang, Mingxia, Longpeng Ding, Jiangchun Chen, Liming Xu, Xinjie Wang, Jingbin Li, and Hongfei Yang. 2024. "SwinLabNet: Jujube Orchard Drivable Area Segmentation Based on Lightweight CNN-Transformer Architecture" Agriculture 14, no. 10: 1760. https://doi.org/10.3390/agriculture14101760
APA StyleLiang, M., Ding, L., Chen, J., Xu, L., Wang, X., Li, J., & Yang, H. (2024). SwinLabNet: Jujube Orchard Drivable Area Segmentation Based on Lightweight CNN-Transformer Architecture. Agriculture, 14(10), 1760. https://doi.org/10.3390/agriculture14101760