Testing a Vision-Based Autonomous Drone Navigation Model in a Forest Environment
Abstract
:1. Introduction
- The localization of visual objects is performed via object detection to give the algorithms we parse in Section 3 contextual information of the scene for further processing. In Section 3, scenes are described using color-based segmentation with a masking approach to provide a comprehensive semantic representation of their relative environments.
- The vision-based navigation model can autonomously guide a drone in a dynamic and remote environment.
- The masking model is innovatively employed on the basis of vision to autonomously navigate the drone in a remote environment and without any connection with the center.
- The masking model is innovatively used based on the vision to autonomously navigate.
- A drone can navigate in remote environments without any connection with the center.
2. Related Works
3. The Proposed Method
Algorithm 1 Clear Path Determination Algorithm |
|
3.1. Object Detection
Algorithm 2 Drone Navigation Algorithm with Object Detection, Semantic Segmentation, Thresholding, and Navigation |
|
- They are trained on large-scale datasets, providing a robust foundation for transfer learning with minimal additional training data.
- They are designed to output bounding boxes alongside class predictions and they facilitate object localization, a critical aspect of object detection.
- They are selected for feature extraction based on their high mAP (mean average precision) scores on benchmark datasets like COCO, indicating superior performance in object detection tasks.
- We evaluated the performance of these feature extraction models on datasets with varying compositions. While trained on real trees, they were also tested on 3D model trees [link to dataset], exhibiting effective performance in both scenarios.
- The green region corresponds to the bounding box of the detected tree, while the red region represents the fixed field of view (FOV) of the drone.
3.2. Masking
3.3. Drone Navigation with Simulator
3.4. Model Training
4. Results and Discussion
4.1. Object Detection Results
4.2. Flight Direction Results
4.3. Autonomous Navigation Results
4.4. Discussion
5. Conclusions
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- The National Coordination Office (NCO). National Coordination Office for Space-Based Positioning, Navigation, and Timing. 2013. Available online: https://www.gps.gov/governance/excom/nco/ (accessed on 18 May 2021).
- Karpenko, S.; Konovalenko, I.; Miller, A.; Miller, B.; Nikolaev, D. UAV control on the basis of 3D landmark bearing-only observations. Sensors 2015, 15, 29802–29820. [Google Scholar] [CrossRef] [PubMed]
- Kato, A.; Obanawa, H.; Hayakawa, Y.; Watanabe, M.; Yamaguchi, Y.; Enoki, T. Fusion between UAV-SFM and terrestrial laser scanner for field validation of satellite remote sensing. In Proceedings of the 2015 IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Milan, Italy, 26–31 July 2015; IEEE: Piscataway, NJ, USA, 2015; pp. 2642–2645. [Google Scholar]
- Pham, H.X.; La, H.M.; Feil-Seifer, D.; Nguyen, L.V. Autonomous UAV Navigation Using Reinforcement Learning. arXiv 2018, arXiv:1801.05086. [Google Scholar]
- Xiao, B.; Yin, S. A new disturbance attenuation control scheme for quadrotor unmanned aerial vehicles. IEEE Trans. Ind. Informatics 2017, 13, 2922–2932. [Google Scholar] [CrossRef]
- Roberge, V.; Tarbouchi, M.; Labonté, G. Comparison of parallel genetic algorithm and particle swarm optimization for real-time UAV path planning. IEEE Trans. Ind. Inform. 2012, 9, 132–141. [Google Scholar] [CrossRef]
- Cavaliere, D.; Loia, V.; Saggese, A.; Senatore, S.; Vento, M. Semantically Enhanced UAVs to Increase the Aerial Scene Understanding. IEEE Trans. Syst. Man Cybern. Syst. 2019, 49, 555–567. [Google Scholar] [CrossRef]
- Yong, S.; Yeong, Y. Human Object Detection in Forest with Deep Learning based on Drone’s Vision. In Proceedings of the 2018 4th International Conference on Computer and Information Sciences (ICCOINS), Kuala Lumpur, Malaysia, 13–14 August 2018; pp. 1–5. [Google Scholar] [CrossRef]
- Yong, S.; Deng, J.D.; Purvis, M.K. Modelling semantic context for novelty detection in wildlife scenes. In Proceedings of the 2010 IEEE International Conference on Multimedia and Expo, Singapore, 19–23 July 2010; pp. 1254–1259. [Google Scholar] [CrossRef]
- Cai, J.; Huang, P.; Zhang, B.; Wang, D. A TSR visual servoing system based on a novel dynamic template matching method. Sensors 2015, 15, 32152–32167. [Google Scholar] [CrossRef] [PubMed]
- Chen, T.; Lu, S. Object-level motion detection from moving cameras. IEEE Trans. Circuits Syst. Video Technol. 2017, 27, 2333–2343. [Google Scholar] [CrossRef]
- Saripalli, S.; Montgomery, J.F.; Sukhatme, G.S. Visually guided landing of an unmanned aerial vehicle. IEEE Trans. Robot. Autom. 2003, 19, 371–380. [Google Scholar] [CrossRef]
- Bruni, V.; Vitulano, D. An improvement of kernel-based object tracking based on human perception. IEEE Trans. Syst. Man Cybern. Syst. 2014, 44, 1474–1485. [Google Scholar] [CrossRef]
- Dang, C.T.; Pham, H.T.; Pham, T.B.; Truong, N.V. Vision based ground object tracking using AR. Drone quadrotor. In Proceedings of the 2013 International Conference on Control, Automation and Information Sciences (ICCAIS), Nha Trang, Vietnam, 25–28 November 2013; IEEE: Piscataway, NJ, USA, 2013; pp. 146–151. [Google Scholar]
- Thiang, I.N.; LuMaw, H.M.T. Vision-based object tracking algorithm with ar. drone. Red 2016, 160, 179. [Google Scholar]
- Szegedy, C.; Toshev, A.; Erhan, D. Deep neural networks for object detection. In Proceedings of the Advances in Neural Information Processing Systems, Lake Tahoe, NV, USA, 5–10 December 2013; pp. 2553–2561. [Google Scholar]
- Erhan, D.; Szegedy, C.; Toshev, A.; Anguelov, D. Scalable object detection using deep neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 2147–2154. [Google Scholar]
- Yong, S.P.; Deng, J.D.; Purvis, M.K. Novelty detection in wildlife scenes through semantic context modelling. Pattern Recognit. 2012, 45, 3439–3450. [Google Scholar] [CrossRef]
- Filonenko, A.; Jo, K.H. Unattended object identification for intelligent surveillance systems using sequence of dual background difference. IEEE Trans. Ind. Inform. 2016, 12, 2247–2255. [Google Scholar]
- Zhou, Z.; Zhang, C.; Xu, C.; Xiong, F.; Zhang, Y.; Umer, T. Energy-efficient industrial internet of uavs for power line inspection in smart grid. IEEE Trans. Ind. Inform. 2018, 14, 2705–2714. [Google Scholar] [CrossRef]
- Huang, H.; Savkin, A.V. An algorithm of reactive collision free 3D deployment of networked unmanned aerial vehicles for surveillance and monitoring. IEEE Trans. Ind. Inform. 2019, 16, 132–140. [Google Scholar] [CrossRef]
- Yang, L.; Sun, Q.; Ye, Z.S. Designing Mission Abort Strategies Based on Early-Warning Information: Application to UAV. IEEE Trans. Ind. Inform. 2019, 16, 277–287. [Google Scholar] [CrossRef]
- Huang, Y.P.; Sithole, L.; Lee, T.T. Structure From Motion Technique for Scene Detection Using Autonomous Drone Navigation. IEEE Trans. Syst. Man Cybern. Syst. 2017, 49, 2559–2570. [Google Scholar] [CrossRef]
- Minaeian, S.; Liu, J.; Son, Y.J. Vision-based target detection and localization via a team of cooperative UAV and UGVs. IEEE Trans. Syst. Man Cybern. Syst. 2016, 46, 1005–1016. [Google Scholar] [CrossRef]
- Dementhon, D.F.; Davis, L.S. Model-based object pose in 25 lines of code. Int. J. Comput. Vis. 1995, 15, 123–141. [Google Scholar] [CrossRef]
- Zhang, H.; Geiger, A.; Urtasun, R. Understanding high-level semantics by modeling traffic patterns. In Proceedings of the IEEE International Conference on Computer Vision, Sydney, Australia, 1–8 December 2013; pp. 3056–3063. [Google Scholar]
- Guo, Y.; Liu, Y.; Georgiou, T.; Lew, M.S. A review of semantic segmentation using deep neural networks. Int. J. Multimed. Inf. Retr. 2018, 7, 87–93. [Google Scholar] [CrossRef]
- Arbeláez, P.; Pont-Tuset, J.; Barron, J.T.; Marques, F.; Malik, J. Multiscale combinatorial grouping. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 328–335. [Google Scholar]
- Zitnick, C.L.; Dollár, P. Edge boxes: Locating object proposals from edges. In Proceedings of the European Conference on Computer Vision, Zurich, Switzerland, 6–12 September 2014; Springer: Cham, Switzerland, 2014; pp. 391–405. [Google Scholar]
- Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich feature hierarchies for accurate object detection and semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 580–587. [Google Scholar]
- Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada, 7–12 December 2015; pp. 91–99. [Google Scholar]
- Uijlings, J.R.; Van De Sande, K.E.; Gevers, T.; Smeulders, A.W. Selective search for object recognition. Int. J. Comput. Vis. 2013, 104, 154–171. [Google Scholar] [CrossRef]
- 3D Robotics. Guiding and Controlling Copter. Available online: https://dronekit-python.readthedocs.io/en/latest/guide/copter/guided_mode.html (accessed on 26 August 2020).
- Girshick, R. Fast r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar]
- He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 770–778. [Google Scholar]
- ArduPilot Dev Team. APM Planner 2 Home. Available online: https://ardupilot.org/planner2/ (accessed on 18 April 2021).
Velocity Operation | Direction |
---|---|
Fly NORTH | |
Fly SOUTH | |
Fly EAST | |
Fly WEST | |
Ascend | |
Descend |
Coordinate x | Flight Direction |
---|---|
5.23 | NORTH |
4.11 | NORTH |
5.68 | NORTH |
3.19 | NORTH |
0.22 | NORTH |
−1.15 | SOUTH |
−0.19 | SOUTH |
−1.18 | SOUTH |
−4.23 | SOUTH |
−5.87 | SOUTH |
Coordinate y | Flight Direction |
---|---|
−2.23 | WEST |
−0.11 | WEST |
−2.68 | WEST |
−4.19 | WEST |
−0.92 | WEST |
2.15 | EAST |
0.33 | EAST |
1.18 | EAST |
3.12 | EAST |
5.56 | EAST |
Video | No. of Frames | No. of Errors | Accuracy |
---|---|---|---|
1 | 500 | 55 | 89.0% |
2 | 500 | 0 | 100.0% |
3 | 500 | 0 | 100.0% |
4 | 500 | 11 | 97.8% |
5 | 500 | 3 | 99.4% |
6 | 500 | 0 | 100.0% |
Average | 97.7% |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Lee, A.; Yong, S.-P.; Pedrycz, W.; Watada, J. Testing a Vision-Based Autonomous Drone Navigation Model in a Forest Environment. Algorithms 2024, 17, 139. https://doi.org/10.3390/a17040139
Lee A, Yong S-P, Pedrycz W, Watada J. Testing a Vision-Based Autonomous Drone Navigation Model in a Forest Environment. Algorithms. 2024; 17(4):139. https://doi.org/10.3390/a17040139
Chicago/Turabian StyleLee, Alvin, Suet-Peng Yong, Witold Pedrycz, and Junzo Watada. 2024. "Testing a Vision-Based Autonomous Drone Navigation Model in a Forest Environment" Algorithms 17, no. 4: 139. https://doi.org/10.3390/a17040139
APA StyleLee, A., Yong, S. -P., Pedrycz, W., & Watada, J. (2024). Testing a Vision-Based Autonomous Drone Navigation Model in a Forest Environment. Algorithms, 17(4), 139. https://doi.org/10.3390/a17040139