Expediting the Convergence of Global Localization of UAVs through Forward-Facing Camera Observation
Abstract
:1. Introduction
- Unlike a downward-facing camera that has a limited observational range, a forward-facing camera mounted on a UAV can capture extensive scene information ahead, even at lower flight altitudes;
- The forward-facing camera is utilized not only for localization but also for detecting obstacles, thereby improving flight safety;
- In some rescue and inspection application scenarios, the forward-facing camera can be used simultaneously for localization as well as for the detection and identification of objects. This will enhance the task execution ability under GNSS denial circumstances.
- We proposed a method for constructing a robust 2.5D grid feature descriptor using two consecutive frames captured by a forward-facing camera to construct a two-view structure from motion (SFM) model. The 3D points derived from the SFM were projected onto the 2.5D grid feature descriptor, with varying weights assigned to each grid based on odometry noise and the 3D points. This descriptor offers terrain information over a large area and provides more comprehensive observations than those obtained using a downward-facing camera;
- We define global localization as a state estimation problem based on the PMF. We quantified the matching information between the described features and the DEM as a probability distribution, integrating it into the PMF framework as an additional observation. We developed two truncation methods to enhance computational efficiency and accelerate the convergence rate of localization. The first method truncated the convolution kernel size for state prediction, and the other truncated the posterior state probability distribution using a sliding window. These approaches aim to improve the computational efficiency and expedite the convergence ratio of localization;
- The experimental results from real flight sequences indicate that our method can accelerate the convergence rate of localization during the takeoff, ascent, and cruise flight stages. Additionally, the method effectively enhances localization accuracy and robustness in complex scenarios such as uneven terrain and ambiguous scenes;
- The proposed method demonstrates high scalability and the potential for seamless integration into any Bayesian filtering state estimation framework. Moreover, the observational model can be extended to include multiple pinhole or fisheye cameras, which further enhance the localization performance.
2. Related Works
3. Method
3.1. Overview of the Method
3.2. Methodology for Constructing Descriptor
3.2.1. Methodology for Constructing 3D Point Cloud Using Two-View SFM
3.2.2. Mapping of 3D Points to Descriptor
3.3. Methodology for Matching and Calculating Likelihood
3.3.1. Expression of the Similarity between the Descriptor and the Patch
3.3.2. Design of
3.3.3. Design of
3.4. Methodology for Multi-Observation Fusion and Truncation
3.4.1. Multi-Observation Fusion Method Based on PMF
3.4.2. Truncation Method
- (a)
- Convolution kernel truncation
- (b)
- Sliding Window-Based probability truncation
4. Experiments
4.1. Experimental Setting
- Forward-facing camera observation thread: Acquire the images captured by the forward-facing camera at time and time and generate descriptors and compute the observation likelihood via the approaches presented in Section 3.2 and Section 3.3 of this article;
- Downward-facing camera observation thread: Acquire the image captured by the downward-facing camera at time and compute the observation likelihood based on a prevalent image-matching methodology;
- Odometry thread: Calculate the odometry information between time and time based on a prevalent visual odometry methodology.
4.2. Experimental Procedure
5. Results and Discussion
5.1. Influence of Type of Observation Mode
5.2. Influence of Sliding Window Probability Truncation
5.3. Influence of the Utilization of Observation from the Forward-Facing Camera during the Takeoff and Ascent Stages
5.4. Influence of Truncating the Convolution Kernel
5.5. Consideration of the Noise Model
5.6. Summary of Discussion
6. Conclusions
Author Contributions
Funding
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- Holland, R.A. The role of visual landmarks in the avian familiar area map. J. Exp. Biol. 2003, 206, 1773–1778. [Google Scholar] [CrossRef] [PubMed]
- Forster, C.; Pizzoli, M.; Scaramuzza, D. SVO: Fast semi-direct monocular visual odometry. In Proceedings of the 2014 IEEE International Conference on Robotics and Automation (ICRA), Hong Kong, China, 31 May–7 June 2014; pp. 15–22. [Google Scholar]
- Qin, T.; Li, P.; Shen, S. Vins-mono: A robust and versatile monocular visual-inertial state estimator. IEEE Trans. Robot. 2018, 34, 1004–1020. [Google Scholar] [CrossRef]
- Campos, C.; Elvira, R.; Rodríguez JJ, G.; Montiel, J.M.; Tardós, J.D. Orb-slam3: An accurate open-source library for visual, visual–inertial, and multimap slam. IEEE Trans. Robot. 2021, 37, 1874–1890. [Google Scholar] [CrossRef]
- Guizilini, V.; Ramos, F. Visual odometry learning for unmanned aerial vehicles. In Proceedings of the 2011 IEEE International Conference on Robotics and Automation, Shanghai, China, 9–13 May 2011; pp. 6213–6220. [Google Scholar]
- Virrankoski, R. Autonomous Sensing using Satellites, Multicopters, Sensors and Actuators (MULTICO)-Final Report; Aalto-yliopisto: Helsinki, Finland, May 2023. [Google Scholar]
- Dellaert, F.; Fox, D.; Burgard, W.; Thrun, S. Monte carlo localization for mobile robots. In Proceedings of the 1999 IEEE International Conference on Robotics and Automation (Cat. No. 99CH36288C), Detroit, MI, USA, 10–15 May 1999; Volume 2, pp. 1322–1328. [Google Scholar]
- Bergman, N. Recursive Bayesian Estimation: Navigation and Tracking Applications. Ph.D. Thesis, Linköping University, Linköping, Sverige, 1999. [Google Scholar]
- Duník, J.; Biswas, S.K.; Dempster, A.G.; Pany, T.; Closas, P. State estimation methods in navigation: Overview and application. IEEE Aerosp. Electron. Syst. Mag. 2020, 35, 16–31. [Google Scholar] [CrossRef]
- Kinnari, J.; Verdoja, F.; Kyrki, V. GNSS-denied geolocalization of UAVs by visual matching of onboard camera images with orthophotos. In Proceedings of the 2021 20th international conference on advanced robotics (ICAR), Ljubljana, Slovenia, 6–10 December 2021; pp. 555–562. [Google Scholar]
- Xu, Y.; Pan, L.; Du, C.; Li, J.; Jing, N.; Wu, J. Vision-based UAVs aerial image localization: A survey. In Proceedings of the 2nd ACM SIGSPATIAL International Workshop on AI for Geographic Knowledge Discovery, Seattle, WA, USA, 6 November 2018; pp. 9–18. [Google Scholar]
- Couturier, A.; Akhloufi, M.A. A review on absolute visual localization for UAV. Robot. Auton. Syst. 2021, 135, 103666. [Google Scholar] [CrossRef]
- Shan, M.; Wang, F.; Lin, F.; Gao, Z.; Tang, Y.Z.; Chen, B.M. Google map aided visual navigation for UAVs in GPS-denied environment. In Proceedings of the 2015 IEEE international conference on robotics and biomimetics (ROBIO), Zhuhai, China, 6–9 December 2015; pp. 114–119. [Google Scholar]
- Yol, A.; Delabarre, B.; Dame, A.; Dartois, J.E.; Marchand, E. Vision-based absolute localization for unmanned aerial vehicles. In Proceedings of the2014 IEEE/RSJ International Conference on Intelligent Robots and Systems, Chicago, IL, USA, 14–18 September 2014; pp. 3429–3434. [Google Scholar]
- Samano, N.; Zhou, M.; Calway, A. Global aerial localisation using image and map embeddings. In Proceedings of the 2021 IEEE International Conference on Robotics and Automation (ICRA), Xi’an, China, 30 May–5 June 2021; pp. 5788–5794. [Google Scholar]
- Jurevičius, R.; Marcinkevičius, V.; Šeibokas, J. Robust GNSS-denied localization for UAV using particle filter and visual odometry. Mach. Vis. Appl. 2019, 30, 1181–1190. [Google Scholar] [CrossRef]
- Van Dalen, G.J.; Magree, D.P.; Johnson, E.N. Absolute localization using image alignment and particle filtering. In Proceedings of the AIAA Guidance, Navigation, and Control Conference, San Diego, CA, USA, 4–8 January 2016; p. 0647. [Google Scholar]
- Wan, X.; Liu, J.; Yan, H.; Morgan, G.L. Illumination-invariant image matching for autonomous UAV localisation based on optical sensing. ISPRS J. Photogramm. Remote Sens. 2016, 119, 198–213. [Google Scholar] [CrossRef]
- Tang, H.; Shi, J.; Miao, X.; Wu, R.; Yang, D. UAV Visual Localization Technology Based on Heterogenous Remote Sensing Image Matching. In Proceedings of the 2022 5th International Conference on Artificial Intelligence and Pattern Recognition, Xiamen China, 23–25 September2022; pp. 773–779. [Google Scholar]
- Cao, Y.; Ren, K.; Chen, Q. Template matching based on convolution neural network for UAV visual localization. Optik 2023, 283, 170920. [Google Scholar] [CrossRef]
- Li, H.; Wang, J.; Wei, Z.; Xu, W. Jointly Optimized Global-Local Visual Localization of UAVs. arXiv 2023, arXiv:2310.08082. [Google Scholar]
- Kinnari, J.; Verdoja, F.; Kyrki, V. Season-invariant GNSS-denied visual localization for UAVs. IEEE Robot. Autom. Lett. 2022, 7, 10232–10239. [Google Scholar] [CrossRef]
- Hanyu, W.A.N.G.; Qiang, S.H.E.N.; Zilong, D.E.N.G.; Xinyi, C.A.O.; Wang, X. Absolute pose estimation of UAV based on large-scale satellite image. Chin. J. Aeronaut. 2023, 37, 219–231. [Google Scholar]
- Tian, Y.; Chen, C.; Shah, M. Cross-view image matching for geo-localization in urban environments. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 3608–3616. [Google Scholar]
- Tian, X.; Shao, J.; Ouyang, D.; Shen, H.T. UAV-satellite view synthesis for cross-view geo-localization. IEEE Trans. Circuits Syst. Video Technol. 2021, 32, 4804–4815. [Google Scholar] [CrossRef]
- Choi, J.; Myung, H. BRM localization: UAV localization in GNSS-denied environments based on matching of numerical map and UAV images. In Proceedings of the 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Las Vegas, NV, USA, 25–29 October 2020; pp. 4537–4544. [Google Scholar]
- Mantelli, M.; Pittol, D.; Neuland, R.; Ribacki, A.; Maffei, R.; Jorge, V.; Prestes, E.; Kolberg, M. A novel measurement model based on abBRIEF for global localization of a UAV over satellite images. Robot. Auton. Syst. 2019, 112, 304–319. [Google Scholar] [CrossRef]
- Kinnari, J.; Renzulli, R.; Verdoja, F.; Kyrki, V. LSVL: Large-scale season-invariant visual localization for UAVs. Robot. Auton. Syst. 2023, 168, 104497. [Google Scholar] [CrossRef]
- Thrun, S. Probabilistic robotics. Commun. ACM 2002, 45, 52–57. [Google Scholar] [CrossRef]
- Hartley, R.; Zisserman, A. Multiple View Geometry in Computer Vision; Cambridge University Press: Cambridge, UK, 2003. [Google Scholar]
- Mosbrucker, A.R.; Major, J.J.; Spicer, K.R.; Pitlick, J. Camera system considerations for geomorphicCamera system considerations for geomorphic applications of SfM photogrammetry. Earth Surf. Process. Landf. 2017, 42, 969–986. [Google Scholar] [CrossRef]
- Fischler, M.A.; Bolles, R.C. Random sample consensus: A paradigm for model fitting with applications to image analysis and automated cartography. Commun. ACM 1981, 24, 381–395. [Google Scholar] [CrossRef]
- Roma, N.; Santos-Victor, J.; Tomé, J. A comparative analysis of cross-correlation matching algorithms using a pyramidal resolution approach. In Empirical Evaluation Methods in Computer Vision; World Scientific Publishing: Stockholm, Sweden, May 2002; pp. 117–142. [Google Scholar]
- Mur-Artal, R.; Tardós, J.D. Orb-slam2: An open-source slam system for monocular, stereo, and rgb-d cameras. IEEE Trans. Robot. 2017, 33, 1255–1262. [Google Scholar] [CrossRef]
- Chiodini, S.; Giubilato, R.; Pertile, M.; Debei, S. Retrieving scale on monocular visual odometry using low-resolution range sensors. IEEE Trans. Instrum. Meas. 2020, 69, 5875–5889. [Google Scholar] [CrossRef]
Experiment Index | Variable-1 | Variable-2 | Variable-3 | ITE | ALE (m) | ALStd (m) | RT (s) |
---|---|---|---|---|---|---|---|
SAMPLE-00 | Type 1 | No | – | 67 | 165.4 | 321.9 | 13.2 |
SAMPLE-01 | Type 1 | Yes | – | 30 | 55.4 | 90.7 | 0.078 |
SAMPLE-02 | Type 2 | No | No | 19 | 37.6 | 107.2 | 13.1 |
SAMPLE-03 | Type 2 | No | Yes | 19 | 37.4 | 105.4 | 13.5 |
SAMPLE-04 | Type 2 | Yes | No | 14 | 34.4 | 77.9 | 0.072 |
SAMPLE-05 | Type 2 | Yes | Yes | 12 | 34.4 | 79.9 | 0.069 |
Experiment Index | Average Localization Error (m) (9 Iterations Prior to Convergence) | Average Localization Standard Deviation (m) (9 Iterations Prior to Convergence) |
---|---|---|
SAMPLE-02 | 771.0 | 1336.6 |
SAMPLE-03 | 660.2 | 1263.9 |
SAMPLE-04 | 1139.4 | 1480.3 |
SAMPLE-05 | 924.0 | 1231.8 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Li, Z.; Jiang, X.; Ma, S.; Ma, X.; Lv, Z.; Ding, H.; Ji, H.; Sun, Z. Expediting the Convergence of Global Localization of UAVs through Forward-Facing Camera Observation. Drones 2024, 8, 335. https://doi.org/10.3390/drones8070335
Li Z, Jiang X, Ma S, Ma X, Lv Z, Ding H, Ji H, Sun Z. Expediting the Convergence of Global Localization of UAVs through Forward-Facing Camera Observation. Drones. 2024; 8(7):335. https://doi.org/10.3390/drones8070335
Chicago/Turabian StyleLi, Zhenyu, Xiangyuan Jiang, Sile Ma, Xiaojing Ma, Zhenyi Lv, Hongliang Ding, Haiyan Ji, and Zheng Sun. 2024. "Expediting the Convergence of Global Localization of UAVs through Forward-Facing Camera Observation" Drones 8, no. 7: 335. https://doi.org/10.3390/drones8070335
APA StyleLi, Z., Jiang, X., Ma, S., Ma, X., Lv, Z., Ding, H., Ji, H., & Sun, Z. (2024). Expediting the Convergence of Global Localization of UAVs through Forward-Facing Camera Observation. Drones, 8(7), 335. https://doi.org/10.3390/drones8070335