Intelligent Methods for Forest Fire Detection Using Unmanned Aerial Vehicles
Abstract
:1. Introduction
- In the first stage, the overall quality of the video stream from the UAV’s technical vision system is improved, including removing motion blur in frames and stabilizing the video stream. This stage helps minimize pictures’ artifacts on video frames and potentially improves the quality of detection in regions of interest (fire and smoke).
- In the second stage, the horizon line is detected, which allows areas with clouds to be cut off as they can be incorrectly recognized as smoke. his improves the quality of detection too.
- Section 2 contains a review of previous methods and approaches for the started tasks.
- All proposed methods are described in Section 3.
- The results of all the described methods are shown in Section 4.
- We discuss the limitations of the proposed algorithms and describe future work in Section 5.
- The conclusion and general results of this research are given in Section 6.
2. Related Works
3. Materials and Methods
- We propose improving the quality of the video stream method: Removing blur and video stream stabilization. This procedure is necessary in cases where, during sudden movements of the UAV caused by wind gusts, the image becomes unclear, which prevents good fire and smoke recognition. To solve this problem, various means of removing grease are used.
- Another approach that significantly affects the quality of recognition of fires is associated with removing part of the image from video frames, covering the upper part with the sky. This is an important step since most of the recognition errors are associated with weather conditions, such as when clouds appear in the sky adjacent to the forest area, as well as reflections during sunsets and sunrises, which leads to false positives of the recognition algorithm.
- Finally, we solve the problem of detecting forest fires with the modified CNN «YOLO». Fine-tuning of this network is provided through the selection of layers and neurons.
3.1. Deconvolution and Video Stream Stabilization
3.2. Image Semantic Segmentation
3.3. Horizon Line Detection
- Calculation of factors that affect the size of the sliding window and the threshold of informativeness for filtering out non-informative points. Factors characterizing the main features of displaying the horizon line on a photo are calculated, and models (equations) of multiple linear regression are constructed:
- 2.
- Normalization of the image brightness in grayscale.
- 3.
- Calculation and normalization of informative areas of the image.
- 4.
- Filtering informative pixels of the image based on the brightness of the areas above and below.
- 5.
- Calculation of factor values , , regression model coefficients , and threshold using Equation (2).
- 6.
- Filtering out points with an informativeness level below the threshold .
- 7.
- Grouping points into lines.
- 8.
- Grouping lines.
- 9.
- Selection of the line with the greatest length. The selected line is assigned as the “central part” of the horizon line. In Figure 3f–h, fragments of the “central part” are indicated in turquoise.
- 10.
- Restoration of gaps between the segments of the “central part” of the horizon line.
- 11.
- Calculation of the “right part” of the horizon line.
- 12.
- Calculation of the “Calculation of the “right part” of the horizon line.
- 13.
- Checking the horizon line for truthfulness. The mean brightness values in grayscale above and below the horizon line are calculated and compared. Let be the number of points on the horizon line, be the coordinates of the -th point, and be the brightness in grayscale of the -th point on the horizon line. If the condition
- 14.
- Smoothing of the “left part” of the horizon line.
- 15.
- Smoothing of the “right part” of the horizon line.
- 16.
- Displaying the horizon line on the image (see Figure 3i,k).
- 17.
- End.
3.4. Detection of Forest Fires with Convolutional Neural Networks
4. Results
4.1. Results of Deconvolution and Video Stream Stabilization
4.2. Results of Image Semantic Segmentation
4.3. Results of Horizon Line Detection
4.4. Detection of Forest Fires with Convolutional Neural Networks
- Yolo 4 explicitly selected the ADAM optimizer.
- Yolo 5 explicitly selected the AdamW optimizer.
- Yolo 6 did not change the optimizer and used the default SGD in the code.
- Yolo 7 did not change the optimizer and used the default SGD in the code.
- Yolo 8 explicitly selected the AdamW optimizer.
5. Discussion
- Fog and snow in the processed image make the algorithm’s operation challenging (increased processing time and decreased accuracy in detecting the horizon line).
- The applicability and effectiveness of the algorithm depend on the orientation of the UAV’s camera.
- Processing satellite images of forest areas. Establishing aerospace monitoring will enhance the speed and accuracy of fire localization.
- Addressing software and hardware acceleration issues to achieve real-time computations for processing multiple parallel incoming video streams.
- Addressing planning and optimization issues in monitoring forest areas using UAVs.
- Enhancing the reliability of UAV control through the development of an intelligent interface with speech and gesture recognition capabilities.
- Expanding the functions of UAVs through the ability to interact with multiple devices in a group.
6. Conclusions
- The conducted research allowed to obtain a solution to two of the most important tasks of improving the video stream from UAVs: Removing blurs and stabilizing the video. The Stripformer neural network removes blurriness frame by frame in a video stream captured by a moving camera, and then the processed data are passed into the buffer of the stabilization module based on the VidGear library, which stabilizes the video stream. The final solution allows for the full cycle of video stream processing with a delay of no more than 0.03 s due to the use of a pipeline method for data processing. As a result, we obtain an enhanced image in real-time mode, which can be utilized for further research.
- A new approach for detecting smoke and fires is presented, applicable to challenging weather conditions and various types of landscapes. The approach is based on an algorithm that segments a portion of the image frame along the horizon line and applies a classifier in the form of a generalized Euclidean–Mahalanobis metric. This classifier relies on extracting a large number of fire and smoke features from the video stream frames. The calculation of horizon line points is based on determining the local contrast of the image within a sliding window, where the window size and the pixel informativeness threshold are determined for each image. Experimental results show that the main advantage of this approach is the ability to distinguish smoky areas from cloud-covered areas, which simplifies the task of early fire detection and, in some cases, leads to an improvement in accuracy of approximately 11%.
- An alternative approach to smoke and fire detection was explored, which does not require explicit feature extraction. The approach is based on analyzing different variants of CNN and fine-tuning them for smoke and fire classification. In the task of fire hotspot detection, versions of YOLO (ver. 4 to 8) were compared. The best results were achieved with YOLO 5m, which yielded an F1 score of 76.75% combined with a processing speed of 45 frames per second, i.e., in real time. In the segmentation task, the neural networks Linknet and U-Net with the ResNet18 backbone showed the best performance (by F1 score). During the experiments, some shortcomings and inaccuracies found in the current implementations of YOLO CNN were identified and addressed.
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Flyox I Amphibious UAV—Aerospace Technology. Available online: https://www.aerospace-technology.com/projects/flyox-i-amphibious-uav/ (accessed on 24 January 2024).
- Unmanned Aerial Systems (UAS) as Firefighters | Commercial UAV News. Available online: https://www.commercialuavnews.com/public-safety/unmanned-aerial-systems-uas-firefighters (accessed on 24 January 2024).
- YOLO: Real-Time Object Detection. Available online: https://pjreddie.com/darknet/yolo/ (accessed on 24 January 2024).
- Khachumov, M.V. Distances, Metrics and Cluster Analysis. Sci. Tech. Inf. Process. 2012, 39, 310–316. [Google Scholar] [CrossRef]
- Islam, A.M.; Masud, F.B.; Ahmed, M.R.; Jafar, A.I.; Ullah, J.R.; Islam, S.; Shatabda, S.; Islam, A.K.M.M. An Attention-Guided Deep-Learning-Based Network with Bayesian Optimization for Forest Fire Classification and Localization. Forests 2023, 14, 2080. [Google Scholar] [CrossRef]
- Parthipan, V.; Dhanasekaran, D. Preventing and Monitoring of Framework for Forest Fire Detection and Data Analysis Using Internet of Things (IoT). Int. J. Eng. Adv. Technol. 2019, 8, 691–695. [Google Scholar]
- Brito, T.; Azevedo, B.F.; Mendes, J.; Zorawski, M.; Fernandes, F.P.; Pereira, A.I.; Rufino, J.; Lima, J.; Costa, P. Data Acquisition Filtering Focused on Optimizing Transmission in a LoRaWAN Network Applied to the WSN Forest Monitoring System. Sensors 2023, 23, 1282. [Google Scholar] [CrossRef]
- El-Madafri, I.; Peña, M.; Olmedo-Torre, N. Dual-Dataset Deep Learning for Improved Forest Fire Detection: A Novel Hierarchical Domain-Adaptive Learning Approach. Mathematics 2024, 12, 534. [Google Scholar] [CrossRef]
- Goyal, S.; Vohra, H.; Singh, A.; Kaur, A. A YOLO based Technique for Early Forest Fire Detection. Int. J. Innov. Technol. Explor. Eng. 2020, 9, 1357–1362. [Google Scholar] [CrossRef]
- Wang, Z.; Zhang, H.; Hou, M.; Shu, X.; Wu, J.; Zhang, X. A Study on Forest Flame Recognition of UAV Based on YOLO-V3 Improved Algorithm. In Recent Advances in Sustainable Energy and Intelligent Systems; Li, K., Coombs, T., He, J., Tian, Y., Niu, Q., Yang, Z., Eds.; Communications in Computer and Information Science; Springer: Singapore, 2021; Volume 1468, pp. 497–503. ISBN 9789811672095. [Google Scholar]
- Kim, S.-Y.; Muminov, A. Forest Fire Smoke Detection Based on Deep Learning Approaches and Unmanned Aerial Vehicle Images. Sensors 2023, 23, 5702. [Google Scholar] [CrossRef] [PubMed]
- Cheng, Y.; Chen, K.; Bai, H.; Mou, C.; Zhang, Y.; Yang, K.; Gao, Y.; Liu, Y. An Efficient Fire Detection Algorithm Based on Multi-scale Convolutional Neural Network. Fire Mater. 2022, 46, 981–992. [Google Scholar] [CrossRef]
- Pradeep Kumar, G.; Rahul, R.; Ravindharan, N. Early Forest Fire Detection Using Machine Learning Algorithms. Int. J. New Technol. Res. 2021, 7, 1–5. [Google Scholar]
- Talaat, F.M.; ZainEldin, H. An Improved Fire Detection Approach Based on YOLO-v8 for Smart Cities. Neural Comput. Appl. 2023, 35, 20939–20954. [Google Scholar] [CrossRef]
- Cruz, H.; Gualotuña, T.; Pinillos, M.; Marcillo, D.; Jácome, S.; Fonseca, C.E.R. Machine Learning and Color Treatment for the Forest Fire and Smoke Detection Systems and Algorithms, a Recent Literature Review. In Artificial Intelligence, Computer and Software Engineering Advances; Botto-Tobar, M., Cruz, H., Díaz Cadena, A., Eds.; Advances in Intelligent Systems and Computing; Springer International Publishing: Cham, Switzerland, 2021; Volume 1326, pp. 109–120. ISBN 978-3-030-68079-4. [Google Scholar]
- Yu, Y.; Yao, M. When Convolutional Neural Networks Meet Laser-Induced Breakdown Spectroscopy: End-to-End Quantitative Analysis Modeling of ChemCam Spectral Data for Major Elements Based on Ensemble Convolutional Neural Networks. Remote Sens. 2023, 15, 3422. [Google Scholar] [CrossRef]
- Geetha, S.; Abhishek, C.S.; Akshayanat, C.S. Machine Vision Based Fire Detection Techniques: A Survey. Fire Technol. 2021, 57, 591–623. [Google Scholar] [CrossRef]
- Ciprián-Sánchez, J.F.; Ochoa-Ruiz, G.; Rossi, L.; Morandini, F. Assessing the Impact of the Loss Function, Architecture and Image Type for Deep Learning-Based Wildfire Segmentation. Appl. Sci. 2021, 11, 7046. [Google Scholar] [CrossRef]
- Favorskaya, M.N. Early Smoke Detection in Outdoor Space: State-of-the-Art, Challenges and Methods. In Advances in Selected Artificial Intelligence Areas; Virvou, M., Tsihrintzis, G.A., Jain, L.C., Eds.; Learning and Analytics in Intelligent Systems; Springer International Publishing: Cham, Switzerland, 2022; Volume 24, pp. 171–208. ISBN 978-3-030-93051-6. [Google Scholar]
- Huo, Y.; Zhang, Q.; Jia, Y.; Liu, D.; Guan, J.; Lin, G.; Zhang, Y. A Deep Separable Convolutional Neural Network for Multiscale Image-Based Smoke Detection. Fire Technol. 2022, 58, 1445–1468. [Google Scholar] [CrossRef]
- Miao, J.; Zhao, G.; Gao, Y.; Wen, Y. Fire Detection Algorithm Based on Improved YOLOv5. In Proceedings of the 2021 International Conference on Control, Automation and Information Sciences (ICCAIS), Xi’an, China, 14–17 October 2021; pp. 776–781. [Google Scholar] [CrossRef]
- Li, Y.; Zhang, W.; Liu, Y.; Jin, Y. A Visualized Fire Detection Method Based on Convolutional Neural Network beyond Anchor. Appl. Intell. 2022, 52, 13280–13295. [Google Scholar] [CrossRef]
- Wang, S.; Zhao, J.; Ta, N.; Zhao, X.; Xiao, M.; Wei, H. A Real-Time Deep Learning Forest Fire Monitoring Algorithm Based on an Improved Pruned + KD Model. J. Real-Time Image Process. 2021, 18, 2319–2329. [Google Scholar] [CrossRef]
- Wang, S.; Chen, T.; Lv, X.; Zhao, J.; Zou, X.; Zhao, X.; Xiao, M.; Wei, H. Forest Fire Detection Based on Lightweight Yolo. In Proceedings of the 2021 33rd Chinese Control and Decision Conference (CCDC), Kunming, China, 22–24 May 2021; pp. 1560–1565. [Google Scholar] [CrossRef]
- Nebaba, S.G. In Methods of evaluation and preparation of images in the video stream for object recognition. In Proceedings of the 28th International Conference on Computer Graphics and Vision “GraphiCon 2018”, Tomsk, Russia, 24–27 September 2018; pp. 450–453. (In Russian). [Google Scholar]
- Sherstjuk, V.; Zharikova, M.; Dorovskaja, I.; Sheketa, V. Assessing Forest Fire Dynamicsin UAV-Based Tactical Monitoring System. In Lecture Notes in Computational Intelligence and Decision Making; Babichev, S., Lytvynenko, V., Wójcik, W., Vyshemyrskaya, S., Eds.; Advances in Intelligent Systems and Computing; Springer International Publishing: Cham, Switzerland, 2021; Volume 1246, pp. 285–301. ISBN 978-3-030-54214-6. [Google Scholar] [CrossRef]
- Chernov, T.S. Mathematical Models and Algorithms for Image Quality Assessment in Optical Recognition Systems. Ph.D. Thesis, FRCCSC, Moscow, Russia, 2018. (In Russian). [Google Scholar]
- Mohammadi, P.; Ebrahimi-Moghadam, A.; Shirani, S. Subjective and Objective Quality Assessment of Image: A Survey. arXiv 2014, arXiv:1406.7799. [Google Scholar] [CrossRef]
- Mittal, A.; Moorthy, A.K.; Bovik, A.C. No-Reference Image Quality Assessment in the Spatial Domain. IEEE Trans. Image Process. 2012, 21, 4695–4708. [Google Scholar] [CrossRef]
- Xue, W.; Mou, X.; Zhang, L.; Bovik, A.C.; Feng, X. Blind Image Quality Assessment Using Joint Statistics of Gradient Magnitude and Laplacian Features. IEEE Trans. Image Process. 2014, 23, 4850–4862. [Google Scholar] [CrossRef]
- Zhang, L.; Zhang, L.; Bovik, A.C. A Feature-Enriched Completely Blind Image Quality Evaluator. IEEE Trans. Image Process. 2015, 24, 2579–2591. [Google Scholar] [CrossRef]
- Korovin, I.S.; Khisamutdinov, M.V. Method of Noise-Free Image Production Based on Video Sequence Handling. AASRI Procedia 2014, 6, 73–81. [Google Scholar] [CrossRef]
- Khisamutdinov, M.; Kalyaev, A.; Korovin, I. New Method of Improving the Quality of Single Images in a Video Sequence. DEStech Trans. Comput. Sci. Eng. 2018, 215, 744. [Google Scholar] [CrossRef] [PubMed]
- Zotin, A.G.; Pakhirka, A.I.; Buryachenko, V.V. Development of video surveillance system with visual quality enhancement. Softw. Syst. 2013, 2, 191–197. (In Russian) [Google Scholar]
- Liu, Y.-J.; Chiu, C.-C.; Yang, J.-H. A Robust Vision-Based Skyline Detection Algorithm under Different Weather Conditions. IEEE Access 2017, 5, 22992–23009. [Google Scholar] [CrossRef]
- Guo, F.; Mai, Y.; Tang, J.; Huang, Y.; Zhu, L. Robust and Automatic Skyline Detection Algorithm Based on MSSDN. J. Adv. Comput. Intell. Intell. Inform. 2020, 24, 750–762. [Google Scholar] [CrossRef]
- Martinez-Sanchez, L.; Borio, D.; d’Andrimont, R.; Van Der Velde, M. Skyline Variations Allow Estimating Distance to Trees on Landscape Photos Using Semantic Segmentation. Ecol. Inform. 2022, 70, 101757. [Google Scholar] [CrossRef]
- Ahmad, T.; Emami, E.; Čadík, M.; Bebis, G. Resource Efficient Mountainous Skyline Extraction Using Shallow Learning. In Proceedings of the 2021 International Joint Conference on Neural Networks (IJCNN), Shenzhen, China, 18–22 July 2021. [Google Scholar] [CrossRef]
- Lin, C.; Chen, W.; Zhou, H. Multi-Visual Feature Saliency Detection for Sea-Surface Targets through Improved Sea-Sky-Line Detection. J. Mar. Sci. Eng. 2020, 8, 799. [Google Scholar] [CrossRef]
- Khachumov, V.M.; Portnov, E.M.; Fedorov, P.A.; Kasimov, R.A.; Naing Linn, A. Development of an Accelerated Method for Calculating Streaming Video Data Obtained from UAVs. In Proceedings of the 2020 8th International Conference on Control, Mechatronics and Automation (ICCMA), Moscow, Russia, 6–8 November 2020; pp. 212–216. [Google Scholar] [CrossRef]
- Zhu, D.; Wan, L.; Gao, W. Fusion Method Evaluation and Classification Suitability Study of Wetland Satellite Imagery. Earth Sci. Res. J. 2019, 23, 339–346. [Google Scholar] [CrossRef]
- Vichevskaya, J.A.; Murynov, A.I. Structural analysis of images based on the use of the informative function. Alm. Sovrem. Nauk. I Obraz. 2010, 4, 53–55. (In Russian) [Google Scholar]
- Tymchuk, A.I. On the choice of gray levels in the problem of texture segmentation of images based on the luminance dependence matrices. Cybern. Program. 2018, 3, 1–9. (In Russian) [Google Scholar] [CrossRef]
- Zhurbin, I.; Bazhenova, A.; Shaura, A.; Zlobina, A. Determining the Sliding Window Size and the Optimal Number of Clusters in the Problem of Texture Segmentation of Multispectral Aerial Photography Data. HFIM 2020, 4, 434–447. (In Russian) [Google Scholar] [CrossRef]
- Emelyanova, Y.G. Algorithm for finding the horizon line in images taken from an unmanned aerial vehicle camera. Aerosp. Instrum. Mak. 2023, 1, 40–53. (In Russian) [Google Scholar] [CrossRef]
- Bakir, H.; Bakir, R. Evaluating the robustness of yolo object detection algorithm in terms of detecting objects in noisy environment. J. Sci. Rep. A 2023, 54, 1–25. [Google Scholar] [CrossRef]
- GoPro DataSet. Available online: https://paperswithcode.com/dataset/gopro (accessed on 24 January 2024).
- Awesome-Deblurring. Available online: https://github.com/subeeshvasu/Awesome-Deblurring (accessed on 24 January 2024).
- Chu, X.; Chen, L.; Chen, C.; Lu, X. Improving Image Restoration by Revisiting Global Information Aggregation. In Computer Vision—ECCV 2022; Avidan, S., Brostow, G., Cissé, M., Farinella, G.M., Hassner, T., Eds.; Lecture Notes in Computer Science; Springer Nature: Cham, Switzerland, 2022; Volume 13667, pp. 53–71. ISBN 978-3-031-20070-0. [Google Scholar]
- Tu, Z.; Talebi, H.; Zhang, H.; Yang, F.; Milanfar, P.; Bovik, A.; Li, Y. MAXIM: Multi-Axis MLP for Image Processing. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 5759–5770. [Google Scholar] [CrossRef]
- Sun, L.; Sakaridis, C.; Liang, J.; Jiang, Q.; Yang, K.; Sun, P.; Ye, Y.; Wang, K.; Gool, L.V. Event-Based Fusion for Motion Deblurring with Cross-Modal Attention. In Computer Vision—ECCV 2022; Avidan, S., Brostow, G., Cissé, M., Farinella, G.M., Hassner, T., Eds.; Lecture Notes in Computer Science; Springer Nature: Cham, Switzerland, 2022; Volume 13678, pp. 412–428. ISBN 978-3-031-19796-3. [Google Scholar]
- Li, D.; Zhang, Y.; Cheung, K.C.; Wang, X.; Qin, H.; Li, H. Learning Degradation Representations for Image Deblurring. In Computer Vision—ECCV 2022; Avidan, S., Brostow, G., Cissé, M., Farinella, G.M., Hassner, T., Eds.; Lecture Notes in Computer Science; Springer Nature: Cham, Switzerland, 2022; Volume 13678, pp. 736–753. ISBN 978-3-031-19796-3. [Google Scholar]
- Mou, C.; Wang, Q.; Zhang, J. Deep Generalized Unfolding Networks for Image Restoration. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022. [Google Scholar] [CrossRef]
- He, W.; Yao, Q.; Yokoya, N.; Uezato, T.; Zhang, H.; Zhang, L. Spectrum-Aware and Transferable Architecture Search for Hyperspectral Image Restoration. In Computer Vision—ECCV 2022; Avidan, S., Brostow, G., Cissé, M., Farinella, G.M., Hassner, T., Eds.; Lecture Notes in Computer Science; Springer Nature: Cham, Switzerland, 2022; Volume 13679, pp. 19–37. ISBN 978-3-031-19799-4. [Google Scholar]
- Tsai, F.-J.; Peng, Y.-T.; Lin, Y.-Y.; Tsai, C.-C.; Lin, C.-W. Stripformer: Strip Transformer for Fast Image Deblurring. In Computer Vision—ECCV 2022; Avidan, S., Brostow, G., Cissé, M., Farinella, G.M., Hassner, T., Eds.; Lecture Notes in Computer Science; Springer Nature: Cham, Switzerland, 2022; Volume 13679, pp. 146–162. ISBN 978-3-031-19799-4. [Google Scholar]
- Wang, Z.; Cun, X.; Bao, J.; Zhou, W.; Liu, J.; Li, H. Uformer: A General U-Shaped Transformer for Image Restoration. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 17662–17672. [Google Scholar] [CrossRef]
- STDAN. Available online: https://github.com/huicongzhang/STDAN (accessed on 24 January 2024).
- MMP RNN. Available online: https://github.com/sollynoay/MMP-RNN (accessed on 24 January 2024).
- Qian, Y.; Barthelemy, J.; Iqbal, U.; Perez, P. V2ReID: Vision-Outlooker-Based Vehicle Re-Identification. Sensors 2022, 22, 8651. [Google Scholar] [CrossRef]
- abhiTronix/Vidgear: VidGear Stable v0.2.6 2022. Available online: https://zenodo.org/records/6926196 (accessed on 12 March 2024).
- Liu, Y.-L.; Lai, W.-S.; Yang, M.-H.; Chuang, Y.-Y.; Huang, J.-B. Hybrid Neural Fusion for Full-Frame Video Stabilization. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021. [Google Scholar] [CrossRef]
- Liu, S.; Tan, P.; Yuan, L.; Sun, J.; Zeng, B. MeshFlow: Minimum Latency Online Video Stabilization. In Computer Vision—ECCV 2016; Leibe, B., Matas, J., Sebe, N., Welling, M., Eds.; Lecture Notes in Computer Science; Springer International Publishing: Cham, Switzerland, 2016; Volume 9910, pp. 800–815. ISBN 978-3-319-46465-7. [Google Scholar]
- Video Stabilization with L1 Optimal Camera Paths. Available online: https://github.com/ishank-juneja/L1-optimal-paths-Stabilization (accessed on 24 January 2024).
- Grundmann, M.; Kwatra, V.; Essa, I. Auto-Directed Video Stabilization with Robust L1 Optimal Camera Paths. In Proceedings of the CVPR 2011, Colorado Springs, CO, USA, 20–25 June 2011; pp. 225–232. [Google Scholar] [CrossRef]
- Video Stabilization Using Homography Transform. Available online: https://github.com/SergejVolkov/video_smoothing (accessed on 24 January 2024).
- Xu, Y.; Zhang, J.; Maybank, S.J.; Tao, D. DUT: Learning Video Stabilization by Simply Watching Unstable Videos. IEEE Trans. Image Process. 2022, 31, 4306–4320. [Google Scholar] [CrossRef]
- Choi, J.; Kweon, I.S. Deep Iterative Frame Interpolation for Full-Frame Video Stabilization. ACM Trans. Graph. 2020, 39, 1–9. [Google Scholar] [CrossRef]
- Wang, M.; Yang, G.-Y.; Lin, J.-K.; Zhang, S.-H.; Shamir, A.; Lu, S.-P.; Hu, S.-M. Deep Online Video Stabilization with Multi-Grid Warping Transformation Learning. IEEE Trans. Image Process. 2019, 28, 2283–2292. [Google Scholar] [CrossRef]
- Abhishek, S.T. Video Stabilization Using Point Feature Matching in OpenCV. Available online: https://learnopencv.com/video-stabilization-using-point-feature-matching-in-opencv/ (accessed on 24 January 2024).
- Kvyetnyy, R.; Sofina, O.; Olesenko, A.; Komada, P.; Sikora, J.; Kalizhanova, A.; Smailova, S. Method of Image Texture Segmentation Using Laws’ Energy Measures. In Proceedings of the Photonics Applications in Astronomy, Communications, Industry, and High-Energy Physics, Wilga, Poland, 7 August 2017; Romaniuk, R.S., Linczuk, M., Eds.; SPIE: Bellingham, WA, USA, 2017; p. 1044561. [Google Scholar]
- Mukhiddinov, M.; Abdusalomov, A.B.; Cho, J. A Wildfire Smoke Detection System Using Unmanned Aerial Vehicle Images Based on the Optimized YOLOv5. Sensors 2022, 22, 9384. [Google Scholar] [CrossRef] [PubMed]
- Shamsoshoara, A.; Afghah, F.; Razi, A.; Zheng, L.; Fulé, P. The Flame Dataset: Aerial Imagery Pile Burn Detection Using Drones (UAVS). 2021. Available online: https://ieee-dataport.org/open-access/flame-dataset-aerial-imagery-pile-burn-detection-using-drones-uavs (accessed on 24 January 2024).
- Alarmod/Forest_Fire Datasets at Hugging Face. Available online: https://huggingface.co/datasets/alarmod/forest_fire (accessed on 24 February 2024).
- Models Supported by Ultralytics. Available online: https://docs.ultralytics.com/models (accessed on 24 February 2024).
- Loshchilov, I.; Hutter, F. Decoupled Weight Decay Regularization. arXiv 2017, arXiv:1711.05101. [Google Scholar] [CrossRef]
- Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Medical Image Computing and Computer-Assisted Intervention—MICCAI 2015, Munich, Germany, 5–9 October 2015; Navab, N., Hornegger, J., Wells, W.M., Frangi, A.F., Eds.; Lecture Notes in Computer Science; Springer International Publishing: Cham, Switzerland, 2015; Volume 9351, pp. 234–241. ISBN 978-3-319-24573-7. [Google Scholar]
- Yesilkaynak, V.B.; Sahin, Y.H.; Unal, G. EfficientSeg: An Efficient Semantic Segmentation Network. arXiv 2020, arXiv:2009.06469. [Google Scholar] [CrossRef]
- Ghali, R.; Akhloufi, M.A.; Mseddi, W.S. Deep Learning and Transformer Approaches for UAV-Based Wildfire Detection and Segmentation. Sensors 2022, 22, 1977. [Google Scholar] [CrossRef]
- Cao, H.; Wang, Y.; Chen, J.; Jiang, D.; Zhang, X.; Tian, Q.; Wang, M. Swin-Unet: Unet-like Pure Transformer for Medical Image Segmentation. In European Conference on Computer Vision; Springer Nature: Cham, Switzerland, 2021. [Google Scholar]
- Baheti, B.; Innani, S.; Gajre, S.; Talbar, S. Eff-UNet: A Novel Architecture for Semantic Segmentation in Unstructured Environment. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Seattle, WA, USA, 14–19 June 2020; pp. 1473–1481. [Google Scholar] [CrossRef]
- Chaurasia, A.; Culurciello, E. LinkNet: Exploiting Encoder Representations for Efficient Semantic Segmentation. In Proceedings of the 2017 IEEE Visual Communications and Image Processing (VCIP), St. Petersburg, FL, USA, 10–13 December 2017; pp. 1–4. [Google Scholar] [CrossRef]
- Lv, W.; Zhao, Y.; Xu, S.; Wei, J.; Wang, G.; Cui, C.; Du, Y.; Dang, Q.; Liu, Y. DETRs Beat YOLOs on Real-Time Object Detection. arXiv 2023, arXiv:2304.08069. [Google Scholar] [CrossRef]
- Wang, C.-Y.; Yeh, I.-H.; Liao, H.-Y.M. YOLOv9: Learning What You Want to Learn Using Programmable Gradient Information. arXiv 2024, arXiv:2402.13616. [Google Scholar] [CrossRef]
Method | Processing Time, s | Fragment 1 | Fragment 2 | ||
---|---|---|---|---|---|
PSNR | SSIM | PSNR | SSIM | ||
Test-time Local Converter (TLC) [49] | 13.5 | 33.012 | 0.939 | 31.186 | 0.929 |
Multi-Axis MLP [50] | 19.5 | 30.556 | 0.917 | 31.748 | 0.934 |
EFNet [51] | 1.10 | 35.184 1 | 0.948 1 | 35.695 1 | 0.962 1 |
Learning degradation [52] | 3.60 | 32.549 | 0.932 | 31.739 | 0.934 |
Deep Generalized Unfolding [53] | 5.80 | 22.582 | 0.735 | 29.923 | 0.903 |
NAFNet width32 [54] | 5.00 | 31.789 | 0.918 | 30.000 | 0.916 |
Stripformer [55] | 0.03 2 | 28.060 | 0.876 | 30.222 | 0.916 |
Uformer [56] | 9.40 | 32.296 | 0.928 | 31.477 | 0.929 |
STDAN [57] | 360 | 31.719 | 0.926 | 30.168 | 0.913 |
MMP-RNN [58] | 108 | 26.668 | 0.844 | 30.413 | 0.910 |
Method | Processing Time, s | Note |
---|---|---|
VidGear [60] | 3.91 | The frame borders are cropped using a border offset and scaling. Initialization of the stabilizer is required, with input data of about 24 frames used. The time required to prepare the stabilizer for operation is approximately 0.70 s. 1 |
FuSta: Hybrid Neural Fusion [61] | 1200 | A graphics accelerator is used. The resulting video from these final frames is smooth, but there are artifacts at the boundaries of the “car” and “tree” objects. |
MeshFlow: Minimum Latency Online Video Stabilization [62] | 1260 | Experiments were conducted with two settings profiles. The Original profile provides excellent quality without borders, and the frame is automatically cropped. The Constant High profile has artifacts in the form of unnaturally tilted objects, such as cars. |
Video Stabilization with L1 optimal camera paths [63] | 0.50 | The resulting video is “wavy”. The camera sometimes moves towards objects in the frame and then back. |
Auto-Directed Video Stabilization with Robust L1 Optimal Camera Paths [64] | 120 | The resulting image is significantly cropped when high levels of stabilization are used. For example, at a threshold of 40 pixels, an image of 1280 × 720 is cropped at the edges to 1180 × 620. Using a lower threshold gives an insufficient level of video stabilization. Processing is divided into two stages: 2 min for preprocessing, followed by fast stabilization using the obtained data. |
Video stabilization using homography transform [65] | 4.18 | The resulting image is significantly cropped, especially at high stabilization coefficients. The best results are obtained with real-time stabilization when the parameter is set to 20. |
DUT: Learning Video Stabilization [66] | 32.0 | A graphics accelerator is used. The method relies on several pre-trained neural networks. The approach is based on self-learning on unstable video data. The car image in the resulting video is slightly offset left and right, but the overall resulting video can be considered of good quality. |
Deep Iterative Frame Interpolation for Full-frame Video Stabilization [67] | 100 | A graphics accelerator is used. Processing is performed iteratively, with the number of iterations controlled by the n_iter parameter, which is set to 3 by default. The resulting video stream is unstable, and the overall quality of the resulting video stream is low. |
PWStableNet [68] | 8.00 | A graphics accelerator is used. The resulting video has jerky movements. |
ANN Version | Precision, % | Recall, % | F1 Score, % | Confidence Threshold, % | [email protected], % | [email protected]:.95, % | Processing Speed, Frames per Second |
---|---|---|---|---|---|---|---|
4 csp mish | 78.2 | 70.4 | 74.1 | 51.0 | 74.5 | 39.2 | 55 |
5s | 79.4 | 71.1 | 75.1 | 51.5 | 76.7 | 40.7 | 49 |
5m with SiLU activation | 79.3 | 74.3 2 | 76.8 3 | 44.4 | 77. 9 4 | 41.3 | 45 |
5m with Mish activation | 80.2 1 | 72.1 | 75.9 | 49.9 | 77.1 | 41.1 | 43 |
5l | 79.5 | 72.3 | 75.7 | 49.7 | 76.8 | 41.8 5 | 35 |
6t | 73.7 | 72.3 | 73.0 | 50.0 | 73.7 | 38.3 | 122 6 |
7x | 73.4 | 69.3 | 71.3 | 23.6 | 72.8 | 35.4 | 39 |
8l | 77.3 | 70.3 | 73.0 | 30.1 | 76.1 | 36.4 | 39 |
8l Ghost | 78.0 | 72.1 | 75.0 | 26.4 | 77.1 | 41.6 | 49 |
ANN Used | Normalized F1 Score, % | Frame Size, px × px | Processing Speed, Frames per Second |
---|---|---|---|
U-Net [76] with one convolution layer added to the coder and decoder | 97.48 | 512 × 512 | 29.09 |
U-Net with VGG16 as a backbone | 99.35 | 512 × 512 | 12.17 |
U-Net with ResNet18 as a backbone | 99.36 1 | 512 × 512 | 37.01 |
EfficientSeg [77] | 97.47 | 512 × 512 | 54.99 |
TransUnet [78] with 6 vit blocks and 6 vit-headers, the mhsa block size is 512, the transformer size is 128 | 97.40 | 512 × 512 | 18.78 |
SwinUnet [79] | 91.81 | 448 × 448 | 26.82 |
EffuNet [80] | 97.54 | 512 × 512 | 56.95 2 |
Linknet [81] with VGG16 as a backbone | 99.30 | 512 × 512 | 11.62 |
Linknet with ResNet18 as a backbone | 99.31 | 512 × 512 | 38.08 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Abramov, N.; Emelyanova, Y.; Fralenko, V.; Khachumov, V.; Khachumov, M.; Shustova, M.; Talalaev, A. Intelligent Methods for Forest Fire Detection Using Unmanned Aerial Vehicles. Fire 2024, 7, 89. https://doi.org/10.3390/fire7030089
Abramov N, Emelyanova Y, Fralenko V, Khachumov V, Khachumov M, Shustova M, Talalaev A. Intelligent Methods for Forest Fire Detection Using Unmanned Aerial Vehicles. Fire. 2024; 7(3):89. https://doi.org/10.3390/fire7030089
Chicago/Turabian StyleAbramov, Nikolay, Yulia Emelyanova, Vitaly Fralenko, Vyacheslav Khachumov, Mikhail Khachumov, Maria Shustova, and Alexander Talalaev. 2024. "Intelligent Methods for Forest Fire Detection Using Unmanned Aerial Vehicles" Fire 7, no. 3: 89. https://doi.org/10.3390/fire7030089
APA StyleAbramov, N., Emelyanova, Y., Fralenko, V., Khachumov, V., Khachumov, M., Shustova, M., & Talalaev, A. (2024). Intelligent Methods for Forest Fire Detection Using Unmanned Aerial Vehicles. Fire, 7(3), 89. https://doi.org/10.3390/fire7030089