A Detection Method for Individual Infected Pine Trees with Pine Wilt Disease Based on Deep Learning
Abstract
:1. Introduction
2. Materials and Methods
2.1. Study Area
2.2. UAV-Based Multispectral Data
2.3. Dataset
2.4. Pre-Processing
2.5. YOLOv5
2.6. MFTD Structure
2.7. MFFT Module
- (1)
- A MaxPooling layer was used to reduce the dimensions of the input feature map and , from to , which was concatenated after flattening each feature map to obtain , a feature matrix with a dimension of .
- (2)
- The position encoding () method of Transformer was used to supplement the timing information for . , the input feature sequence of Embedding, can be expressed as:
- (3)
- was input into the Multi-Head Attention (MHA) module of Embedding. The MHA is the concatenating result after multiple Self-Attention parallel calculations, with its structure shown in Figure 10. Self-Attention can be described as a query mechanism, which applies vector Query () to retrieve the target Key () and the corresponding Value () of the Key and takes the weighted sum of all Values as the output. , , and , as the input of MHA, are the matrixes of feature sequence S under the mapping of weights, , , and , which are expressed as:
- (4)
- After and were added, they were normalized through the Normal Layer [36] and then input into the full-connection Feed-Forward Network (FFN), which contains two linear transformations and a Relu activation function. The output F is expressed as:
- (5)
- The output feature sequence of Embedding was obtained after adding and through the Normal Layer. Two feature maps with the dimensions of were obtained by splitting M′ and the reverse operation of flattening in step (1). After upsampling, the visible and multispectral enhancement feature maps, and , were output.
2.8. Evaluation Indicators
3. Results
3.1. Implementation Details
3.2. Evaluation of Network Fusion Modes
3.3. Ablation Experiments
3.4. Experiments of YOLO Series
4. Discussion
5. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Wu, W.; Zhang, Z.; Zheng, L.; Han, C.; Wang, X. Research Progress on the Early Monitoring of Pine Wilt Disease Using Hyperspectral Techniques. Sensors 2020, 20, 3729. [Google Scholar] [CrossRef] [PubMed]
- Ye, J. Epidemic status of pine wilt disease in China and its prevention and control techniques and counter measures. Sci. Silvae Sin. 2019, 55, 10. [Google Scholar]
- Announcement of the State Forestry and Grassland Administration (2022 No. 5) (Pinewood Nematode Epidemic Area in 2022). Available online: http://www.forestry.gov.cn/ (accessed on 6 April 2022).
- Umebayashi, T.; Yamada, T.; Fukuhara, K.; Endo, R.; Kusumoto, D.; Fukuda, K. In situ observation of pinewood nematode in wood. Eur. J. Plant Pathol. 2017, 147, 463–467. [Google Scholar] [CrossRef]
- Kim, S.-R.; Lee, W.-K.; Lim, C.-H.; Kim, M.; Kafatos, M.C.; Lee, S.-H.; Lee, S.S. Hyperspectral analysis of pine wilt disease to determine an optimal detection index. Forests 2018, 9, 115. [Google Scholar] [CrossRef] [Green Version]
- Shin, S.-C. Pine wilt Disease; Zhao, B.G., Futai, K., Sutherland, J.R., Eds.; Springer: Tokyo, Japan, 2008; pp. 26–32. [Google Scholar]
- Pajares, G. Overview and Current Status of Remote Sensing Applications Based on Unmanned Aerial Vehicles (UAVs). Photogramm. Eng. Remote Sens. 2015, 81, 281–329. [Google Scholar] [CrossRef] [Green Version]
- Tang, L.; Shao, G. Drone remote sensing for forestry research and practices. J. For. Res. 2015, 26, 791–797. [Google Scholar] [CrossRef]
- Li, W.; Shen, S.; He, P.; Hao, D.; Fang, Y.; Tao, L.; Zhang, S. A precisely positioning technique by remote sensing the dead trees in stands with inexpensive small UAV. China For. Sci. Technol. 2014, 28, 102–105. [Google Scholar]
- Yuan, Y.; Hu, X. Random forest and objected-based classification for forest pest extraction from UAV aerial imagery. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2016, XLI-B1, 1093–1098. [Google Scholar] [CrossRef] [Green Version]
- Zhong, T.; Liu, W.; Luo, Y.; Hung, C.C. A New Type-2 Fuzzy Algorithm for Unmanned Aerial Vehicle Image Segmentation. Int. J. Multimed. Ubiquitous Eng. 2017, 12, 75–90. [Google Scholar] [CrossRef]
- Takenaka, Y.; Katoh, M.; Deng, S.; Cheung, K. 25–27 October 2017 Detecting forests damaged by pine wilt disease at the individual tree level using airborne laser data and worldview-2/3 images over two seasons. In The ISPRS-International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences; Shinshu University Library: Jyväskylä, Finland, 2017; pp. 181–184. [Google Scholar]
- Hu, G.; Yin, C.; Wan, M. Recognition of diseased Pinus trees in UAV images using deep learning and AdaBoost classifier. Biosyst. Eng. 2020, 194, 138–151. [Google Scholar] [CrossRef]
- Deng, X.; Tong, Z.; Lan, Y.; Huang, Z. Detection and Location of Dead Trees with Pine Wilt Disease Based on Deep Learning and UAV Remote Sensing. AgriEngineering 2020, 2, 294–307. [Google Scholar] [CrossRef]
- Tao, H.; Li, C.; Zhao, D.; Deng, S.; Hu, H.; Xu, X.; Jing, W. Deep learning-based dead pine tree detection from unmanned aerial vehicle images. Int. J. Remote Sens. 2020, 41, 8238–8255. [Google Scholar] [CrossRef]
- Wu, H. A study of the potential of using worldview-2 of images for the detection of red attack pine tree. In Proceedings of the Eighth International Conference on Digital Image Processing (ICDIP 2016), Chengdu, China, 20–22 May 2016. [Google Scholar]
- Run, Y.; Youqing, L.; Quan, Z.; Xz, A.; Dw, A.; Lra, B. Early detection of pine wilt disease using deep learning algorithms and UAV-based multispectral imagery. For. Ecol. Manag. 2021, 497, 119493. [Google Scholar]
- Shaoqing, R.; Kaiming, H.; Ross, G.; Jian, S. Faster R-CNN: Towards real-time object detection with region proposal networks. In Proceedings of the International Conference on Neural Information Processing Systems (NIPS), Montreal, QC, Canada, 7 December 2015. [Google Scholar]
- Bochkovskiy, A.; Wang, C.-Y.; Liao, H.-Y.M. Yolov4: Optimal speed and accuracy of object detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
- Shi, S.; Guo, C.; Jiang, L.; Wang, Z.; Li, H. PV-RCNN: Point-Voxel Feature Set Abstraction for 3D Object Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 14 June 2020. [Google Scholar]
- Ha, Q.; Watanabe, K.; Karasawa, T.; Ushiku, Y.; Harada, T. MFNet: Towards real-time semantic segmentation for autonomous vehicles with multi-spectral scenes. In Proceedings of the International Conference on Intelligent Robots and Systems (IROS), Vancouver, BC, Canada, 25 September 2017. [Google Scholar]
- Vollenweider, P.; Günthardt-Goerg, M. Diagnosis of abiotic and biotic stress factors using the visible symptoms in foliage. Environ. Pollut. 2006, 140, 562–571. [Google Scholar] [CrossRef] [PubMed]
- Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27 June 2016. [Google Scholar]
- Redmon, J.; Farhadi, A. YOLO9000: Better, faster, stronger. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21 July 2017. [Google Scholar]
- Redmon, J.; Farhadi, A. YOLOv3: An Incremental Improvement. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18 June 2018. [Google Scholar]
- Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. In Proceedings of the Annual Conference on Neural Information Processing Systems (NIPS), Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
- Carion, N.; Massa, F.; Synnaeve, G.; Usunier, N.; Kirillov, A.; Zagoruyko, S. End-to-end object detection with transformers. In Proceedings of the European Conference on Computer Vision (ECCV), Online, 23 August 2020. [Google Scholar]
- Song, H.; Sun, D.; Chun, S.; Jampani, V.; Han, D.; Heo, B.; Kim, W.; Yang, M.-H. ViDT: An Efficient and Effective Fully Transformer-based Object Detector. In Proceedings of the International Conference on Learning Representations (ICLR), Online, 3–7 May 2021. [Google Scholar]
- Jocher, G. Ultralytics-YOLOv5. Available online: https://github.com/ultralytics/yolov5 (accessed on 4 October 2022).
- Tolias, G.; Sicre, R.; Jégou, H. Particular object retrieval with integral max-pooling of CNN activations. In Proceedings of the International Conference on Learning Representations (ICLR), San Juan, Puerto Rico, 2–4 May 2016. [Google Scholar]
- Lin, T.Y.; Dollar, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature Pyramid Networks for Object Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21 July 2017. [Google Scholar]
- Li, H.; Xiong, P.; An, J.; Wang, L. Pyramid Attention Network for Semantic Segmentation. In Proceedings of the British Machine Vision Conference, Newcastle, UK, 3–6 September 2018. [Google Scholar]
- Wang, C.Y.; Liao, H.; Wu, Y.H.; Chen, P.Y.; Yeh, I.H. CSPNet: A New Backbone that can Enhance Learning Capability of CNN. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Seattle, WA, USA, 14 June 2020. [Google Scholar]
- He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27 June 2016. [Google Scholar]
- Liu, J.; Zhang, S.; Wang, S.; Metaxas, D.N. Multispectral Deep Neural Networks for Pedestrian Detection. In Proceedings of the British Machine Vision Conference (BMVC), York, UK, 19–22 September 2016. [Google Scholar]
- Ba, J.L.; Kiros, J.R.; Hinton, G.E. Layer normalization. Stat 2016, 1050, 21. [Google Scholar]
- Lin, T.Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft COCO: Common Objects in Context. In Proceedings of the European Conference on Computer Vision (ECCV), Zurich, Switzerland, 6–12 September 2014. [Google Scholar]
- Loshchilov, I.; Hutter, F. SGDR: Stochastic Gradient Descent with Warm Restarts. In Proceedings of the International Conference on Learning Representations (ICLR), Toulon, France, 24–26 April 2017. [Google Scholar]
- Syifa, M.; Park, S.-J.; Lee, C.-W. Detection of the pine wilt disease tree candidates for drone remote sensing using artificial intelligence techniques. Engineering 2020, 6, 919–926. [Google Scholar] [CrossRef]
- Iordache, M.-D.; Mantas, V.; Baltazar, E.; Pauly, K.; Lewyckyj, N. A machine learning approach to detecting pine wilt disease using airborne spectral imagery. Remote Sens. 2020, 12, 2280. [Google Scholar] [CrossRef]
- Qin, J.; Wang, B.; Wu, Y.; Lu, Q.; Zhu, H. Identifying pine wood nematode disease using UAV images and deep learning algorithms. Remote Sens. 2021, 13, 162. [Google Scholar] [CrossRef]
Items | Parameters | Values |
---|---|---|
DJI Phantom4 multispectral UAV | Length × Width × Height/(mm × mm × mm) | 540 × 595 × 255 |
Sensor | CMOS × 6, 1/2.9″ | |
Light filter | Blue: 450 nm ± 16 nm Green: 560 nm ± 16 nm Red: 650 nm ± 16 nm Red edge: 730 nm ± 16 nm NIR: 840 nm ± 26 nm | |
Image resolution/pixels | 1600 × 1300 |
Study Areas | 1 | 2 | 3 | 4 |
---|---|---|---|---|
Tree species | Chinese pines | Korean pines | Korean pines | Chinese pines |
Date | 24 June 2021 | 13 May 2021 | 10 April 2021 | 15 July 2021 |
Flight height/m | 120 | 120 | 60 | 120 |
Center coordinates | 124°10′39″ E 41°54′53″ N | 124°10′45″ E 41°54′48″ N | 124°14′26″ E 41°55′48″ N | 124°16′44″ E 41°57′33″ N |
Spatial resolution/(cm/pixel) | 0.2–1.5 | 0.2–1.5 | 0.75 | 0.2–1.5 |
Image sets amount | 180 | 150 | 239 | 135 |
Surface area/hm2 | 4.2 | 3.9 | 3.5 | 2.6 |
Infection Stage | Early Stage | Middle Stage | Late Stage | ||
---|---|---|---|---|---|
Description | Slightly wilted crown tops | Part of pine needles wilted and turning yellow or red | The whole crowns wilted and discoloration | Crowns turning reddish brown | Pine needles falling off |
Korean pines | |||||
Chinese pines |
Model (YOLOv5l) | Network Fusion Mode | Image Type | mAP@50 | AP@50 | Parameter Amount/MB | ||
---|---|---|---|---|---|---|---|
Early Stage | Middle Stage | Late Stage | |||||
K0 | \ | RGB | 79.9 | 72.9 | 87.3 | 78.6 | 93.7 |
K1 | \ | R | 77.8 | 70.6 | 81.6 | 81.3 | 93.7 |
K2 | Input Fusion | RGB + R | 81.9 | 74.9 | 87.7 | 83.2 | 118.4 |
K3 | Half Fusion | RGB + R | 83.2 | 77.4 | 89.5 | 82.6 | 148.1 |
K4 | Late Fusion | RGB + R | 82.9 | 75.2 | 89.3 | 84.3 | 187.3 |
Model (YOLOv5l) | Network Fusion Mode | Image Type | mAP@50 | AP@50 | Parameter Amount/MB | ||
---|---|---|---|---|---|---|---|
Early Stage | Middle Stage | Late Stage | |||||
C0 | \ | RGB | 76.0 | 69.6 | 85.5 | 72.8 | 93.7 |
C1 | \ | R | 73.6 | 68.1 | 74.4 | 78.2 | 93.7 |
C2 | Input Fusion | RGB + R | 79.5 | 75.3 | 82.3 | 80.8 | 118.4 |
C3 | Half Fusion | RGB + R | 82.1 | 78.4 | 85.6 | 82.3 | 148.1 |
C4 | Late Fusion | RGB + R | 82.0 | 78.1 | 87.0 | 80.8 | 187.3 |
Model (YOLOv5l) | Data Preprocessing | MFFT | mAP@50 | AP@50 | ||||
---|---|---|---|---|---|---|---|---|
[Frgb1, Fr1] | [Frgb2, Fr2] | [Frgb3, Fr3] | Early Stage | Middle Stage | Late Stage | |||
K3 | √ | 83.2 | 77.4 | 89.5 | 82.6 | |||
K5 | √ | √ | 83.3 | 77.7 | 89.0 | 83.1 | ||
K6 | √ | √ | 84.3 | 78.0 | 89.9 | 84.8 | ||
K7 | √ | √ | 83.5 | 76.1 | 92.1 | 82.2 | ||
K8 | √ | √ | √ | 84.4 | 79.4 | 90.5 | 83.4 | |
K9 | √ | √ | √ | 85.4 | 78.0 | 93.6 | 84.7 | |
K10 | √ | √ | √ | 86.3 | 81.9 | 92.8 | 84.3 | |
K11(MFTD) | √ | √ | √ | √ | 88.5 | 87.2 | 93.5 | 84.8 |
K12 | √ | √ | √ | 83.2 | 82.2 | 86.7 | 80.7 |
Model (YOLOv5l) | Data Preprocessing | MFFT | mAP@50 | AP@50 | ||||
---|---|---|---|---|---|---|---|---|
[Frgb1, Fr1] | [Frgb2, Fr2] | [Frgb3, Fr3] | Early Stage | Middle Stage | Late Stage | |||
C3 | √ | 82.1 | 78.4 | 85.6 | 82.3 | |||
C5 | √ | √ | 83.7 | 80.3 | 87.7 | 83.3 | ||
C6 | √ | √ | 82.6 | 81.1 | 86.3 | 80.5 | ||
C7 | √ | √ | 83.4 | 78.7 | 87.6 | 84.0 | ||
C8 | √ | √ | √ | 84.9 | 82.4 | 88.8 | 83.5 | |
C9 | √ | √ | √ | 84.6 | 79.2 | 91.9 | 82.8 | |
C10 | √ | √ | √ | 85.3 | 81.0 | 90.1 | 84.7 | |
C11(MFTD) | √ | √ | √ | √ | 86.8 | 81.2 | 92.9 | 86.2 |
C12 | √ | √ | √ | 84.9 | 79.2 | 90.4 | 85.0 |
Model | YOLOx | Image Type | mAP@50 | AP@50 | Parameter Amount/MB | FPS/HZ | ||
---|---|---|---|---|---|---|---|---|
Early Stage | Middle Stage | Late Stage | ||||||
K13 | v3 | RGB | 73.9 | 70.1 | 80.7 | 71.1 | 248.2 | 80.3 |
K14 | v4 | RGB | 78.7 | 72.7 | 86.1 | 77.2 | 257.9 | 65.7 |
K15 | v5s | RGB | 77.9 | 71.3 | 87.4 | 75.1 | 14.4 | 146.3 |
K16 | R | 77.8 | 69.5 | 82.8 | 81.1 | 14.4 | 146.3 | |
K17 | v5m | RGB | 79.2 | 73.4 | 86.3 | 77.9 | 42.5 | 97.1 |
K18 | R | 77.6 | 72.7 | 76.1 | 84.0 | 42.5 | 97.1 | |
K0 | v5l | RGB | 79.9 | 72.9 | 87.3 | 78.6 | 93.7 | 86.1 |
K1 | R | 77.8 | 70.6 | 81.6 | 81.3 | 93.7 | 86.1 | |
K19 | v5x | RGB | 79.3 | 75.2 | 86.2 | 76.5 | 175.1 | 58.9 |
K20 | R | 78.4 | 71.0 | 82.4 | 81.8 | 175.1 | 58.9 | |
K21(MFTD) | v5s | RGB + R | 80.8 | 77.1 | 84.6 | 80.7 | 89.6 | 49.2 |
K22(MFTD) | v5m | RGB + R | 85.3 | 84.3 | 90.1 | 81.5 | 216.8 | 40.2 |
K11(MFTD) | v5l | RGB + R | 88.5 | 87.2 | 93.5 | 84.8 | 413.4 | 35.6 |
K23(MFTD) | v5x | RGB + R | 85.9 | 84.7 | 93.1 | 80.0 | 690.1 | 30.4 |
Model | YOLOx | Image Type | mAP@50 | AP@50 | Parameter Amount/MB | FPS/HZ | ||
---|---|---|---|---|---|---|---|---|
Early Stage | Middle Stage | Late Stage | ||||||
C13 | v3 | RGB | 66.9 | 64.3 | 75.9 | 60.5 | 248.2 | 80.3 |
C14 | v4 | RGB | 73.3 | 67.2 | 80.6 | 72.1 | 257.5 | 65.7 |
C15 | v5s | RGB | 68.6 | 65.0 | 75.5 | 65.2 | 14.4 | 146.3 |
C16 | R | 64.5 | 62.3 | 64.9 | 66.3 | 14.4 | 146.3 | |
C17 | v5m | RGB | 72.6 | 68.8 | 82.2 | 66.7 | 42.5 | 97.1 |
C18 | R | 69.2 | 68.0 | 68.6 | 70.9 | 42.5 | 97.1 | |
C0 | v5l | RGB | 76.0 | 69.6 | 85.5 | 72.8 | 93.7 | 86.1 |
C1 | R | 73.6 | 68.1 | 74.4 | 78.2 | 93.7 | 86.1 | |
C19 | v5x | RGB | 75.6 | 70.6 | 86.0 | 70.3 | 175.1 | 58.9 |
C20 | R | 73.3 | 63.9 | 75.6 | 80.5 | 175.1 | 58.9 | |
C21(MFTD) | v5s | RGB + R | 80.2 | 79.7 | 81.6 | 79.3 | 89.6 | 49.2 |
C22(MFTD) | v5m | RGB + R | 82.0 | 80.2 | 84.4 | 81.3 | 216.8 | 40.2 |
C11(MFTD) | v5l | RGB + R | 86.8 | 81.2 | 92.9 | 86.2 | 413.4 | 35.6 |
C23(MFTD) | v5x | RGB + R | 84.7 | 80.8 | 88.5 | 84.6 | 690.1 | 30.4 |
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2022 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Zhou, Y.; Liu, W.; Bi, H.; Chen, R.; Zong, S.; Luo, Y. A Detection Method for Individual Infected Pine Trees with Pine Wilt Disease Based on Deep Learning. Forests 2022, 13, 1880. https://doi.org/10.3390/f13111880
Zhou Y, Liu W, Bi H, Chen R, Zong S, Luo Y. A Detection Method for Individual Infected Pine Trees with Pine Wilt Disease Based on Deep Learning. Forests. 2022; 13(11):1880. https://doi.org/10.3390/f13111880
Chicago/Turabian StyleZhou, Yan, Wenping Liu, Haojie Bi, Riqiang Chen, Shixiang Zong, and Youqing Luo. 2022. "A Detection Method for Individual Infected Pine Trees with Pine Wilt Disease Based on Deep Learning" Forests 13, no. 11: 1880. https://doi.org/10.3390/f13111880
APA StyleZhou, Y., Liu, W., Bi, H., Chen, R., Zong, S., & Luo, Y. (2022). A Detection Method for Individual Infected Pine Trees with Pine Wilt Disease Based on Deep Learning. Forests, 13(11), 1880. https://doi.org/10.3390/f13111880