YOLO-GP: A Multi-Scale Dangerous Behavior Detection Model Based on YOLOv8
Abstract
:1. Introduction
- In this study, an innovative symmetric structure of grouped pointwise convolutional (GPConv) is designed, which enhances the model’s feature representation and expressiveness by integrating feature fusion in the channel dimension and combining various feature extraction methods.
- A dual-branch aggregation module (DAM) is designed to replace the C2f module of the original model to obtain richer gradient flow information, to solve the problem of the baseline model’s poor accuracy in locating dangerous behavioral targets and its poor ability to discriminate between small targets such as smoking and phone usage.
- By fusing the innovative efficient spatial pyramid pooling (ESPP) module to the neck of the model, we can effectively improve the recognition ability of dangerous behaviors through multi-scale feature capture and feature fusion so that the model can accurately understand and differentiate dangerous behavior targets involving different scales.
- To solve the problem of insufficient channel correlation, which leads to the poor performance of the model in detecting dangerous behaviors in complex scenes, a channel feature enhancement network (CFE-Net) is designed to enable the model to better understand the interactions between different channels, to achieve the purpose of improving the accuracy of the model in detecting dangerous behaviors in complex scenes.
2. Related Work
2.1. Traditional Methods for Dangerous Behavior Detection Algorithm
2.2. Dangerous Behavior Detection Based on Deep Learning
3. Material and Methods
3.1. YOLO-GP Algorithm Overview
3.2. Improvement Strategies
3.2.1. Grouped Pointwise Convolutional
3.2.2. Dual-Branch Aggregation Module
3.2.3. Efficient Spatial Pyramid Pooling
3.2.4. Channel Feature Enhancement Network
3.3. Experimental Environment and Parameter Settings
3.4. Datasets
3.5. Evaluation Indicators
4. Experimental Results and Analysis
4.1. Ablation Experiment
- In the DBD dataset, the design of the DAM effectively enhances the Backbone and Neck structures of the YOLOv8n network, aiding in better gradient flow information capture. This helps improve the model’s accuracy in localizing dangerous behavior targets, particularly addressing the challenge of discriminating small targets such as smoking or phone usage. In the CSSID dataset, this improvement also effectively addresses localization issues in complex construction site environments with diverse target scales;
- On both datasets, the ESPP module enhances the model’s ability to capture and fuse multi-scale features. This contributes to better recognition of dangerous behaviors, enabling the model to more accurately understand and differentiate targets involving various scales.
- Integrating the CFE-Net enhances the model’s understanding of inter-channel interactions, improving dangerous behavior detection performance in complex scenarios. In the DBD dataset, this helps address insufficient channel correlations, thereby better understanding the correlation between different dangerous behaviors. The CSSID dataset assists the model in handling various safety equipment, personnel, and objects present in construction site environments, enhancing the model’s robustness and generalization capability.
4.2. Convergence Curve
4.3. Activation Function Parameter Selection Experiment
4.4. Validation of the Validity of the ESPP Module
4.5. Detection Results for Different Categories of Targets
4.6. Visualization Results and Analysis
- Analysis of visualization results under DBD dataset.From Figure 11, it can be observed that the first and second columns of the visualized results exhibit characteristics such as high image grayscale and dim environments, which may adversely affect target detection. Grayscale images can result in blurred or lost target features and reduced contrast, thereby impacting the accuracy and stability of the algorithm. Dim environments may cause unclear target details and increased background noise, making it challenging for the target detection algorithm to correctly identify targets. The third and fourth columns of images both feature complex backgrounds and small targets. In such cases, the complex background may cause confusion between the target and the background, making it difficult for the algorithm to accurately locate and identify the target. Additionally, the presence of small targets may obscure the features of the target in the image, increasing the probability of false positives and false negatives in the detection algorithm. In these scenarios, the YOLO-GP model in the task of detecting dangerous behaviors among workers exhibits a significant reduction in the number of red and blue boxes in its visualizations compared to the baseline model. This reduction indicates a decrease in the probability of false positives and false negatives, thereby enhancing the reliability of detection.
- Analysis of visualization results under the CSSID dataset.In Figure 12, we observe characteristics such as cluttered backgrounds and significant differences in target scales. The cluttered background makes it challenging for the algorithm to distinguish targets from the surrounding environment, increasing the likelihood of false positives, especially when small targets are present. On the other hand, significant differences in target scales may lead to an imbalance in how the algorithm handles targets of different sizes, potentially resulting in detection errors or missing small targets. While any model has certain limitations, as shown in Figure 12, the YOLO-GP model also exhibits some false detections (red boxes) and missed targets (blue boxes). However, compared to the YOLOv8n model, the YOLO-GP model demonstrates better adaptability and robustness, achieving more accurate target detection and maintaining stable performance across different scales and background environments. Therefore, these visualized results further validate the superiority of the YOLO-GP model in tackling complex backgrounds and multi-scale target detection tasks.
4.7. Multi-Model Comparative Experiments
5. Discussion
Author Contributions
Funding
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- Lane, M.K.M.; Rudel, H.E.; Wilson, J.A.; Erythropel, H.C.; Backhaus, A.; Gilcher, E.B.; Ishii, M.; Jean, C.F.; Lin, F.; Muellers, T.D.; et al. Green Chemistry as Just Chemistry. Nat. Sustain. 2023, 6, 502–512. [Google Scholar] [CrossRef]
- Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich Feature Hierarchies for Accurate Object Detection and Semantic Segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Columbus, OH, USA, 23–28 June 2014. [Google Scholar]
- Girshick, R.B. Fast R-CNN. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar]
- Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef] [PubMed]
- Jiang, P.; Ergu, D.; Liu, F.; Cai, Y.; Ma, B. A Review of Yolo Algorithm Developments. Procedia Comput. Sci. 2022, 199, 1066–1073. [Google Scholar] [CrossRef]
- Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.-Y.; Berg, A.C. SSD: Single Shot Multibox Detector. In Proceedings of the 14th European Conference on Computer Vision (ECCV 2016), Amsterdam, The Netherlands, 11–14 October 2016; Part I. pp. 21–37. [Google Scholar]
- Bharati, P.; Pramanik, A. Deep learning techniques—R-CNN to mask R-CNN: A survey. In Computational Intelligence in Pattern Recognition, Proceedings of the CIPR 2019, Howrah, India, 19–20 January 2019; Springer: Singapore, 2020; pp. 657–668. [Google Scholar]
- Cai, Z.; Vasconcelos, N. Cascade R-CNN: High quality object detection and instance segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 43, 1483–1498. [Google Scholar] [CrossRef] [PubMed]
- Qian, X.; Wang, X.; Yang, S.; Lei, J. LFF-YOLO: A YOLO algorithm with lightweight feature fusion network for multi-scale defect detection. IEEE Access 2022, 10, 130339–130349. [Google Scholar] [CrossRef]
- Ju, M.; Luo, J.; Wang, Z.; Luo, H. Adaptive feature fusion with attention mechanism for multi-scale target detection. Neural Comput. Appl. 2021, 33, 2769–2781. [Google Scholar] [CrossRef]
- Wang, Q.; Wu, B.; Zhu, P.; Li, P.; Zuo, W.; Hu, Q. ECA-Net: Efficient channel attention for deep convolutional neural networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 11534–11542. [Google Scholar]
- Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. CBAM: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
- Qin, Z.; Zhang, P.; Wu, F.; Li, X. FCANet: Frequency channel attention networks. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 783–792. [Google Scholar]
- Zhang, Z.; Lan, C.; Zeng, W.; Jin, X.; Chen, Z. Relation-aware global attention for person re-identification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 3186–3195. [Google Scholar]
- Tan, M.; Pang, R.; Le, Q.V. EfficientDet: Scalable and efficient object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 10781–10790. [Google Scholar]
- Rubaiyat, A.H.M.; Toma, T.T.; Kalantari-Khandani, M.; Rahman, S.A.; Chen, L.; Ye, Y.; Pan, C.S. Automatic Detection of Helmet Uses for Construction Safety. In Proceedings of the 2016 IEEE/WIC/ACM International Conference on Web Intelligence Workshops (WIW), Omaha, NE, USA, 13–16 October 2016; pp. 135–142. [Google Scholar] [CrossRef]
- Seshadri, K.; Juefei-Xu, F.; Pal, D.K.; Savvides, M.; Thor, C.P. Driver Cell Phone Usage Detection on Strategic Highway Research Program (SHRP2) Face View Videos. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Boston, MA, USA, 7–12 June 2015; pp. 35–43. [Google Scholar]
- Wang, J. Driver Cell Phone Usage Detection Based on Semisupervised Support Vector Machine. Master’s Thesis, Hunan University, Changsha, China, 2018. [Google Scholar]
- Pan, G.; Yuan, Q.; Fan, C.; Qiao, H.; Wang, Z. Smoking Detection Algorithm Based on Mixture Gaussian Model and Frame Difference Method. Comput. Eng. Des. 2015, 36, 1290–1294. [Google Scholar]
- Ai, B. Research on Indoor Cigarette Smoke Detection Algorithm Based on Video Surveillance. Master’s Thesis, Yanshan University, Qinhuangdao, China, 2016. [Google Scholar]
- Guo, Q.; Liu, N.; Wang, Z.; Sun, Y. Overview of Object Detection Algorithms Based on Deep Learning. J. Detect. Control 2023, 45, 10–20,26. [Google Scholar]
- Dey, A.K.; Goel, B.; Chellappan, S. Context-Driven Detection of Distracted Driving Using Images from In-Car Cameras. Internet Things 2021, 14, 100380. [Google Scholar] [CrossRef]
- Senyurek, V.Y.; Imtiaz, M.H.; Belsare, P.; Tiffany, S.; Sazonov, E. A Comparison of SVM and CNN-LSTM Based Approach for Detecting Smoke Inhalations from Respiratory Signal. In Proceedings of the 2019 41st Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Berlin, Germany, 23–27 July 2019; pp. 3262–3265. [Google Scholar]
- Han, G.; Li, Q. Rapid Smoking Detection Algorithm Based on Faster R-CNN. J. Xi’An Univ. Posts Telecommun. 2020. [Google Scholar] [CrossRef]
- Wang, Y. Research on Early Warning of Unsafe Behavior of Construction Workers Based on Convolutional Neural Network. Master’s Thesis, Xi’an University of Architecture and Technology, Xi’an, China, 2021. [Google Scholar]
- Chen, S.; Tang, W.; Ji, T.; Zhu, H.; Ouyang, Y.; Wang, W. Detection of Safety Helmet Wearing Based on Improved Faster R-CNN. In Proceedings of the 2020 International Joint Conference on Neural Networks (IJCNN), Glasgow, UK, 19–24 July 2020; pp. 1–7. [Google Scholar]
- Aboah, A.; Wang, B.; Bagci, U.; Adu-Gyamfi, Y. Real-Time Multi-Class Helmet Violation Detection Using Few-Shot Data Sampling Technique and Yolov8. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–20 June 2023; pp. 5349–5357. [Google Scholar]
- Fan, X.; Wang, F.; Pang, S.; Wang, J.; Wang, W. Safety Helmet Wearing Detection Based on EfficientDet Algorithm. In Proceedings of the 2nd International Conference on Artificial Intelligence, Automation, and High-Performance Computing (AIAHPC 2022), Xi’an, China, 15–16 October 2022; SPIE: Bellingham, WA, USA, 2022; Volume 12348, pp. 302–309. [Google Scholar]
- Yang, T.; Yang, J.; Meng, J. Driver’s Illegal Driving Behavior Detection with SSD Approach. In Proceedings of the 2021 IEEE 2nd International Conference on Pattern Recognition and Machine Learning (PRML), Hangzhou, China, 13–15 November 2021; pp. 109–114. [Google Scholar]
- Zhao, Z.; Zhao, H.; Ye, C.; Xu, X.; Hao, K.; Yan, H.; Zhang, L.; Xu, Y. FPN-D-Based Driver Smoking Behavior Detection Method. IETE J. Res. 2023, 69, 5497–5506. [Google Scholar] [CrossRef]
- She, Y.; Zhang, X. Improved YOLOX Method for Small Target Smoking Detection Algorithm. In Proceedings of the International Conference on Cyber Security, Artificial Intelligence, and Digital Economy (CSAIDE 2023), Hangzhou, China, 14–15 September 2023; SPIE: Bellingham, WA, USA, 2023; Volume 12718, pp. 452–460. [Google Scholar]
- Lei, Y.; Zhu, W.; Liao, H. Improved YOLOv8n Algorithm for Safety Helmet Detection in Complex Scenes. Softw. Eng. 2023, 26, 46–51. [Google Scholar] [CrossRef]
- Zhang, T.; Qi, G.J.; Xiao, B.; Wang, J. Interleaved Group Convolutions. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 4373–4382. [Google Scholar]
- Hua, B.S.; Tran, M.K.; Yeung, S.K. Pointwise Convolutional Neural Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 984–993. [Google Scholar]
- Wang, C.Y.; Liao, H.Y.M.; Wu, Y.H.; Chen, P.Y.; Hsieh, J.W.; Yeh, I.H. CSPNet: A New Backbone That Can Enhance Learning Capability of CNN. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Seattle, WA, USA, 14–19 June 2020; pp. 390–391. [Google Scholar]
- Zhang, X.; Zhang, Y.; Hu, M.; Ju, X. Insulator Defect Detection Based on YOLO and SPP-Net. In Proceedings of the 2020 International Conference on Big Data & Artificial Intelligence & Software Engineering (ICBASE), Nanjing, China, 17–19 January 2020; pp. 403–407. [Google Scholar]
- Hu, J.; Shen, L.; Sun, G. Squeeze-and-Excitation Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar]
- Roboflow Universe Projects. Construction Site Safety Dataset [Dataset]. Roboflow Universe. 2023. Available online: https://universe.roboflow.com/roboflow-universe-projects/construction-site-safety (accessed on 1 August 2023).
- Bono, F.M.; Radicioni, L.; Cinquemani, S. A novel approach for quality control of automated production lines working under highly inconsistent conditions. Eng. Appl. Artif. Intell. 2023, 122, 106149. [Google Scholar] [CrossRef]
- Wan, D.; Lu, R.; Shen, S.; Xu, T.; Lang, X.; Ren, Z. Mixed Local Channel Attention for Object Detection. Eng. Appl. Artif. Intell. 2023, 123, 106442. [Google Scholar] [CrossRef]
- Li, C.; Li, L.; Jiang, H.; Weng, K.; Geng, Y.; Li, L.; Ke, Z.; Li, Q.; Cheng, M.; Nie, W.; et al. YOLOv6: A Single-Stage Object Detection Framework for Industrial Applications. arXiv 2022, arXiv:2209.02976. [Google Scholar]
- Wang, C.Y.; Bochkovskiy, A.; Liao, H.Y.M. YOLOv7: Trainable Bag-of-Freebies Sets New State-of-the-Art for Real-Time Object Detectors. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–23 June 2023; pp. 7464–7475. [Google Scholar]
- Weber, M.; Wang, H.; Qiao, S.; Xie, J.; Collins, M.D.; Zhu, Y.; Yuan, L.; Kim, D.; Yu, Q.; Cremers, D.; et al. Deeplab2: A Tensorflow Library for Deep Labeling. arXiv 2021, arXiv:2106.09748. [Google Scholar]
- Liu, S.; Huang, D. Receptive Field Block Net for Accurate and Fast Object Detection. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 385–400. [Google Scholar]
- Redmon, J.; Farhadi, A. YOLOv3: An Incremental Improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
- Jocher, G.; Stoken, A.; Borovec, J.; NanoCode; Chaurasia, A.; Xie, T.; Liu, C.; Abhiram, V.; Laughing; tkianai; et al. Ultralytics/yolov5: V5.0-YOLOv5-P6 1280 Models, AWS, Supervise.ly and YouTube Integrations. Zenodo 2021. [Google Scholar] [CrossRef]
- Wu, X.; Qian, S.; Yang, M. Detection of Safety Helmet-Wearing Based on the YOLO-CA Model. Comput. Mater. Contin. 2023, 77, 3349–3366. [Google Scholar]
Hyperparameter | Values |
---|---|
Batch size | 32 |
Epoch | 100 |
NMS IoU | 0.7 |
Initial Learning Rate | 0.01 |
Final Learning Rate | 0.01 |
Weight decay | 0.0005 |
Momentum | 0.937 |
Input size | 640 |
Dataset | Models | Precision | Recall | [email protected] | [email protected]:.95 | GFLOPs | F1-Score | Para/M |
---|---|---|---|---|---|---|---|---|
DBD | Baseline | 80.1 | 73.3 | 77.6 | 44.9 | 8.9 | 76.5 | 3.16 |
Baseline + DAM | 84.1 | 72.4 | 78.9 | 45.3 | 13.7 | 77.8 | 5.49 | |
Baseline + ESPP | 79.4 | 74.6 | 78.4 | 45.6 | 10.1 | 76.9 | 4.56 | |
Baseline + CFE-Net | 80.3 | 74.0 | 78.5 | 45.4 | 15.1 | 77.8 | 3.38 | |
YOLO-GP | 82.6 | 73.6 | 79.2 | 45.6 | 15.1 | 77.8 | 7.27 | |
CSSID | Baseline | 80.6 | 54.6 | 61.4 | 31.4 | 8.9 | 65.1 | 3.16 |
Baseline + DAM | 85.4 | 57.0 | 65.0 | 35.0 | 13.7 | 68.4 | 5.49 | |
Baseline + ESPP | 81.1 | 57.8 | 64.5 | 34.4 | 10.1 | 67.5 | 4.56 | |
Baseline + CFE-Net | 72.8 | 53.7 | 58.6 | 30.4 | 10.1 | 61.8 | 3.38 | |
YOLO-GP | 81.3 | 59.8 | 66.3 | 35.0 | 15.1 | 68.9 | 7.27 |
Negative_Slope | Precision | Recall | [email protected] | [email protected]:.95 |
---|---|---|---|---|
0.2 | 78.3 | 73.0 | 77.0 | 44.5 |
0.3 | 77.5 | 72.5 | 76.3 | 44.7 |
0.4 | 80.3 | 74.0 | 78.5 | 45.4 |
0.5 | 81.0 | 71.5 | 77.2 | 44.6 |
0.6 | 78.5 | 71.8 | 77.5 | 44.5 |
Dataset | Method | Precision | Recall | [email protected] | [email protected]:.95 | GFLOPs | Para/M |
---|---|---|---|---|---|---|---|
DBD | Baseline + SPPF | 80.1 | 73.3 | 77.6 | 44.9 | 8.9 | 3.16 |
Baseline + SPPCSPC | 81.2 | 73.7 | 78.1 | 45.6 | 10.1 | 4.72 | |
Baseline + SPPFCSPC | 78.2 | 72.8 | 77.5 | 45.2 | 10.1 | 4.77 | |
Baseline + SimSPPF | 80.9 | 74.2 | 77.7 | 45.1 | 8.9 | 3.16 | |
Baseline + ASPP | 77.6 | 75.8 | 77.9 | 45.2 | 10.5 | 5.22 | |
Baseline + RFB | 80.8 | 72.6 | 77.3 | 44.7 | 9.0 | 3.32 | |
Baseline + ESPP | 79.4 | 74.6 | 78.4 | 45.6 | 10.1 | 4.56 | |
CSSID | Baseline + SPPF | 80.6 | 54.6 | 61.4 | 31.4 | 8.9 | 3.16 |
Baseline + SPPCSPC | 78.8 | 58.6 | 64.2 | 34.6 | 10.1 | 4.72 | |
Baseline + SPPFCSPC | 80.7 | 56.5 | 62.5 | 32.8 | 10.1 | 4.77 | |
Baseline + SimSPPF | 72.4 | 56.2 | 60.7 | 30.7 | 8.9 | 3.16 | |
Baseline + ASPP | 77.2 | 54.5 | 61.0 | 30.8 | 10.5 | 5.22 | |
Baseline + RFB | 78.8 | 58.3 | 63.0 | 32.6 | 9.0 | 3.32 | |
Baseline + ESPP | 81.1 | 57.8 | 64.5 | 34.4 | 10.1 | 4.56 |
Baseline | YOLO-GP | ||||||||
---|---|---|---|---|---|---|---|---|---|
Dataset | Category | Precision | Recall | [email protected] | [email protected]:.95 | Precision | Recall | [email protected] | [email protected]:.95 |
DBD | Helmet | 89.9 | 84.6 | 88.4 | 44.9 | 93.5↑ | 85.0↑ | 89.9↑ | 45.6↑ |
Head | 92.3 | 90.8 | 94.0 | 61.2 | 91.9 | 91.2↑ | 94.4↑ | 61.4↑ | |
Phone | 77.4 | 66.5 | 72.5 | 36.2 | 80.5↑ | 67.7↑ | 74.1↑ | 36.9↑ | |
Smoke | 61.0 | 51.5 | 55.6 | 21.5 | 64.4↑ | 50.5 | 58.2↑ | 22.1↑ | |
CSSID | Hardhat | 96.4 | 60.8 | 73.6 | 42.5 | 91.7 | 64.6↑ | 74.3↑ | 43.0↑ |
Mask | 89.4 | 80.0 | 85.6 | 49.5 | 92.0↑ | 81.0↑ | 87.4↑ | 51.2↑ | |
No-hardhat | 78.9 | 43.4 | 49.2 | 22.2 | 74.3 | 49.3↑ | 55.5↑ | 23.7↑ | |
No-mask | 72.2 | 37.8 | 47.3 | 20.0 | 74.2↑ | 44.6↑ | 52.1↑ | 20.5↑ | |
No-safty vest | 74.7 | 41.5 | 48.9 | 24.8 | 76.7↑ | 52.8↑ | 61.1↑ | 29.8↑ | |
Person | 84.2 | 54.5 | 63.4 | 30.6 | 80.5 | 62.7↑ | 69.6↑ | 36.0↑ | |
Safty corn | 88.9 | 70.5 | 77.0 | 36.6 | 74.7 | 75.0↑ | 76.1 | 38.1↑ | |
Safty vest | 83.7 | 62.5 | 67.4 | 37.2 | 80.9 | 61.9 | 67.4 | 39.2↑ | |
Machinery | 65.5 | 70.9 | 72.4 | 33.8 | 75.1↑ | 74.5↑ | 77.7↑ | 45.1↑ | |
Vehicle | 72.7 | 23.8 | 29.5 | 17.2 | 93.1↑ | 32.0↑ | 41.4↑ | 23.3↑ |
Dataset | Method | Precision | Recall | [email protected] | [email protected]:.95 | GFLOPs | Inference Time/ms | Para/M |
---|---|---|---|---|---|---|---|---|
DBD | YOLOv3-tiny | 77.8 | 71.2 | 74.5 | 40.8 | 19.1 | 0.7 | 12.17 |
YOLOv5 | 76.8 | 73.3 | 74.8 | 40.0 | 4.2 | 9.7 | 1.78 | |
YOLOv6 | 75.7 | 70.8 | 74.2 | 43.8 | 13.1 | 0.5 | 4.50 | |
YOLOv7-tiny | 76.5 | 71.8 | 74.6 | 39.4 | 13.2S | 7.8 | 6.02 | |
YOLOv8n | 80.1 | 73.3 | 77.6 | 44.9 | 8.9 | 9.9 | 3.16 | |
YOLO-CA | 81.8 | 72.8 | 76.7 | 41.6 | 12.6 | 13.3 | 5.88 | |
YOLO-GP (Ours) | 82.6 | 73.6 | 79.2 | 45.6 | 15.1 | 14.2 | 7.27 | |
CSSID | YOLOv3-tiny | 78.7 | 55.0 | 61.1 | 32.8 | 19.1 | 1.1 | 12.17 |
YOLOv5 | 67.5 | 51.8 | 55.1 | 22.6 | 4.3 | 9.1 | 1.78 | |
YOLOv6 | 82.6 | 54.9 | 62.1 | 32.6 | 13.1 | 0.7 | 4.50 | |
YOLOv7-tiny | 69.2 | 53.9 | 56.4 | 23.4 | 13.2S | 7.0 | 6.02 | |
YOLOv8n | 80.6 | 54.6 | 61.4 | 31.4 | 8.9 | 8.1 | 3.16 | |
YOLO-CA | 75.8 | 60.0 | 65.1 | 28.8 | 12.6 | 11.9 | 5.88 | |
YOLO-GP (Ours) | 81.3 | 59.8 | 66.3 | 35.0 | 15.1 | 11.7 | 7.27 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Liu, B.; Yu, C.; Chen, B.; Zhao, Y. YOLO-GP: A Multi-Scale Dangerous Behavior Detection Model Based on YOLOv8. Symmetry 2024, 16, 730. https://doi.org/10.3390/sym16060730
Liu B, Yu C, Chen B, Zhao Y. YOLO-GP: A Multi-Scale Dangerous Behavior Detection Model Based on YOLOv8. Symmetry. 2024; 16(6):730. https://doi.org/10.3390/sym16060730
Chicago/Turabian StyleLiu, Bushi, Cuiying Yu, Bolun Chen, and Yue Zhao. 2024. "YOLO-GP: A Multi-Scale Dangerous Behavior Detection Model Based on YOLOv8" Symmetry 16, no. 6: 730. https://doi.org/10.3390/sym16060730
APA StyleLiu, B., Yu, C., Chen, B., & Zhao, Y. (2024). YOLO-GP: A Multi-Scale Dangerous Behavior Detection Model Based on YOLOv8. Symmetry, 16(6), 730. https://doi.org/10.3390/sym16060730