Active Hard Sample Learning for Violation Action Recognition in Power Grid Operation
Abstract
:1. Introduction
- To improve the training efficiency of the network, we design a hard instance sampling (HIS) module with multi-strategy fusion to reduce the iteration time of the training. This module identifies hard samples based on the consistency of models or samples and performs iterative learning through active learning to ensure continuous updates of the network.
- We develop uncertainty evaluation (UE) and the instance discrimination strategy (IDS) to eliminate redundant data while ensuring sample diversity and balance and merge the two manners to assess the contributions of samples effectively, so as to obtain valuable hard samples for the network.
- Experimental results demonstrate significant improvements in both model performance and training efficiency, showcasing the effectiveness of the proposed method in complex power grid environments.
2. Methods
2.1. Hard Instance Sampling Module
2.2. Optimization
3. Experiments
3.1. Evaluation Criteria
3.2. Experimental Details
3.3. Results Comparison and Analysis
4. Discussion
5. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- Mohamed, A.A.R.; Best, R.J.; Liu, X.; Morrow, D.J. Two-phase BESS optimization methodology to enhance the distribution network power quality and mitigate violations. IET Renew. Power Gener. 2023, 17, 2895–2908. [Google Scholar] [CrossRef]
- Zhou, F.; Lu, H.; Jiang, C. Violation with concerns of safety: A study on non-compliant behavior and the antecedent and consequent effects in power grid construction. Saf. Sci. 2024, 170, 106353. [Google Scholar] [CrossRef]
- Ban, G.; Fu, L.; Jiang, L.; Du, H.; Li, A.; He, Y.; Zhou, J. Two-stage dynamic risk identification of complex work personnel behavior based on image screening. Power Syst. Big Data 2024, 27, 58–69. [Google Scholar]
- Feichtenhofer, C.; Fan, H.; Malik, J.; He, K. Slowfast networks for video recognition. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 6202–6211. [Google Scholar]
- Feichtenhofer, C. X3d: Expanding architectures for efficient video recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 203–213. [Google Scholar]
- Bertasius, G.; Wang, H.; Torresani, L. Is space-time attention all you need for video understanding? In Proceedings of the International Conference on Machine Learning, Virtual, 18–24 July 2021; p. 4. [Google Scholar]
- Alafif, T.; Hadi, A.; Allahyani, M.; Alzahrani, B.; Alhothali, A.; Alotaibi, R.; Barnawi, A. Hybrid classifiers for spatio-temporal abnormal behavior detection, tracking, and recognition in massive Hajj crowds. Electronics 2023, 12, 1165. [Google Scholar] [CrossRef]
- Meydani, A.; Shahinzadeh, H.; Ramezani, A.; Moazzami, M.; Nafisi, H.; Askarian-Abyaneh, H. Comprehensive Review of Artificial Intelligence Applications in Smart Grid Operations. In Proceedings of the International Conference on Technology and Energy Management, Behshahr, Iran, 14–15 February 2024; pp. 1–13. [Google Scholar]
- Meng, L.; Ban, G.; Liu, F.; Qiu, W.; He, D.; Zhang, L.; Wang, S. Illegal action classification in power grid operation based on cross-domain few-shot learning. Power Syst. Big Data 2024, 27, 69–76. [Google Scholar]
- Wang, J.; Zhou, H.; Sun, H.; Su, Z.; Li, X. A Violation Behaviors Detection Method for Substation Operators based on YOLOv5 And Pose Estimation. In Proceedings of the IEEE China International Youth Conference on Electrical Engineering, Wuhan, China, 3–5 November 2022; pp. 1–5. [Google Scholar]
- Li, X.; Chen, H.; Tian, Y.; Zhong, Y.; Jiang, G. Identification Method of Underground Operation Violation Behavior Based on Deep Learning. In Proceedings of the International Conference on Algorithms, High Performance Computing and Artificial Intelligence, Guangzhou, China, 21–23 October 2022; pp. 547–550. [Google Scholar]
- Elezi, I.; Yu, Z.; Anandkumar, A.; Leal-Taixe, L.; Alvarez, J.M. Not all labels are equal: Rationalizing the labeling costs for training object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 14492–14501. [Google Scholar]
- Parvaneh, A.; Abbasnejad, E.; Teney, D.; Haffari, G.R.; Van Den Hengel, A.; Shi, J.Q. Active learning by feature mixing. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 12237–12246. [Google Scholar]
- Wu, Z.; Li, H.; Zheng, Y.; Xiong, C.; Jiang, Y.G.; Davis, L.S. A coarse-to-fine framework for resource efficient video recognition. Int. J. Comput. Vis. 2021, 129, 2965–2977. [Google Scholar] [CrossRef]
- Liu, Z.; Ning, J.; Cao, Y.; Wei, Y.; Zhang, Z.; Lin, S.; Hu, H. Video swin transformer. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 3202–3211. [Google Scholar]
- Li, W.; Pan, G.; Wang, C.; Xing, Z.; Han, Z. From coarse to fine: Hierarchical structure-aware video summarization. ACM Trans. Multimed. Comput. Commun. Appl. 2022, 18, 1–16. [Google Scholar] [CrossRef]
- Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 39, 1137–1149. [Google Scholar] [CrossRef] [PubMed]
- Tarvainen, A.; Valpola, H. Mean teachers are better role models: Weight-averaged consistency targets improve semi-supervised deep learning results. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
- Jain, P.; Kapoor, A. Active learning for large multi-class problems. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 762–769. [Google Scholar]
- Li, X.; Guo, Y. Adaptive active learning for image classification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Portland, OR, USA, 23–28 June 2013; pp. 859–866. [Google Scholar]
- Yang, Y.; Ma, Z.; Nie, F.; Chang, X.; Hauptmann, A.G. Multi-class active learning by uncertainty sampling with diversity maximization. Int. J. Comput. Vis. 2015, 113, 113–127. [Google Scholar] [CrossRef]
- Heilbron, F.C.; Lee, J.Y.; Jin, H.; Ghanem, B. What do i annotate next? an empirical study of active learning for action localization. In Proceedings of the European Conference on Computer Vision, Munich, Germany, 8–14 September 2018; pp. 199–216. [Google Scholar]
- Xing, Z.; Dai, Q.; Hu, H.; Chen, J.; Wu, Z.; Jiang, Y.G. Svformer: Semi-supervised video transformer for action recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 18816–18826. [Google Scholar]
- NFPA-70E. Available online: https://www.nfpa.org/codes-and-standards/nfpa-70e-standard-development/70e (accessed on 27 November 2024).
- He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
- Kolesnikov, A.; Dosovitskiy, A.; Weissenborn, D.; Heigold, G.; Uszkoreit, J.; Beyer, L.; Minderer, M.; Dehghani, M.; Houlsby, N.; Gelly, S.; et al. An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale. In Proceedings of the International Conference on Learning Representations, Vienna, Austria, 4 May 2021. [Google Scholar]
- Deng, J.; Dong, W.; Socher, R.; Li, L.J.; Li, K.; Fei-Fei, L. Imagenet: A large-scale hierarchical image database. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar]
- Dave, I.R.; Rizve, M.N.; Chen, C.; Shah, M. Timebalance: Temporally-invariant and temporally-distinctive video representations for semi-supervised action recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 2341–2352. [Google Scholar]
- Zhang, Y.; Bai, Y.; Wang, H.; Xu, Y.; Fu, Y. Look more but care less in video recognition. In Proceedings of the Advances in Neural Information Processing Systems, New Orleans, LA, USA, 28 November–9 December 2022; pp. 30813–30825. [Google Scholar]
Types | Categories | 30% | 60% | ||
---|---|---|---|---|---|
BS-R-50 | AHSL-R-50 | BS-R-50 | AHSL-R-50 | ||
personal protective equipment | safety helmet | 85.67% | 90.54% | 92.23% | 95.18% |
work clothes | 73.28% | 76.82% | 78.56% | 82.32% | |
insulated gloves | 81.96% | 85.24% | 85.52% | 86.46% | |
insulated shoes | 68.75% | 72.51% | 73.13% | 74.67% | |
safety harness | 66.64% | 69.78% | 72.43% | 76.56% | |
others | smoking | 83.29% | 88.07% | 91.03% | 93.45% |
crossing security boundary | 78.95% | 84.32% | 86.36% | 92.45% | |
throwing wires or tools | 83.17% | 87.76% | 90.45% | 94.78% | |
Average Accuracy | 77.74% | 81.88% | 83.71% | 86.98% |
Types | Categories | 30% | 60% | ||
---|---|---|---|---|---|
BS-ViT-S | AHSL-ViT-S | BS-ViT-S | AHSL-ViT-S | ||
personal protective equipment | safety helmet | 87.52% | 91.78% | 92.55% | 95.89% |
work clothes | 73.87% | 77.25% | 77.92% | 81.78% | |
insulated gloves | 81.87% | 85.16% | 84.63% | 86.96% | |
insulated shoes | 69.65% | 73.52% | 72.32% | 74.45% | |
safety harness | 67.49% | 71.13% | 70.85% | 77.09% | |
others | smoking | 86.92% | 90.27% | 92.14% | 94.07% |
crossing security boundary | 82.37% | 86.58% | 85.06% | 92.83% | |
throwing wires or tools | 85.74% | 88.94% | 89.75% | 94.98% | |
Average Accuracy | 79.43% | 83.08% | 83.15% | 87.26% |
Method | Average Accuracy ↑ | Recall ↑ | Precision ↑ | F1 ↑ | FPR ↓ | FNR ↓ |
---|---|---|---|---|---|---|
BS-R-50 | 83.71% | 77.32% | 80.31% | 78.79% | 12.75% | 15.25% |
AHSL-R-50 (w/o UE) | 84.86% | 79.85% | 82.47% | 81.14% | 11.00% | 13.00% |
AHSL-R-50 (w/o IDS) | 85.15% | 82.34% | 84.89% | 83.60% | 9.25% | 11.25% |
AHSL-R-50 | 86.98% | 85.23% | 86.55% | 85.88% | 8.00% | 9.00% |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Meng, L.; He, D.; Ban, G.; Xi, G.; Li, A.; Zhu, X. Active Hard Sample Learning for Violation Action Recognition in Power Grid Operation. Information 2025, 16, 67. https://doi.org/10.3390/info16010067
Meng L, He D, Ban G, Xi G, Li A, Zhu X. Active Hard Sample Learning for Violation Action Recognition in Power Grid Operation. Information. 2025; 16(1):67. https://doi.org/10.3390/info16010067
Chicago/Turabian StyleMeng, Lingwen, Di He, Guobang Ban, Guanghui Xi, Anjun Li, and Xinshan Zhu. 2025. "Active Hard Sample Learning for Violation Action Recognition in Power Grid Operation" Information 16, no. 1: 67. https://doi.org/10.3390/info16010067
APA StyleMeng, L., He, D., Ban, G., Xi, G., Li, A., & Zhu, X. (2025). Active Hard Sample Learning for Violation Action Recognition in Power Grid Operation. Information, 16(1), 67. https://doi.org/10.3390/info16010067