Object Recognition Scheme for Digital Transformation in Marine Science and Engineering
Abstract
:1. Introduction
- We briefly analyze the drawbacks that come from marine science and engineering environments using DL technologies.
- We propose a novel scheme, called NoOP, which enables DL technologies to recognize objects (i.e., lines, tags, and symbols) in P&ID documents without learning models in a large dataset.
- We implement NoOP and evaluate its performance and accuracy with the real-world P&ID document. In addition, to clearly understand the effectiveness of NoOP, we show how to recognize lines, tags, and symbols by dividing the description into several steps.
2. Background
2.1. What Is P&ID Documents?
2.2. Recognition Schemes Based on Deep Learning
3. Related Work
4. Design and Implementation
4.1. Tag Detection
4.2. Symbol Detection
4.3. Line Detection
4.4. Semantic Generation
5. Evaluation
5.1. Experimental Setup and Workload
5.2. The Results of Tag and Symbol Detection
5.3. The Results of Line Detection
5.4. The Results of the Semantic Generation
5.5. The Robustness of NoOP
5.6. The Elapsed Time of NoOP
6. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
Abbreviations
DL | deep learning |
CNN | convolution neural network |
DNN | deep neural network |
YOLO | you only look once |
CAD | computer-aided design |
portable document format | |
P&ID | piping and instrumentation diagram |
DX | digital transformation |
OCR | optical character recognition |
FCN | fully convolution networks |
ICDAR | International Conference of Document Analysis and Recognition |
VGG | visual geometry group |
EAST | efficient and accurate scene text detector |
CRAFT | character-region awareness for text detection |
CTPN | connectionist text proposal symbol network |
LSTM | long short-term memory |
elements of captured symbol bounding box candidate | |
upper left coordinate of the text bounding box | |
upper left coordinate of the symbol bounding box | |
width of the text bounding box | |
width of the symbol bounding box | |
height of the text bounding box | |
height of the symbol bounding box |
References
- Martin, A.; Paul, B.; Jianmin, C.; Zhifeng, C.; Andy, D.; Jeffrey, D.; Matthieu, D.; Sanja, Y.G.; Geoffrey, I.; Michael, I.; et al. TensorFlow: A System for Large-Scale Machine Learning. In Proceedings of the Symposium on Operating Systems Design and Implementation, Savannah, GA, USA, 2–4 November 2016; pp. 265–283. [Google Scholar]
- Wang, T.; Toh, W.Q.; Zhang, H.; Sui, X.; Li, S.; Liu, Y.; Jing, W. RoboCoDraw: Robotic Avatar Drawing with GAN-Based Style Transfer and Time-Efficient Path Optimization. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 10402–10409. [Google Scholar]
- Brown, T.; Mann, B.; Ryder, N.; Subbiah, M.; Kaplan, J.D.; Dhariwal, P.; Neelakantan, A.; Shyam, P.; Sastry, G.; Askell, A.; et al. Language Models are Few-Shot Learners. In Proceedings of the Advances in Neural Information Processing Systems 33, Proceedings of the Annual Conference on Neural Information Processing Systems 2020, NeurIPS 2020, Virtual, 6–12 December 2020; Volume 33, pp. 1877–1901.
- Cui, L.; Biswal, S.; Glass, L.M.; Lever, G.; Sun, J.; Xiao, C. CONAN: Complementary Pattern Augmentation for Rare Disease Detection. In Proceedings of the AAAI Conference on Artificial Intelligence, Washington, DC, USA, 7–12 February 2020; Volume 34, pp. 614–621. [Google Scholar]
- Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada, 7–12 December 2015; Volume 28. [Google Scholar]
- Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. SSD: Single Shot MultiBox Detector. In Proceedings of the European Conference On Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; pp. 21–37. [Google Scholar]
- Redmon, J.; Farhadi, A. YOLOv3: An Incremental Improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
- Tan, M.; Le, Q. EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks. In Proceedings of the International Conference on Machine Learning, Long Beach, CA, USA, 11–13 June 2019; Volume 97, pp. 6105–6114. [Google Scholar]
- Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You Only Look Once: Unified, Real-Time Object Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
- Ketkar, N. Introduction to Keras. In Deep Learning with Python: A Hands-On Introduction; Apress: Berkely, CA, USA, 2017; pp. 97–111. Available online: https://link.springer.com/book/10.1007/978-1-4842-2766-4 (accessed on 24 April 2023).
- Park, H.; Ham, S.H.; Kim, T.; An, D. Object Recognition and Tracking in Moving Videos for Maritime Autonomous Surface Ships. J. Mar. Sci. Eng. 2022, 10, 841. [Google Scholar] [CrossRef]
- Li, L.; Jiang, L.; Zhang, J.; Wang, S.; Chen, F. A Complete YOLO-Based Ship Detection Method for Thermal Infrared Remote Sensing Images under Complex Backgrounds. Remote Sens. 2022, 14, 1534. [Google Scholar] [CrossRef]
- Kim, M.; Choi, W.; Kim, B.C.; Kim, H.; Seol, J.H.; Woo, J.; Ko, K.H. A Vision-based System for Monitoring Block Assembly in Shipbuilding. Comput.-Aided Des. 2015, 59, 98–108. [Google Scholar] [CrossRef]
- Choi, Y.; Park, J.H.; Jang, B. A Risk Estimation Approach based on Deep Learning in Shipbuilding Industry. In Proceedings of the IEEE International Conference on Information and Communication Technology Convergence, Jeju, Republic of Korea, 16–18 October 2019; pp. 1438–1441. [Google Scholar]
- Kong, M.C.; Roh, M.I.; Kim, K.S.; Lee, J.; Kim, J.; Lee, G. Object Detection Method for Ship Safety Plans using Deep Learning. Ocean. Eng. 2022, 246, 110587. [Google Scholar] [CrossRef]
- Smith, R. Modernization Efforts: Cleaning up the Code and Adding New LSTM Technology. 2016. Available online: https://tesseract-ocr.github.io/docs/das_tutorial2016/6ModernizationEfforts.pdf (accessed on 20 April 2023).
- Smith, R. Tesseract-OCR Library. Available online: https://github.com/tesseract-ocr/tesseract (accessed on 20 April 2023).
- Illingworth, J.; Kittler, J. A Survey of the Hough Transform. J. Comput. Vision Graph. Image Process. 1988, 44, 87–116. [Google Scholar] [CrossRef]
- Alekhin, A. OpenCV Library. Available online: https://opencv.org/ (accessed on 30 March 2023).
- Tan, W.C.; Chen, I.M.; Tan, H.K. Automated Identification of Components in Raster Piping and Instrumentation Diagram with Minimal Pre-processing. In Proceedings of the IEEE International Conference on Automation Science and Engineering (CASE), Fort Worth, TX, USA, 21–25 August 2016; pp. 1301–1306. [Google Scholar]
- Long, J.; Shelhamer, E.; Darrell, T. Fully Convolutional Networks for Semantic Segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
- Zhou, X.; Yao, C.; Wen, H.; Wang, Y.; Zhou, S.; He, W.; Liang, J. EAST: An Efficient and Accurate Scene Text Detector. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 9365–9374. [Google Scholar]
- Baek, Y.; Lee, B.; Han, D.; Yun, S.; Lee, H. Character Region Awareness for Text Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–19 June 2019; pp. 9365–9374. [Google Scholar]
- Smith, R. An Overview of the Tesseract OCR Engine. In Proceedings of the International Conference on Document Analysis and Recognition, Curitiba, Brazil, 23–26 September 2007; Volume 2, pp. 629–633. [Google Scholar]
- Yu, E.S.; Cha, J.M.; Lee, T.; Kim, J.; Mun, D. Features Recognition from Piping and Instrumentation Diagrams in Image Format Using a Deep Learning Network. J. Energies 2019, 12, 4425. [Google Scholar] [CrossRef]
- Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-scale Image Recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
- Fujiyoshi, H.; Hirakawa, T.; Yamashita, T. Deep learning-based Image Recognition for Autonomous Driving. J. Int. Assoc. Traffic Saf. Sci. 2019, 43, 244–252. [Google Scholar] [CrossRef]
- Sanchez-Gonzalez, P.L.; Díaz-Gutiérrez, D.; Leo, T.J.; Núñez-Rivas, L.R. Toward Digitalization of Maritime Transport? Sensors 2019, 19, 926. [Google Scholar] [CrossRef] [PubMed]
- Park, S.; Huh, J.H. Study on PLM and Big Data Collection for the Digital Transformation of the Shipbuilding Industry. J. Mar. Sci. Eng. 2022, 10, 1488. [Google Scholar] [CrossRef]
- Pang, T.Y.; Pelaez Restrepo, J.D.; Cheng, C.T.; Yasin, A.; Lim, H.; Miletic, M. Developing a Digital Twin and Digital Thread Framework for an ‘Industry 4.0’ Shipyard. Appl. Sci. 2021, 11, 1097. [Google Scholar] [CrossRef]
- Lee, G.A.; Yang, U.; Son, W.; Kim, Y.; Jo, D.; Kim, K.H.; Choi, J.S. Virtual Reality Content-Based Training for Spray Painting Tasks in the Shipbuilding Industry. ETRI J. 2010, 32, 695–703. [Google Scholar] [CrossRef]
- Zheng, Q.; Tian, X.; Yu, Z.; Jiang, N.; Elhanashi, A.; Saponara, S.; Yu, R. Application of Wavelet-Packet Transform Driven Deep Learning Method in PM2.5 Concentration Prediction: A Case Study of Qingdao, China. Sustain. Cities Soc. 2023, 92, 1–13. [Google Scholar] [CrossRef]
- Rahul, R.; Paliwal, S.; Sharma, M.; Vig, L. Automatic information extraction from piping and instrumentation diagrams. In Proceedings of the International Conference on Pattern Recognition Applications and Methods, Prague, Czech Republic, 19–21 February 2019; pp. 163–172. [Google Scholar]
- Kang, S.O.; Lee, E.B.; Baek, H.K. A Digitization and Conversion Tool for Imaged Drawings to Intelligent Piping and Instrumentation Diagrams (P&Id). Energies 2019, 12, 2593. [Google Scholar]
- Moon, Y.; Lee, J.; Mun, D.; Lim, S. Deep Learning-Based Method to Recognize Line Objects and Flow Arrows from Image-Format Piping and Instrumentation Diagrams for Digitization. J. Appl. Sci. 2021, 11, 10054. [Google Scholar] [CrossRef]
- Kim, H.; Lee, W.; Kim, M.; Moon, Y.; Lee, T.; Cho, M.; Mun, D. Deep-learning-based Recognition of Symbols and Texts at an Industrially Applicable Level from Images of High-density Piping and Instrumentation Diagrams. J. Expert Syst. Appl. 2021, 183, 115337. [Google Scholar] [CrossRef]
- Fahn, C.S.; Wang, J.F.; Lee, J.Y. A Topology-based Component Extractor for Understanding Electronic Circuit Diagrams. Comput. Vision Graph. Image Process. 1988, 44, 119–138. [Google Scholar] [CrossRef]
- Kato, H.; Inokuchi, S. The Recognition Method for Roughly Hand-drawn Logical Diagrams Based on Hybrid Utilization of Multi-layered Knowledge. In Proceedings of the 10th International Conference on Pattern Recognition, Atlantic City, NJ, USA, 16–21 June 1990; Volume 1, pp. 578–582. [Google Scholar]
- De, P.; Mandal, S.; Bhowmick, P. Recognition of electrical symbols in document images using morphology and geometric analysis. In Proceedings of the 2011 International Conference on Image Information Processing, Shimla, India, 3–5 November 2011; pp. 1–6. [Google Scholar]
- Guo, G.; Wang, H.; Bell, D.; Bi, Y.; Greer, K. KNN Model-Based Approach in Classification. In Proceedings of the OTM Confederated International Conferences CoopIS, DOA, and ODBASE 2003 Catania, Sicily, Italy, 3–7 November 2003; pp. 986–996. [Google Scholar]
- Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollar, P. Focal Loss for Dense Object Detection. In Proceedings of the IEEE 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar]
- Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet Classification with Deep Convolutional Neural Networks. J. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
- Tian, Z.; Huang, W.; He, T.; He, P.; Qiao, Y. Detecting Text in Natural Image with Connectionist Text Proposal Network. In Proceedings of the 14th European Conference on Computer Vision—ECCV 2016, Amsterdam, The Netherlands, 11–14 October 2016; pp. 56–72. [Google Scholar]
- Sun, C.; Shrivastava, A.; Singh, S.; Gupta, A. Revisiting Unreasonable Effectiveness of Data in Deep Learning Era. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017. [Google Scholar]
- Hoffstaetter, S. PyTesseract. Available online: https://github.com/madmaze/pytesseract (accessed on 30 March 2023).
- Boncelet, C. Image noise models. In The Essential Guide to Image Processing; Elsevier: Amsterdam, The Netherlands, 2009; pp. 143–167. [Google Scholar]
- Toh, K.K.V.; Ibrahim, H.; Mahyuddin, M.N. Salt-and-pepper noise detection and reduction using fuzzy switching median filter. IEEE Trans. Consum. Electron. 2008, 54, 1956–1961. [Google Scholar] [CrossRef]
- Wikipedia. Available online: http://en.wikipedia.org/w/index.php?title=Watermark&oldid=1161923484 (accessed on 16 September 2023).
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Choi, J.; An, D.; Kang, D. Object Recognition Scheme for Digital Transformation in Marine Science and Engineering. J. Mar. Sci. Eng. 2023, 11, 1914. https://doi.org/10.3390/jmse11101914
Choi J, An D, Kang D. Object Recognition Scheme for Digital Transformation in Marine Science and Engineering. Journal of Marine Science and Engineering. 2023; 11(10):1914. https://doi.org/10.3390/jmse11101914
Chicago/Turabian StyleChoi, Jinseo, Donghyeok An, and Donghyun Kang. 2023. "Object Recognition Scheme for Digital Transformation in Marine Science and Engineering" Journal of Marine Science and Engineering 11, no. 10: 1914. https://doi.org/10.3390/jmse11101914
APA StyleChoi, J., An, D., & Kang, D. (2023). Object Recognition Scheme for Digital Transformation in Marine Science and Engineering. Journal of Marine Science and Engineering, 11(10), 1914. https://doi.org/10.3390/jmse11101914