Capacity Constraint Analysis Using Object Detection for Smart Manufacturing
Abstract
:1. Introduction
- A novel non-invasive theoretical framework for analyzing the capacity of manufacturing facilities by using OD methods to categorize workstations into different states.
- Collection and annotation of a real-world dataset from a production floor for use in training an OD model.
- Comprehensive experimentation and evaluation of the proposed framework in a real-world facility over 6 months, demonstrating its practical applicability and effectiveness.
2. Related Work
2.1. Evolution of YOLO Models
2.2. Applications of OD in Smart Manufacturing
3. Capacity Constraint Analysis
Cycle Time Study
- Stop Watches Tasks performed by workers can be manually timed in the manufacturing environment. This traditional method is commonly used for establishing benchmarks in manual operations. However, reliance on human observation limits the scalability and accuracy of this method, and may not accurately represent normal working conditions due to observer bias and the Hawthorne effect. Additionally, manual timing cannot easily be integrated into the broader data ecosystems that smart manufacturing relies on for real-time decision-making.
- Video Recording with Offline Analysis This allows for efficient analysis of processes by reviewing recorded footage. Video recording enables detailed post hoc analysis to identify bottlenecks and inefficiencies that may not be visible in real time. However, it suffers from delays in feedback, as the analysis only occurs after the fact. Additionally, storage and management of large video datasets can be challenging.
- Breaking Activities into Tasks and Subtasks This can help in understanding task performance and supports line balancing, especially in complex manufacturing processes. By decomposing activities into granular subtasks, manufacturers can identify specific areas for optimization; however, this approach is time-consuming and requires significant initial input to define tasks accurately. Additionally, real-time adaptability to changes in the manufacturing environment might be limited without advanced automation tools.
- Working with Predetermined Standard Times This approach offers deep insights into task performance by using historical data and industry standards to establish benchmarks. While predetermined times provide a solid foundation for efficiency analysis, they may not align perfectly with actual timings due to variability in human and machine performance. Moreover, this method may not capture the nuances of novel or highly customized manufacturing processes, requiring continuous updating of standard times.
- Sensor-based Tracking This involves using data from Industrial Internet of Things (IIoT) sensors and workflow systems for real-time productivity analysis, the results of which can be fed into predictive analytics models to forecast potential delays and optimize production schedules. However, while this method is efficient, it does b not provide insights into the root causes of productivity changes, as sensors typically provide raw data which lack context. Integration with other data sources such as quality control systems is necessary in order to gain a comprehensive understanding.
- Visual Tracking This method combines non-intrusive real-time data collection with identification of improvement opportunities by visual tracking of objects and processes. In smart manufacturing scenarios, visual tracking can be implemented through advanced computer vision systems that monitor the production line, providing real-time feedback to operators and management. These systems can detect anomalies, track the flow of goods, and even analyze worker movements for ergonomic improvements. Although the upfront investment in visual tracking technology can be substantial, it is increasingly becoming cost-effective due to advancements in AI and ML. Additionally, the integration of visual tracking with other smart manufacturing systems can lead to a more holistic view of production efficiency.
4. Methodology
4.1. Dataset Description
4.2. Training
4.3. Evaluation Metrics
4.4. YOLOv8 Model Performance
4.5. Complete Pipeline
5. Insights into Manufacturing Facility
Challenges
6. Conclusions
6.1. Implementation of the Results in Practice
6.2. Limitations of the Proposed Framework
6.3. Future Direction
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- World Bank. World Bank Open Data; Manufacturing: Value Added (% of GDP); World Bank: Washington, DC, USA, 2023. [Google Scholar]
- Causa, O.; Abendschein, M.; Luu, N.; Soldani, E.; Soriolo, C. The post-COVID-19 rise in labour shortages. OECD Econ. Dep. Work. Pap. 2022. [Google Scholar] [CrossRef]
- Canaj, K.; Sood, S.; Johnston, C. Analysis on Labour Challenges in Canada, Second Quarter of 2023; Government of Canada, Statistics Canada: Ottawa, ON, Canada, 2023. [Google Scholar]
- Bomal, L.A. Labour Shortage CFIB. 2023. Available online: https://www.cfib-fcei.ca/en/media/labour-shortages-cost-ontario-small-businesses-over-16b-in-lost-revenue (accessed on 25 January 2024).
- Gervasi, R.; Barravecchia, F.; Mastrogiacomo, L.; Franceschini, F. Applications of affective computing in human-robot interaction: State-of-art and challenges for manufacturing. Proc. Inst. Mech. Eng. Part B J. Eng. Manuf. 2023, 237, 815–832. [Google Scholar] [CrossRef]
- Poudel, L.; Elagandula, S.; Zhou, W.; Sha, Z. Decentralized and centralized planning for multi-robot additive manufacturing. J. Mech. Des. 2023, 145, 012003. [Google Scholar] [CrossRef]
- Liu, L.; Zou, Z.; Greene, R.L. The Effects of Type and Form of Collaborative Robots in Manufacturing on Trustworthiness, Risk Perceived, and Acceptance. Int. J. Hum.-Comput. Interact. 2023, 40, 2697–2710. [Google Scholar] [CrossRef]
- Pansara, R. From Fields to Factories A Technological Odyssey in Agtech and Manufacturing. Int. J. Manag. Educ. Sustain. Dev. 2023, 6, 1–12. [Google Scholar]
- Ahmad, H.M.; Rahimi, A. Deep learning methods for object detection in smart manufacturing: A survey. J. Manuf. Syst. 2022, 64, 181–196. [Google Scholar] [CrossRef]
- Puttemans, S.; Callemein, T.; Goedemé, T. Building Robust Industrial Applicable Object Detection Models using Transfer Learning and Single Pass Deep Learning Architectures. In Proceedings of the 13th International Joint Conference on Computer Vision, Imaging and Computer Graphics Theory and Applications (VISIGRAPP 2018), Funchal, Portugal, 27–29 January 2018; pp. 209–217. [Google Scholar] [CrossRef]
- Wang, J.; Fu, P.; Gao, R.X. Machine vision intelligence for product defect inspection based on deep learning and Hough transform. J. Manuf. Syst. 2019, 51, 52–60. [Google Scholar] [CrossRef]
- Redmon, J.; Farhadi, A. YOLO9000: Better, Faster, Stronger. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 6517–6525. [Google Scholar] [CrossRef]
- Farahnakian, F.; Koivunen, L.; Mäkilä, T.; Heikkonen, J. Towards Autonomous Industrial Warehouse Inspection. In Proceedings of the 2021 26th International Conference on Automation and Computing (ICAC), Portsmouth, UK, 2–4 September 2021; pp. 1–6. [Google Scholar] [CrossRef]
- Li, T.; Huang, B.; Li, C.; Huang, M. Application of convolution neural network object detection algorithm in logistics warehouse. J. Eng. 2019, 2019, 9053–9058. [Google Scholar] [CrossRef]
- Jocher, G.; Chaurasia, A.; Qiu, J. YOLO by Ultralytics. 2023. Available online: https://github.com/ultralytics/ultralytics (accessed on 20 January 2024).
- Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich Feature Hierarchies for Accurate Object Detection and Semantic Segmentation. In Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 28 June 2014; pp. 580–587. [Google Scholar] [CrossRef]
- Girshick, R. Fast r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar]
- Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef]
- Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. SSD: Single Shot MultiBox Detector. In Proceedings of the Computer Vision—ECCV, Amsterdam, The Netherlands, 11–14 October 2016; Leibe, B., Matas, J., Sebe, N., Welling, M., Eds.; Springer: Cham, Switzerland, 2016; pp. 21–37. [Google Scholar]
- Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You Only Look Once: Unified, Real-Time Object Detection. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar] [CrossRef]
- Terven, J.; Cordova-Esparza, D. A Comprehensive Review of YOLO: From YOLOv1 and Beyond. arXiv 2023, arXiv:2304.00501. [Google Scholar] [CrossRef]
- Redmon, J.; Farhadi, A. YOLOv3: An Incremental Improvement. arXiv 2018, arXiv:804.02767. [Google Scholar]
- Lin, T.Y.; Dollar, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature Pyramid Networks for Object Detection. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 26 July 2017; pp. 936–944. [Google Scholar] [CrossRef]
- Bochkovskiy, A.; Wang, C.Y.; Liao, H.Y.M. YOLOv4: Optimal Speed and Accuracy of Object Detection. arXiv 2020, arXiv:abs/2004.10934. [Google Scholar]
- Wang, C.Y.; Mark Liao, H.Y.; Wu, Y.H.; Chen, P.Y.; Hsieh, J.W.; Yeh, I.H. CSPNet: A New Backbone that can Enhance Learning Capability of CNN. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Seattle, WA, USA, 14–19 June 2020; pp. 1571–1580. [Google Scholar] [CrossRef]
- Liu, S.; Qi, L.; Qin, H.; Shi, J.; Jia, J. Path Aggregation Network for Instance Segmentation. arXiv 2018, arXiv:1803.01534. [Google Scholar]
- Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. Cbam: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
- Misra, D. Mish: A Self Regularized Non-Monotonic Activation Function. arXiv 2020, arXiv:1908.08681. [Google Scholar]
- Zheng, Z.; Wang, P.; Liu, W.; Li, J.; Ye, R.; Ren, D. Distance-IoU loss: Faster and better learning for bounding box regression. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 12993–13000. [Google Scholar]
- Jocher, G.; Stoken, A.; Borovec, J.; NanoCode012; ChristopherSTAN; Liu, C.; Laughing; tkianai; Hogan, A.; lorenzomammana; et al. ultralytics/yolov5: v3.1—Bug Fixes and Performance Improvements. 2020. Available online: https://zenodo.org/records/4154370 (accessed on 20 January 2024).
- Elfwing, S.; Uchibe, E.; Doya, K. Sigmoid-Weighted Linear Units for Neural Network Function Approximation in Reinforcement Learning. arXiv 2017, arXiv:1702.03118. [Google Scholar] [CrossRef]
- Li, C.; Li, L.; Jiang, H.; Weng, K.; Geng, Y.; Li, L.; Ke, Z.; Li, Q.; Cheng, M.; Nie, W.; et al. YOLOv6: A Single-Stage Object Detection Framework for Industrial Applications. arXiv 2022, arXiv:2209.02976. [Google Scholar]
- Wang, C.Y.; Bochkovskiy, A.; Liao, H.Y.M. YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 7464–7475. [Google Scholar]
- Wang, C.Y.; Liao, H.Y.M.; Yeh, I.H. Designing Network Design Strategies Through Gradient Path Analysis. arXiv 2022, arXiv:2211.04800. [Google Scholar]
- Ding, X.; Zhang, X.; Ma, N.; Han, J.; Ding, G.; Sun, J. Repvgg: Making vgg-style convnets great again. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 13733–13742. [Google Scholar]
- Ge, Z.; Liu, S.; Wang, F.; Li, Z.; Sun, J. YOLOX: Exceeding YOLO Series in 2021. arXiv 2021, arXiv:2107.08430. [Google Scholar]
- Li, X.; Wang, W.; Wu, L.; Chen, S.; Hu, X.; Li, J.; Tang, J.; Yang, J. Generalized focal loss: Learning qualified and distributed bounding boxes for dense object detection. Adv. Neural Inf. Process. Syst. 2020, 33, 21002–21012. [Google Scholar]
- Good, I.J. Rational decisions. J. R. Stat. Soc. Ser. B (Methodol.) 1952, 14, 107–114. [Google Scholar] [CrossRef]
- King, R. Brief Summary of YOLOv8 Model Structure. 2023. Available online: https://github.com/ultralytics/ultralytics/issues/189 (accessed on 11 March 2024).
- Zendehdel, N.; Chen, H.; Leu, M.C. Real-time tool detection in smart manufacturing using You-Only-Look-Once (YOLO)v5. Manuf. Lett. 2023, 35, 1052–1059. [Google Scholar] [CrossRef]
- Liu, M.; Chen, Y.; Xie, J.; He, L.; Zhang, Y. LF-YOLO: A lighter and faster yolo for weld defect detection of X-ray image. IEEE Sensors J. 2023, 23, 7430–7439. [Google Scholar] [CrossRef]
- Wang, J.; Dai, H.; Chen, T.; Liu, H.; Zhang, X.; Zhong, Q.; Lu, R. Toward surface defect detection in electronics manufacturing by an accurate and lightweight YOLO-style object detector. Sci. Rep. 2023, 13, 7062. [Google Scholar] [CrossRef] [PubMed]
- Zhao, C.; Shu, X.; Yan, X.; Zuo, X.; Zhu, F. RDD-YOLO: A modified YOLO for detection of steel surface defects. Measurement 2023, 214, 112776. [Google Scholar] [CrossRef]
- Vu, T.T.H.; Pham, D.L.; Chang, T.W. A YOLO-based Real-time Packaging Defect Detection System. Procedia Comput. Sci. 2023, 217, 886–894. [Google Scholar] [CrossRef]
- Zhao, Z.; Yang, X.; Zhou, Y.; Sun, Q.; Ge, Z.; Liu, D. Real-time detection of particleboard surface defects based on improved YOLOV5 target detection. Sci. Rep. 2021, 11, 21777. [Google Scholar] [CrossRef]
- Hu, J.; Shen, L.; Sun, G. Squeeze-and-Excitation Networks. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar] [CrossRef]
- Rahimi, A.; Anvaripour, M.; Hayat, K. Object Detection using Deep Learning in a Manufacturing Plant to Improve Manual Inspection. In Proceedings of the 2021 IEEE International Conference on Prognostics and Health Management (ICPHM), Detroit, MI, USA, 7–9 June 2021; pp. 1–7. [Google Scholar] [CrossRef]
- Ahmad, H.M.; Rahimi, A.; Hayat, K. Deep Learning Transforming the Manufacturing Industry: A Case Study. In Proceedings of the 2021 IEEE 23rd Int Conf on High Performance Computing andCommunications; 7th Int Conf on Data Science and Systems; 19th Int Conf on Smart City; 7th Int Conf on Dependability in Sensor, Cloud and Big Data Systems and Application (HPCC/DSS/SmartCity/DependSys), Haikou, China, 20–22 December 2021; pp. 1286–1291. [Google Scholar] [CrossRef]
- Liu, Z.; Ye, K. YOLO-IMF: An Improved YOLOv8 Algorithm for Surface Defect Detection in Industrial Manufacturing Field. In Proceedings of the Metaverse—-METAVERSE, Honolulu, HI, USA, 23 September 2023; He, S., Lai, J., Zhang, L.J., Eds.; Springer: Cham, Switzerland, 2023. Lecture Notes in Computer Science. pp. 15–28. [Google Scholar] [CrossRef]
- Luo, B.; Kou, Z.; Han, C.; Wu, J. A “Hardware-Friendly” Foreign Object Identification Method for Belt Conveyors Based on Improved YOLOv8. Appl. Sci. 2023, 13, 11464. [Google Scholar] [CrossRef]
- Ma, N.; Zhang, X.; Zheng, H.T.; Sun, J. Shufflenet v2: Practical guidelines for efficient cnn architecture design. In Proceedings of the European Conference on Computer Vision (ECCV), Tel Aviv, Israel, 23–27 October 2018; pp. 116–131. [Google Scholar]
- Ahmad, H.M.; Rahimi, A. SH17: A Dataset for Human Safety and Personal Protective Equipment Detection in Manufacturing Industry. arXiv 2024, arXiv:2407.04590. [Google Scholar]
- Krummenacher, G.; Ong, C.S.; Koller, S.; Kobayashi, S.; Buhmann, J.M. Wheel Defect Detection with Machine Learning. IEEE Trans. Intell. Transp. Syst. 2018, 19, 1176–1187. [Google Scholar] [CrossRef]
- O’Brien, K.; Humphries, J. Object Detection using Convolutional Neural Networks for Smart Manufacturing Vision Systems in the Medical Devices Sector. Procedia Manuf. 2019, 38, 142–147. [Google Scholar] [CrossRef]
- Wei, H.; Laszewski, M.; Kehtarnavaz, N. Deep Learning-Based Person Detection and Classification for Far Field Video Surveillance. In Proceedings of the 2018 IEEE 13th Dallas Circuits and Systems Conference (DCAS), Dallas, TX, USA, 12 November 2018; pp. 1–4. [Google Scholar] [CrossRef]
- Luo, C.; Yu, L.; Yang, E.; Zhou, H.; Ren, P. A benchmark image dataset for industrial tools. Pattern Recognit. Lett. 2019, 125, 341–348. [Google Scholar] [CrossRef]
- Goldratt, E.M.; Cox, J. The goal: A Process of Ongoing Improvement; Routledge: London, UK, 2016. [Google Scholar]
- Goldratt, E.M. Theory of Constraints; North River Croton-on-Hudson; The North River Press: Mt. Clemens, MI, USA, 1990. [Google Scholar]
- Christian Terwiesch. How to Measure and Improve Labor Productivity. Knowledge at Wharton. Available online: https://knowledge.wharton.upenn.edu/article/how-to-measure-and-improve-labor-productivity/ (accessed on 26 January 2024).
- Lin, T.Y.; Maire, M.; Belongie, S.; Bourdev, L.; Girshick, R.; Hays, J.; Perona, P.; Ramanan, D.; Zitnick, C.L.; Dollár, P. Microsoft COCO: Common Objects in Context. arXiv 2015, arXiv:1405.0312. [Google Scholar] [CrossRef]
- Neubeck, A.; Van Gool, L. Efficient Non-Maximum Suppression. In Proceedings of the 18th International Conference on Pattern Recognition (ICPR’06), Hong Kong, China, 20–24 August 2006; Volume 3, pp. 850–855. [Google Scholar] [CrossRef]
- Loshchilov, I.; Hutter, F. SGDR: Stochastic Gradient Descent with Warm Restarts. In Proceedings of the 5th International Conference on Learning Representations, ICLR 2017—Conference Track Proceedings, Toulon, France, 24–26 April 2017. [Google Scholar]
- You, Y.; Gitman, I.; Ginsburg, B. Large Batch Training of Convolutional Networks. arXiv 2017, arXiv:1708.03888. [Google Scholar]
- Bjorck, N.; Gomes, C.P.; Selman, B.; Weinberger, K.Q. Understanding Batch Normalization. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; Bengio, S., Wallach, H., Larochelle, H., Grauman, K., Cesa-Bianchi, N., Garnett, R., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2018; Volume 31. [Google Scholar]
- Powers, D. Evaluation: From Precision, Recall and F-Measure to ROC, Informedness, Markedness and Correlation. J. Mach. Learn. Technol. 2011, 2, 37–63. [Google Scholar]
- Wojke, N.; Bewley, A.; Paulus, D. Simple Online and Realtime Tracking with a Deep Association Metric. In Proceedings of the 2017 IEEE International Conference on Image Processing (ICIP), Beijing, China, 17–20 September 2017; pp. 3645–3649. [Google Scholar]
- Chui, C.K.; Chen, G. Kalman Filtering with Real-Time Applications. In Springer Series in Information Sciences; Springer: Berlin/Heidelberg, Germany, 1987. [Google Scholar] [CrossRef]
State | Symbol | Material | Worker |
---|---|---|---|
Productive | ✓ | ✓ | |
Non-productive | ✓ | ||
Downtime | ✓ | ||
Idle time |
Model | GPU (ms) | Parameters (M) | |
---|---|---|---|
YOLOv8-n | 18.4 | 1.21 | 3.5 |
YOLOv8-s | 27.7 | 1.40 | 11.4 |
YOLOv8-m | 33.6 | 2.26 | 26.2 |
YOLOv8-l | 34.9 | 2.43 | 44.1 |
YOLOv8-x | 36.3 | 3.56 | 68.7 |
Model Size | P (%) | R (%) | ||||
All | Worker | Chair | All | Worker | Chair | |
Nano | 89.2 | 84.4 | 93.9 | 87.1 | 86.0 | 88.2 |
Medium | 89.9 | 85.4 | 94.4 | 88.8 | 89.5 | 88.0 |
Large | 89.8 | 85.7 | 93.8 | 89.0 | 90.2 | 87.7 |
Model Size | mAP50 (%) | (%) | ||||
All | Worker | Chair | All | Worker | Chair | |
Nano | 92.6 | 91.7 | 93.5 | 64.7 | 64.8 | 64.6 |
Medium | 94.4 | 93.8 | 95.0 | 68.8 | 69.7 | 68.0 |
Large | 93.8 | 93.8 | 93.9 | 68.9 | 70.0 | 67.7 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Ahmad, H.M.; Rahimi, A.; Hayat, K. Capacity Constraint Analysis Using Object Detection for Smart Manufacturing. Automation 2024, 5, 545-563. https://doi.org/10.3390/automation5040031
Ahmad HM, Rahimi A, Hayat K. Capacity Constraint Analysis Using Object Detection for Smart Manufacturing. Automation. 2024; 5(4):545-563. https://doi.org/10.3390/automation5040031
Chicago/Turabian StyleAhmad, Hafiz Mughees, Afshin Rahimi, and Khizer Hayat. 2024. "Capacity Constraint Analysis Using Object Detection for Smart Manufacturing" Automation 5, no. 4: 545-563. https://doi.org/10.3390/automation5040031
APA StyleAhmad, H. M., Rahimi, A., & Hayat, K. (2024). Capacity Constraint Analysis Using Object Detection for Smart Manufacturing. Automation, 5(4), 545-563. https://doi.org/10.3390/automation5040031