Deep-Learning-Based Automated Building Construction Progress Monitoring for Prefabricated Prefinished Volumetric Construction
Abstract
:1. Introduction
- -
- Elimination of window detections that are erroneous or irrelevant to progress monitoring of the target building by developing a filtering process, which consists of building mask filtering and arrangement of detected windows into columns.
- -
- Prediction of missed window detections in the mid-section and near-ground regions of the constructed building via identification of anomalies within each window column.
2. Related Work
2.1. 3D Image-Based Solutions for Progress Monitoring
2.2. 2D Object Detection for Progress Monitoring
3. Methodology
3.1. WAVBCPM Overview
- (i)
- The detection module is responsible for localising windows on the target building within an input image. This is achieved by first conducting two-scale window detection, which is performed by extracting patches from the input image before conducting window detection on both the image and extracted patches. Output window detections are then concatenated together before duplicated detections are merged. The two-scale window detection approach is designed to improve window detection robustness by accounting for targeted windows at smaller scales. Next, the window detections are filtered using a building mask obtained from an instance segmentation model and arranged into columns via a vertical vectorisation algorithm to eliminate irrelevant and erroneous detections, such as in Figure 3a. The implemented filtering mechanisms are designed to mimic how humans are able to identify and focus on the monitored building based on its distinct features and repetitive construction patterns during manual progress monitoring. If no window columns were output, building construction is still at an early stage. In such cases, the number of detected windows is directly used to estimate building construction progress.
- (ii)
- As building construction is a ground-up process, a human would count assembled modules in the lower and mid-section regions of the constructed building even if they are occluded. This is reproduced algorithmically by sending the arranged window columns into the rectification module, which identifies and rectifies missed window detections due to occlusion or poor detection, as illustrated in Figure 3b. Missed window detections are classified as either mid-section or near-ground to be resolved separately. Note that missed window detections on the top-most storey of a constructed PPVC building were not considered for rectification because the top-most storey could still be under construction. Furthermore, occlusions that block the window components of PPVC modules are usually found only in the lower to middle regions of the constructed PPVC building. For the case of mid-section missing windows, erroneous regions are identified by significant gaps between windows in each column. Each erroneous gap is defined by the detected windows above and below it. For the case of near-ground missing windows, there are no detected windows below the erroneous region. To account for near-ground missing windows, a horizontal line vector is estimated to localise the lowest windowed storey of the constructed building. Using the obtained line vector, near-ground missing windows are identified and rectified based on the location of the bottom-most windows in each column.
- (iii)
- Lastly, the progress estimation module extracts the average number of windows per column from the rectified window columns to predict building construction progress. If building construction is near completion, end-stage detection is conducted to bypass any residual detection errors or occlusion. This is achieved by identifying the roof components that are installed along with the top-most PPVC modules on a completed building.
3.2. Detection Module
3.2.1. Two-Scale Window Detection
3.2.2. Window Detection Filtering
Algorithm 1. Window detection filtering pseudocode |
- (i)
- Vectorisation check 1: A median box size is derived for each column by arranging the areas of its bounding boxes in ascending order and selecting the area value in the middle. If there are an even number of predicted window columns, then an average of the two areas in the middle is used. Bounding boxes within the column that deviate from the derived median box size by more than 75% are eliminated.
- (ii)
- Vectorisation check 2: The median box size and the median gap height from each column are used to derive an overall median box and gap height in a similar way to measure 1. If both the median box and gap height for a predicted column differ by more than 50% from the derived overall median box and gap height, the predicted window column is likely detected from a nearby building in the background and is thus eliminated. Figure 7 illustrates a scenario where irrelevant window detections that were unable to be filtered by the predicted building mask were eliminated using the conducted vectorisation checks.
3.3. Rectification Module
3.3.1. Mid-Section Missing Window Rectification
3.3.2. Near-Ground Missing Window Rectification
3.4. Progress Estimation Module
3.4.1. End-Stage Detection
3.4.2. Window Columns Check and Storey-Based PPVC Progress Estimation
4. Evaluating the Progress Estimation of WAVBCPM
4.1. Experiment Setup
4.2. Results
4.3. Discussion
5. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- Yang, J.; Park, M.-W.; Vela, P.A.; Golparvar-Fard, M. Construction performance monitoring via still images, time-lapse photos, and video streams: Now, tomorrow, and the future. Adv. Eng. Inform. 2015, 29, 211–224. [Google Scholar] [CrossRef]
- Abioye, S.O.; Oyedele, L.O.; Akanbi, L.; Ajayi, A.; Davila Delgado, J.M.; Bilal, M.; Akinade, O.O.; Ahmed, A. Artificial intelligence in the construction industry: A review of present status, opportunities and future challenges. J. Build. Eng. 2021, 44, 103299. [Google Scholar] [CrossRef]
- Son, H.; Kim, C.; Cho, Y. Automated Schedule Updates Using As-Built Data and a 4D Building Information Model. J. Manag. Eng. 2017, 33, 04017012. [Google Scholar] [CrossRef]
- Han, K.K.; Golparvar-Fard, M. Potential of big visual data and building information modeling for construction performance analytics: An exploratory study. Autom. Constr. 2017, 73, 184–198. [Google Scholar] [CrossRef]
- Fang, Q.; Li, H.; Luo, X.; Ding, L.; Luo, H.; Rose, T.M.; An, W. Detecting non-hardhat-use by a deep learning method from far-field surveillance videos. Autom. Constr. 2018, 85, 1–9. [Google Scholar] [CrossRef]
- Fang, W.; Ding, L.; Zhong, B.; Love, P.E.D.; Luo, H. Automated detection of workers and heavy equipment on construction sites: A convolutional neural network approach. Adv. Eng. Inform. 2018, 37, 139–149. [Google Scholar] [CrossRef]
- Ilyas, M.; Khaw, H.Y.; Selvaraj, N.M.; Jin, Y.; Zhao, X.; Cheah, C.C. Robot-Assisted Object Detection for Construction Automation: Data and Information-Driven Approach. IEEE/ASME Trans. Mechatron. 2021, 26, 2845–2856. [Google Scholar] [CrossRef]
- Kim, M.-K.; Wang, Q.; Park, J.-W.; Cheng, J.C.P.; Sohn, H.; Chang, C.-C. Automated dimensional quality assurance of full-scale precast concrete elements using laser scanning and BIM. Autom. Constr. 2016, 72, 102–114. [Google Scholar] [CrossRef]
- Lee, D.; Nie, G.-Y.; Han, K. Vision-based inspection of prefabricated components using camera poses: Addressing inherent limitations of image-based 3D reconstruction. J. Build. Eng. 2023, 64, 105710. [Google Scholar] [CrossRef]
- Woldeamanuel, M.M.; Kim, T.; Cho, S.; Kim, H.-K. Estimation of concrete strength using thermography integrated with deep-learning-based image segmentation: Case studies and economic analysis. Expert Syst. Appl. 2023, 213, 119249. [Google Scholar] [CrossRef]
- Chen, W.; Li, C.; Guo, H. A lightweight face-assisted object detection model for welding helmet use. Expert Syst. Appl. 2023, 221, 119764. [Google Scholar] [CrossRef]
- Iannizzotto, G.; Lo Bello, L.; Patti, G. Personal Protection Equipment detection system for embedded devices based on DNN and Fuzzy Logic. Expert Syst. Appl. 2021, 184, 115447. [Google Scholar] [CrossRef]
- Seo, J.; Han, S.; Lee, S.; Kim, H. Computer vision techniques for construction safety and health monitoring. Adv. Eng. Inform. 2015, 29, 239–251. [Google Scholar] [CrossRef]
- Wang, Z.; Zhang, Q.; Yang, B.; Wu, T.; Lei, K.; Zhang, B.; Fang, T. Vision-Based Framework for Automatic Progress Monitoring of Precast Walls by Using Surveillance Videos during the Construction Phase. J. Comput. Civ. Eng. 2021, 35, 04020056. [Google Scholar] [CrossRef]
- Zheng, Z.; Zhang, Z.; Pan, W. Virtual prototyping- and transfer learning-enabled module detection for modular integrated construction. Autom. Constr. 2020, 120, 103387. [Google Scholar] [CrossRef]
- Hwang, B.-G.; Shan, M.; Looi, K.-Y. Key constraints and mitigation strategies for prefabricated prefinished volumetric construction. J. Clean. Prod. 2018, 183, 183–193. [Google Scholar] [CrossRef]
- Han, Y.; Zhu, W.-z. The Development of Modular Building in China. In Proceedings of the International Conference on Applied Mechanics, Electronics and Mechatronics Engineering, Beijing, China, 28–29 May 2016; pp. 204–207. [Google Scholar] [CrossRef]
- Pan, W.; Yang, Y.; Yang, L. High-Rise Modular Building: Ten-Year Journey and Future Development. In Proceedings of the Construction Research Congress 2018, New Orleans, LA, USA, 2–4 April 2018; pp. 523–532. [Google Scholar] [CrossRef]
- Razkenari, M.; Fenner, A.; Shojaei, A.; Hakim, H.; Kibert, C. Perceptions of offsite construction in the United States: An investigation of current practices. J. Build. Eng. 2020, 29, 101138. [Google Scholar] [CrossRef]
- Prefabricated Prefinished Volumetric Construction (PPVC). Available online: https://www1.bca.gov.sg/buildsg/productivity/design-for-manufacturing-and-assembly-dfma/prefabricated-prefinished-volumetric-construction-ppvc (accessed on 4 October 2024).
- Rahimian, P.F.; Seyedzadeh, S.; Oliver, S.; Rodriguez, S.; Dawood, N. On-demand monitoring of construction projects through a game-like hybrid application of BIM and machine learning. Autom. Constr. 2020, 110, 103012. [Google Scholar] [CrossRef]
- Alaloul, W.S.; Qureshi, A.H.; Musarat, M.A.; Saad, S. Evolution of close-range detection and data acquisition technologies towards automation in construction progress monitoring. J. Build. Eng. 2021, 43, 102877. [Google Scholar] [CrossRef]
- Ma, G.; Wu, M.; Wu, Z.; Yang, W. Single-shot multibox detector- and building information modeling-based quality inspection model for construction projects. J. Build. Eng. 2021, 38, 102216. [Google Scholar] [CrossRef]
- Tran, H.; Nguyen, T.N.; Christopher, P.; Bui, D.-K.; Khoshelham, K.; Ngo, T.D. A digital twin approach for geometric quality assessment of as-built prefabricated façades. J. Build. Eng. 2021, 41, 102377. [Google Scholar] [CrossRef]
- Omar, H.; Mahdjoubi, L.; Kheder, G. Towards an automated photogrammetry-based approach for monitoring and controlling construction site activities. Comput. Ind. 2018, 98, 172–182. [Google Scholar] [CrossRef]
- Bognot, J.; Candido, C.; Blanco, A.; Montelibano, J. Building Construction Progress Monitoring Using Unmanned Aerial System (UAS), Low-Cost Photogrammetry, and Geographic Information System (GIS). ISPRS Ann. Photogramm. Remote Sens. Spat. Inf. Sci. 2018, 4, 41–47. [Google Scholar] [CrossRef]
- Kopsida, M.; Brilakis, I.; Vela, P.A. A review of automated construction progress monitoring and inspection methods. In Proceedings of the 32nd International Conference of CIB W78, Eindhoven, The Netherlands, 27–29 October 2015; pp. 421–431. Available online: https://itc.scix.net/pdfs/w78-2015-paper-044.pdf (accessed on 30 April 2024).
- Aslan, M.F.; Durdu, A.; Sabanci, K.; Mutluer, M.A. CNN and HOG based comparison study for complete occlusion handling in human tracking. Measurement 2020, 158, 107704. [Google Scholar] [CrossRef]
- Wei, W.; Lu, Y.; Zhong, T.; Li, P.; Liu, B. Integrated vision-based automated progress monitoring of indoor construction using mask region-based convolutional neural networks and BIM. Autom. Constr. 2022, 140, 104327. [Google Scholar] [CrossRef]
- Li, Z.; Li, D. Action recognition of construction workers under occlusion. J. Build. Eng. 2022, 45, 103352. [Google Scholar] [CrossRef]
- Chen, K.; Reichard, G.; Xu, X.; Akanmu, A. Automated crack segmentation in close-range building façade inspection images using deep learning techniques. J. Build. Eng. 2021, 43, 102913. [Google Scholar] [CrossRef]
- Ultralytics/Yolov5: v6.1-TensorRT, TensorFlow Edge TPU and OpenVINO Export and Inference. Zenodo. 2022. Available online: https://zenodo.org/records/6222936 (accessed on 13 June 2024).
- BTO Top Tracker. Available online: https://www.btohq.com/bto-top-tracker (accessed on 4 June 2024).
- LabelImg. Tzutalin. 2015. Available online: https://github.com/tzutalin/labelImg (accessed on 1 July 2024).
- Labelme: Image Polygonal Annotation with Python. K. Wada. 2018. Available online: https://github.com/wkentaro/labelme (accessed on 1 June 2024).
Test set A (40 images) | |||
Method | TP | FP | FN |
POD (YOLOv5-S) | 4709 | 1119 | 530 |
POD (YOLOv5-M) | 4765 | 1227 | 474 |
POD (YOLOv5-L) | 4771 | 1168 | 468 |
POD (Cascade Mask R-CNN (Swin-T) | 4110 | 594 | 1129 |
POD (Cascade Mask R-CNN (Swin-S) | 3990 | 500 | 1249 |
POD (Cascade Mask R-CNN (Swin-B) | 4240 | 660 | 999 |
WAVBCPM | 5180 | 101 | 59 |
Test set B (60 images) | |||
Method | TP | FP | FN |
POD (YOLOv5-S) | 5134 | 1011 | 452 |
POD (YOLOv5-M) | 5152 | 1041 | 434 |
POD (YOLOv5-L) | 5107 | 1058 | 479 |
POD (Cascade Mask R-CNN (Swin-T) | 4791 | 762 | 795 |
POD (Cascade Mask R-CNN (Swin-S) | 4722 | 667 | 864 |
POD (Cascade Mask R-CNN (Swin-B) | 4872 | 833 | 714 |
WAVBCPM | 5504 | 226 | 82 |
Test set A (40 images) | |||||
Method | Absolute Deviation | Time Taken Per Image | |||
Max (%) | Mean (%) | Min (s) | Max (s) | Mean (s) | |
POD (YOLOv5-S) | 56.25 | 10.07 | 0.03 | 0.54 | 0.33 |
POD (YOLOv5-M) | 56.25 | 9.69 | 0.04 | 0.55 | 0.33 |
POD (YOLOv5-L) | 56.25 | 9.74 | 0.04 | 0.55 | 0.34 |
POD (Cascade Mask R-CNN (Swin-T) | 65.99 | 13.20 | 0.13 | 3.00 | 0.65 |
POD (Cascade Mask R-CNN (Swin-S) | 65.99 | 13.20 | 0.15 | 2.78 | 0.65 |
POD (Cascade Mask R-CNN (Swin-B) | 61.95 | 13.02 | 0.18 | 3.04 | 0.78 |
WAVBCPM | 2.27 | 0.14 | 1.16 | 15.19 | 6.86 |
Test set B (60 images) | |||||
Method | Absolute Deviation | Time Taken Per Image | |||
Max (%) | Mean (%) | Min (s) | Max (%) | Mean (%) | |
POD (YOLOv5-S) | 58.34 | 9.08 | 0.02 | 0.14 | 0.07 |
POD (YOLOv5-M) | 63.89 | 8.56 | 0.02 | 0.17 | 0.08 |
POD (YOLOv5-L) | 72.22 | 9.23 | 0.02 | 0.15 | 0.08 |
POD (Cascade Mask R-CNN (Swin-T) | 77.78 | 11.02 | 0.11 | 0.81 | 0.24 |
POD (Cascade Mask R-CNN (Swin-S) | 62.5 | 10.78 | 0.13 | 0.8 | 0.26 |
POD (Cascade Mask R-CNN (Swin-B) | 77.78 | 10.28 | 0.14 | 0.82 | 0.29 |
WAVBCPM | 2.60 | 0.26 | 0.78 | 8.74 | 2.42 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Chua, W.P.; Cheah, C.C. Deep-Learning-Based Automated Building Construction Progress Monitoring for Prefabricated Prefinished Volumetric Construction. Sensors 2024, 24, 7074. https://doi.org/10.3390/s24217074
Chua WP, Cheah CC. Deep-Learning-Based Automated Building Construction Progress Monitoring for Prefabricated Prefinished Volumetric Construction. Sensors. 2024; 24(21):7074. https://doi.org/10.3390/s24217074
Chicago/Turabian StyleChua, Wei Png, and Chien Chern Cheah. 2024. "Deep-Learning-Based Automated Building Construction Progress Monitoring for Prefabricated Prefinished Volumetric Construction" Sensors 24, no. 21: 7074. https://doi.org/10.3390/s24217074
APA StyleChua, W. P., & Cheah, C. C. (2024). Deep-Learning-Based Automated Building Construction Progress Monitoring for Prefabricated Prefinished Volumetric Construction. Sensors, 24(21), 7074. https://doi.org/10.3390/s24217074