Automated Building Detection from Airborne LiDAR and Very High-Resolution Aerial Imagery with Deep Neural Network
Abstract
:1. Introduction
2. Related Work
3. Study Area, Materials, and Methods
3.1. Study Area
3.2. Data Used
3.3. Methodology Applied
3.3.1. Overview of CNN
3.3.2. The Proposed Architecture of CNN
Deep Feature Extraction
Preprocessing
Implementation Network
Training and Testing Process
Quantitative Evaluation
4. Results and Discussion
4.1. Experimental Scenes and LiDAR Point Clouds
4.2. CNN Building Extraction Results
5. Conclusions
Author Contributions
Funding
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- Al-Najjar, H.A.H.; Kalantar, B.; Pradhan, B.; Saeidi, V. Land Cover Classification from fused DSM and UAV Images Using Convolutional Neural Networks. Remote Sens. 2019, 11, 1461. [Google Scholar] [CrossRef] [Green Version]
- Pradhan, B.; Al-Najjar, H.A.H.; Sameen, M.I.; Tsang, I.; Alamri, A.M. Unseen land cover classification fromhigh-resolution orthophotos using integration of zero-shot learning and convolutional neural networks. Remote Sens. 2020, 12, 1676. [Google Scholar] [CrossRef]
- Kalantar, B.; Ueda, N.; Al-Najjar, H.A.H.; Halin, A.A. Assessment of convolutional neural network architectures for earthquake-induced building damage detection based on pre-and post-event orthophoto images. Remote Sens. 2020, 12, 3529. [Google Scholar] [CrossRef]
- Dong, Y.; Zhang, L.; Cui, X.; Ai, H.; Xu, B. Extraction of buildings from multiple-view aerial images using a feature-level-fusion strategy. Remote Sens. 2018, 10, 1947. [Google Scholar] [CrossRef] [Green Version]
- Rottensteiner, F.; Sohn, G.; Gerke, M.; Wegner, J.D.; Breitkopf, U.; Jung, J. Results of the ISPRS benchmark on urban object detection and 3D building reconstruction. ISPRS J. Photogramm. Remote Sens. 2014, 93, 256–271. [Google Scholar] [CrossRef]
- Alidoost, F.; Arefi, H. A CNN-Based Approach for Automatic Building Detection and Recognition of Roof Types Using a Single Aerial Image. PFG J. Photogramm. Remote Sens. Geoinf. Sci. 2018, 86, 235–248. [Google Scholar] [CrossRef]
- Gibril, M.B.A.; Kalantar, B.; Al-Ruzouq, R.; Ueda, N.; Saeidi, V.; Shanableh, A.; Mansor, S.; Shafri, H.Z.M. Mapping heterogeneous urban landscapes from the fusion of digital surface model and unmanned aerial vehicle-based images using adaptive multiscale image segmentation and classification. Remote Sens. 2020, 12, 1081. [Google Scholar] [CrossRef] [Green Version]
- Grigillo, D.; Kanjir, U. Urban Object Extraction from Digital Surface Model and Digital Aerial Images. ISPRS Ann. Photogramm. Remote Sens. Spat. Inf. Sci. 2012, 3, 215–220. [Google Scholar] [CrossRef] [Green Version]
- Tomljenovic, I.; Höfle, B.; Tiede, D.; Blaschke, T. Building extraction from Airborne Laser Scanning data: An analysis of the state of the art. Remote Sens. 2015, 7, 3826–3862. [Google Scholar] [CrossRef] [Green Version]
- Fritsch, D.; Klein, M.; Gressin, A.; Mallet, C.; Demantké, J.Ô.; David, N.; Heo, J.; Jeong, S.; Park, H.H.K.; Jung, J.; et al. Generating 3D city models without elevation data. ISPRS J. Photogramm. Remote Sens. 2017, 64, 1–18. [Google Scholar] [CrossRef] [Green Version]
- Yang, H.; Wu, P.; Yao, X.; Wu, Y.; Wang, B.; Xu, Y. Building extraction in very high resolution imagery by dense-attention networks. Remote Sens. 2018, 10, 1768. [Google Scholar] [CrossRef] [Green Version]
- Fernandes, D.; Silva, A.; Névoa, R.; Simões, C.; Gonzalez, D.; Guevara, M.; Novais, P.; Monteiro, J.; Melo-Pinto, P. Point-cloud based 3D object detection and classification methods for self-driving applications: A survey and taxonomy. Inf. Fusion 2021, 68, 161–191. [Google Scholar] [CrossRef]
- Zhou, Z.; Gong, J. Automated residential building detection from airborne LiDAR data with deep neural networks. Adv. Eng. Inform. 2018, 36, 229–241. [Google Scholar] [CrossRef]
- Park, Y.; Guldmann, J.-M.M. Creating 3D city models with building footprints and LIDAR point cloud classification: A machine learning approach. Comput. Environ. Urban Syst. 2019, 75, 76–89. [Google Scholar] [CrossRef]
- Chang, Z.; Yu, H.; Zhang, Y.; Wang, K. Fusion of hyperspectral CASI and airborne LiDAR data for ground object classification through residual network. Sensors 2020, 20, 3961. [Google Scholar] [CrossRef] [PubMed]
- Elsayed, H.; Zahran, M.; Elshehaby, A.R.; Salah, M.; Elshehaby, A. Integrating Modern Classifiers for Improved Building Extraction from Aerial Imagery and LiDAR Data. Am. J. Geogr. Inf. Syst. 2019, 2019, 213–220. [Google Scholar] [CrossRef]
- Xie, Y.; Zhu, J.; Cao, Y.; Feng, D.; Hu, M.; Li, W.; Zhang, Y.; Fu, L. Refined Extraction of Building Outlines from High-Resolution Remote Sensing Imagery Based on a Multifeature Convolutional Neural Network and Morphological Filtering. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 13, 1842–1855. [Google Scholar] [CrossRef]
- Awrangjeb, M.; Ravanbakhsh, M.; Clive, S.F. Automatic detection of residential buildings using LIDAR data and multispectral imagery. ISPRS Jounal Photogramm. Remote Sens. 2013, 53, 1689–1699. [Google Scholar] [CrossRef] [Green Version]
- Gilani, S.A.N.; Awrangjeb, M.; Lu, G. Fusion of LiDAR data and multispectral imagery for effective building detection based on graph and connected component analysis. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2015, 40, 65–72. [Google Scholar] [CrossRef] [Green Version]
- Maltezos, E.; Doulamis, A.; Doulamis, N.; Ioannidis, C. Building extraction from LiDAR data applying deep convolutional neural networks. IEEE Geosci. Remote Sens. Lett. 2019, 16, 155–159. [Google Scholar] [CrossRef]
- Chen, C.; Gong, W.; Hu, Y.; Chen, Y.; Ding, Y.; Vi, C.; Vi, W.G. Learning Oriented Region-based Convolutional Neural Networks for Building Detection in Satellite Remote Sensing Images. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2017, XLII, 6–9. [Google Scholar] [CrossRef] [Green Version]
- Wierzbicki, D.; Matuk, O.; Bielecka, E. Polish cadastre modernization with remotely extracted buildings from high-resolution aerial orthoimagery and airborne LiDAR. Remote Sens. 2021, 13, 611. [Google Scholar] [CrossRef]
- Lu, T.; Ming, D.; Lin, X.; Hong, Z.; Bai, X.; Fang, J. Detecting building edges from high spatial resolution remote sensing imagery using richer convolution features network. Remote Sens. 2018, 10, 1496. [Google Scholar] [CrossRef] [Green Version]
- Li, W.; Liu, H.; Wang, Y.; Li, Z.; Jia, Y.; Gui, G. Deep Learning-Based Classification Methods for Remote Sensing Images in Urban Built-Up Areas. IEEE Access 2019, 7, 36274–36284. [Google Scholar] [CrossRef]
- Lai, X.; Yang, J.; Li, Y.; Wang, M. A building extraction approach based on the fusion of LiDAR point cloud and elevation map texture features. Remote Sens. 2019, 11, 1636. [Google Scholar] [CrossRef] [Green Version]
- Wen, C.; Yang, L.; Li, X.; Peng, L.; Chi, T. Directionally constrained fully convolutional neural network for airborne LiDAR point cloud classification. ISPRS J. Photogramm. Remote Sens. 2020, 162, 50–62. [Google Scholar] [CrossRef] [Green Version]
- Al-Najjar, H.A.H.; Pradhan, B.; Sarkar, R.; Beydoun, G.; Alamri, A. A New Integrated Approach for Landslide Data Balancing and Spatial Prediction Based on Generative Adversarial. Remote Sens. 2021, 13, 4011. [Google Scholar] [CrossRef]
- Al-Najjar, H.A.H.; Pradhan, B. Spatial landslide susceptibility assessment using machine learning techniques assisted by additional data created with generative adversarial networks. Geosci. Front. 2021, 12, 625–637. [Google Scholar] [CrossRef]
- Mousa, Y.A.; Helmholz, P.; Belton, D.; Bulatov, D. Building detection and regularisation using DSM and imagery information. Photogramm. Rec. 2019, 34, 85–107. [Google Scholar] [CrossRef] [Green Version]
- Sherrah, J. Fully Convolutional Networks for Dense Semantic Labelling of High-Resolution Aerial Imagery. arXiv 2016, arXiv:Abs/1606.02585. [Google Scholar]
- Wen, Q.; Jiang, K.; Wang, W.; Liu, Q.; Guo, Q.; Li, L.; Wang, P. Automatic building extraction from google earth images under complex backgrounds based on deep instance segmentation network. Sensors 2019, 19, 333. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Ghamisi, P.; Li, H.; Soergel, U.; Zhu, X.X. Hyperspectral and LiDAR fusion using deep three-stream convolutional neural networks. Remote Sens. 2018, 10, 1649. [Google Scholar] [CrossRef] [Green Version]
- Xie, L.; Zhu, Q.; Hu, H.; Wu, B.; Li, Y.; Zhang, Y.; Zhong, R. Hierarchical regularization of building boundaries in noisy aerial laser scanning and photogrammetric point clouds. Remote Sens. 2018, 10, 1996. [Google Scholar] [CrossRef] [Green Version]
- Bittner, K.; Adam, F.; Cui, S.; Körner, M.; Reinartz, P. Building Footprint Extraction From VHR Remote Sensing Images Combined With Normalized DSMs Using Fused Fully Convolutional Networks. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2018, 11, 2615–2629. [Google Scholar] [CrossRef] [Green Version]
- Gilani, S.A.N.; Awrangjeb, M.; Lu, G. An Automatic Building Extraction and Regularisation Technique Using LiDAR Point Cloud Data and Orthoimage. Remote Sens. 2016, 8, 258. [Google Scholar] [CrossRef] [Green Version]
- Nahhas, F.H.; Shafri, H.Z.M.; Sameen, M.I.; Pradhan, B.; Mansor, S. Deep Learning Approach for Building Detection Using LiDAR-Orthophoto Fusion. J. Sens. 2018, 2018, 1–12. [Google Scholar] [CrossRef] [Green Version]
- Li, D.; Shen, X.; Yu, Y.; Guan, H.; Li, J.; Zhang, G.; Li, D. Building extraction from airborne multi-spectral LiDAR point clouds based on graph geometric moments convolutional neural networks. Remote Sens. 2020, 12, 3186. [Google Scholar] [CrossRef]
- Zhang, L.; Li, Z.; Li, A.; Liu, F. Large-scale urban point cloud labeling and reconstruction. ISPRS J. Photogramm. Remote. Sens. 2018, 138, 86–100. [Google Scholar] [CrossRef]
- Seydi, S.T.; Hasanlou, M.; Amani, M. A new end-to-end multi-dimensional CNN framework for land cover/land use change detection in multi-source remote sensing datasets. Remote Sens. 2020, 12, 2010. [Google Scholar] [CrossRef]
- Seydi, S.T.; Rastiveis, H. A Deep Learning Framework for Roads Network Damage Assessment Using Post-Earthquake Lidar Data. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2019, XLII, 12–14. [Google Scholar] [CrossRef] [Green Version]
- Seydi, S.T.; Hasanlou, M. A New Structure for Binary and Multiple Hyperspectral Change Detection Based on Spectral Unmixing and Convolutional Neural Network. Measurement 2021, 186, 110137. [Google Scholar] [CrossRef]
- Seydi, S.T.; Hasanlou, M.; Amani, M.; Huang, W. Oil Spill Detection Based on Multiscale Multidimensional Residual CNN for Optical Remote Sensing Imagery. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 10941–10952. [Google Scholar] [CrossRef]
Classes | Non-Building | Building | Total | User Accuracy | Commission Error |
---|---|---|---|---|---|
(a) Area1 (DSM) | |||||
Non-Building | 96.45% | 18.24% | 79.19% | 94.92% | 5.08% |
Building | 3.55% | 81.76% | 20.81% | 86.70% | 13.30% |
Total | 100.00% | 100.00% | 100.00% | ||
Producer’s Accuracy | 96.45% | 81.76% | OA | Kappa Coefficient | |
Omission Error | 3.55% | 18.24% | 93.20% | 0.798 | |
(b) Area1 (Optical) | |||||
Non-Building | 92.80% | 31.05% | 79.17% | 91.35% | 8.65% |
Building | 7.20% | 68.95% | 20.83 | 73.06% | 26.94% |
Total | 100% | 100% | 100% | ||
Producer’s Accuracy | 92.80% | 68.95% | OA | Kappa Coefficient | |
Omission Error | 7.20% | 31.05% | 87.53% | 0.630 | |
(c) Area1 (Fusion) | |||||
Non-Building | 96.45% | 18.24% | 79.19% | 94.92% | 5.08% |
Building | 3.55% | 81.76% | 20.81% | 86.70% | 13.30% |
Total | 100% | 100% | 100% | ||
Producer’s Accuracy | 96.45% | 81.76 % | OA | Kappa Coefficient | |
Omission Error | 3.55% | 18.24% | 93.20% | 0.787 |
Classes | Non-Building | Building | Total | User Accuracy | Commission Error |
---|---|---|---|---|---|
(a) Area2 (DSM) | |||||
Non-Building | 97.76% | 2.61% | 79.25% | 90.80% | 9.20% |
Building | 2.24% | 72.39% | 20.75% | 92.06% | 7.94% |
Total | 100% | 100% | 100% | ||
Producer’s Accuracy | 97.76% | 72.39% | OA | Kappa Coefficient | |
Omission Error | 2.24% | 27.61% | 91.06% | 0.753 | |
(b) Area2 (Optical) | |||||
Non-Building | 84.56% | 19.29% | 67.33% | 92.44% | 7.56% |
Building | 15.44% | 80.71% | 32.67% | 65.20% | 34.80% |
Total | 100% | 100% | 100% | ||
Producer’s Accuracy | 84.56% | 80.71% | OA | Kappa Coefficient | |
Omission Error | 15.44% | 19.29% | 83.5407% | 0.606 | |
(c) Area2 (Fusion) | |||||
Non-Building | 93.72% | 4.10% | 70.07% | 98.46% | 1.54% |
Building | 6.28% | 95.90% | 29.93% | 84.56% | 15.44% |
Total | 100% | 100% | 100% | ||
Producer’s Accuracy | 93.72% | 95.90% | OA | Kappa Coefficient | |
Omission Error | 6.28% | 4.10% | 94.29% | 0.859 |
Classes | Non-Building | Building | Total | User Accuracy | Commission Error |
---|---|---|---|---|---|
(a) Area3 (DSM) | |||||
Non-Building | 85.30% | 7.43% | 57.24% | 95.32% | 4.68% |
Building | 14.70% | 92.57% | 42.76% | 78.00% | 22.00% |
Total | 100% | 100% | 100% | ||
Producer’s Accuracy | 85.30% | 92.57% | OA | Kappa Coefficient | |
Omission Error | 14.70% | 7.43% | 87.91% | 0.748 | |
(b) Area3 (Optical) | |||||
Non-Building | 83.73% | 23.00% | 57.77% | 82.97% | 17.03% |
Building | 16.27% | 77.00% | 42.23% | 77.95% | 22.05% |
Total | 100% | 100% | 100% | ||
Producer’s Accuracy | 83.73% | 77.00% | OA | Kappa Coefficient | |
Omission Error | 16.27% | 23.00% | 80.85% | 0.608 | |
(c) Area3 (Fusion) | |||||
Non-Building | 92.51% | 8.98% | 56.79% | 93.24% | 6.76% |
Building | 7.49% | 91.02% | 43.21% | 90.07% | 9.93% |
Total | 100% | 100% | 100% | ||
Producer’s Accuracy | 92.51% | 91.02% | OA | Kappa Coefficient | |
Omission Error | 7.49% | 8.98% | 91.8715% | 0.834 |
Datasets/Study Area | A1 | A2 | A3 | |||
---|---|---|---|---|---|---|
OA | Kappa | OA | Kappa | OA | Kappa | |
DSM | 93.21% | 0.798 | 91.07% | 0.753 | 87.92% | 0.748 |
Optical | 87.54% | 0.630 | 83.54% | 0.606 | 80.85% | 0.608 |
Fusion | 93.21% | 0.798 | 94.30% | 0.859 | 91.87% | 0.834 |
Study Area | Fusion | Result Using Morphology | Discussion |
---|---|---|---|
A1 | OA = 93.09% Kappa Coefficient = 0.78 | A morphological dilation filter was used to correct the extracted buildings, compensating for the effects of shadow on building boundaries. Before applying morphological operation at area A1, the OA and the kappa coefficient were 93.22% and 0.798, respectively. After applying the OA, and kappa coefficient rose to 93.09% and 0.788. It implies the operation is effective for area A1. | |
A3 | OA = 92.16% Kappa Coefficient = 0.840 | The same operation was performed on area A3. The OA and kappa coefficient value was 92.16%, 0.840, and after this operation on A3, the new OA and kappa coefficient became 91.87%, 0.834. at the same time, the performance was poor for A3. Hence morphological operation cannot provide total refinement of all forms of building objects. |
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2021 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Ojogbane, S.S.; Mansor, S.; Kalantar, B.; Khuzaimah, Z.B.; Shafri, H.Z.M.; Ueda, N. Automated Building Detection from Airborne LiDAR and Very High-Resolution Aerial Imagery with Deep Neural Network. Remote Sens. 2021, 13, 4803. https://doi.org/10.3390/rs13234803
Ojogbane SS, Mansor S, Kalantar B, Khuzaimah ZB, Shafri HZM, Ueda N. Automated Building Detection from Airborne LiDAR and Very High-Resolution Aerial Imagery with Deep Neural Network. Remote Sensing. 2021; 13(23):4803. https://doi.org/10.3390/rs13234803
Chicago/Turabian StyleOjogbane, Sani Success, Shattri Mansor, Bahareh Kalantar, Zailani Bin Khuzaimah, Helmi Zulhaidi Mohd Shafri, and Naonori Ueda. 2021. "Automated Building Detection from Airborne LiDAR and Very High-Resolution Aerial Imagery with Deep Neural Network" Remote Sensing 13, no. 23: 4803. https://doi.org/10.3390/rs13234803
APA StyleOjogbane, S. S., Mansor, S., Kalantar, B., Khuzaimah, Z. B., Shafri, H. Z. M., & Ueda, N. (2021). Automated Building Detection from Airborne LiDAR and Very High-Resolution Aerial Imagery with Deep Neural Network. Remote Sensing, 13(23), 4803. https://doi.org/10.3390/rs13234803