2D&3DHNet for 3D Object Classification in LiDAR Point Cloud
Abstract
:1. Introduction
- Each cell in the 3D Hough space voted by the relevant numerous coplanar points represents the normal vector of their plane as a common feature. Thus, the Hough descriptors have a wide effective receptive field. The Hough features are voted by a set of unordered coplanar points, which satisfies the permutation invariant requirement of deep neural networks. Furthermore, the extracted planar features are stable against the point loss and non-uniform density of LiDAR point clouds.
- The multi-scale critical point sampling method is developed to extract critical points for retaining the local spatial structure of the object. This way, the redundant points are removed to reduce the local features computation time cost.
- To preserve the local features, the grid-based dynamic nearest neighbors algorithm is developed to select a certain number of points nearby the critical points for discrimination of local features generation.
- The fusion of 3D global and 2D local Hough features enables discriminative structure retrieval and improves classification accuracy.
2. Related Works
3. Object Recognition Method
3.1. Global Hough Features Extraction
3.2. Local Hough Feature Generation
3.3. 2D&3DHNet
4. Experiments
4.1. Experimental Environment
4.2. Global Hough Feature Analysis
4.3. Local Hough Feature Analysis
4.4. Classification Performance Using Global Hough Features
4.5. Classification Performance of 2D&3DHNet
4.6. Algorithms Comparison
5. Conclusions
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- Yan, F.; Wang, J.; He, G.; Chang, H.; Zhuang, Y. Sparse semantic map building and relocalization for UGV using 3D point clouds in outdoor environments. Neurocomputing 2020, 400, 333–342. [Google Scholar] [CrossRef]
- Eitel, J.U.; Höfle, B.; Vierling, L.A.; Abellán, A.; Asner, G.P.; Deems, J.S.; Glennie, C.L.; Joerg, P.C.; LeWinter, A.L.; Magney, T.S.; et al. Beyond 3-D: The new spectrum of lidar applications for earth and ecological sciences. Remote Sens. Environ. 2016, 186, 372–392. [Google Scholar] [CrossRef] [Green Version]
- Eagleston, H.; Marion, J.L. Application of airborne LiDAR and GIS in modeling trail erosion along the Appalachian Trail in New Hampshire, USA. Landsc. Urban Plan. 2020, 198, 103765. [Google Scholar] [CrossRef]
- Li, J.; Xu, Y.; Macrander, H.; Atkinson, L.; Thomas, T.; Lopez, M.A. GPU-based lightweight parallel processing toolset for LiDAR data for terrain analysis. Environ. Model. Softw. 2019, 117, 55–68. [Google Scholar] [CrossRef]
- Weinmann, M.; Jutzi, B.; Hinz, S.; Mallet, C. Semantic point cloud interpretation based on optimal neighborhoods, relevant features and efficient classifiers. ISPRS J. Photogramm. Remote Sens. 2015, 105, 286–304. [Google Scholar] [CrossRef]
- Charles, R.Q.; Su, H.; Kaichun, M.; Guibas, L.J. PointNet: Deep learning on point sets for 3D classification and segmentation. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 77–85. [Google Scholar]
- Tian, Y.; Song, W.; Chen, L.; Sung, Y.; Kwak, J.; Sun, S. Fast Planar Detection System Using a GPU-Based 3D Hough Transform for LiDAR Point Clouds. Appl. Sci. 2020, 10, 1744. [Google Scholar] [CrossRef] [Green Version]
- Acheampong, R.A.; Cugurullo, F.; Gueriau, M.; Dusparic, I. Can autonomous vehicles enable sustainable mobility in future cities? Insights and policy challenges from user preferences over different urban transport options. Cities 2021, 112, 103134. [Google Scholar] [CrossRef]
- Dowling, R.; McGuirk, P. Autonomous vehicle experiments and the city. Urban Geogr. 2022, 43, 409–426. [Google Scholar] [CrossRef]
- Cugurullo, F. Urban artificial intelligence: From automation to autonomy in the smart city. Front. Sustain. Cities 2020, 2, 38. [Google Scholar] [CrossRef]
- Petrović, Đ.; Mijailović, R.; Pešić, D. Traffic accidents with autonomous vehicles: Type of collisions, manoeuvres and errors of conventional vehicles’ drivers. Transp. Res. Procedia 2020, 45, 161–168. [Google Scholar] [CrossRef]
- Wang, C.; Shu, Q.; Wang, X.; Guo, B.; Liu, P.; Li, Q. A random forest classifier based on pixel comparison features for urban LiDAR data. ISPRS J. Photogramm. Remote Sens. 2018 148, 75–86. [CrossRef]
- Lehtomäki, M.; Jaakkola, A.; Hyyppä, J.; Lampinen, J.; Kaartinen, H.; Kukko, A.; Puttonen, E.; Hyyppä, H. Object Classification and Recognition from Mobile Laser Scanning Point Clouds in a Road Environment. IEEE Trans. Geosci. Remote Sens. 2015, 54, 1226–1239. [Google Scholar] [CrossRef] [Green Version]
- Miao, X.; Heaton, J.S. A comparison of random forest and Adaboost tree in ecosystem classification in east Mojave Desert. In Proceedings of the 2010 18th International Conference on Geoinformatics, Beijing, China, 18–20 June 2010; pp. 1–6. [Google Scholar]
- Li, W.; Wang, F.; Xia, G. A geometry-attentional network for ALS point cloud classification. ISPRS J. Photogramm. Remote Sens. 2020, 164, 26–40. [Google Scholar] [CrossRef]
- Su, H.; Maji, S.; Kalogerakis, E.; Learned-Miller, E. Multi-view Convolutional Neural Networks for 3D Shape Recognition. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 945–953. [Google Scholar]
- Feng, Y.; Zhang, Z.; Zhao, X.; Ji, R.; Gao, Y. GVCNN: Group-View Convolutional Neural Networks for 3D Shape Recognition. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 264–272. [Google Scholar]
- Yu, T.; Meng, J.; Yuan, J. Multi-view Harmonized Bilinear Network for 3D Object Recognition. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 186–194. [Google Scholar]
- Biasutti, P.; Lepetit, V.; Aujol, J.; Brédif, M.; Bugeau, A. LU-Net: An Efficient Network for 3D LiDAR Point Cloud Semantic Segmentation Based on End-to-End-Learned 3D Features and U-Net. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW), Seoul, Korea, 27–28 October 2019; pp. 942–950. [Google Scholar]
- Alonso, I.; Riazuelo, L.; Montesano, L.; Murillo, A.C. 3D-MiniNet: Learning a 2D Representation from Point Clouds for Fast and Efficient 3D LIDAR Semantic Segmentation. IEEE Robot. Autom. Lett. 2020, 5, 5432–5439. [Google Scholar] [CrossRef]
- Maturana, D.; Scherer, S. VoxNet: A 3D Convolutional Neural Network for real-time object recognition. In Proceedings of the 2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Hamburg, Germany, 28 September–2 October 2015; pp. 922–928. [Google Scholar]
- Wang, C.; Cheng, M.; Sohel, F.; Bennamoun, M.; Li, J. NormalNet: A voxel-based CNN for 3D object classification and retrieval. Neurocomputing 2019, 323, 139–147. [Google Scholar] [CrossRef]
- Le, T.; Duan, Y. PointGrid: A Deep Network for 3D Shape Understanding. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; pp. 9204–9214. [Google Scholar]
- Wang, L.; Huang, Y.; Shan, J.; He, L. MSNet: Multi-Scale Convolutional Network for Point Cloud Classification. Remote Sens. 2018, 10, 612. [Google Scholar] [CrossRef] [Green Version]
- Hua, B.; Tran, M.; Yeung, S. Pointwise convolutional neural networks. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 984–993. [Google Scholar]
- Qi, C.R.; Yi, L.; Su, H.; Guibas, L.J. Pointnet++: Deep hierarchical feature learning on point sets in a metric space. In Proceedings of the NIPS’17: Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4-9 December 2017; Curran Associates Inc.: Red Hook, NY, USA, 2017; pp. 5105–5114. [Google Scholar]
- Wang, Z.; Zhang, L.; Li, R.; Zheng, Y.; Zhu, Z. A Deep Neural Network With Spatial Pooling (DNNSP) for 3-D Point Cloud Classification. IEEE Trans. Geosci. Remote Sens. 2018, 56, 4594–4604. [Google Scholar] [CrossRef]
- Duda, R.O.; Hart, P.E. Use of the Hough Transformation to Detect Lines and Curves in Pictures; Technical Report; Sri International, Artificial Intelligence Center: Menlo Park, CA, USA, 1971. [Google Scholar]
- Hu, X.; Tao, C.V.; Hu, Y. Automatic road extraction from dense urban area by integrated processing of high resolution imagery and lidar data. International Archives of Photogrammetry. Remote Sens. Spat. Inf. Sci. 2004, 35 Pt B3, 288–292. [Google Scholar]
- Tarsha-Kurdi, F.; Landes, T.; Grussenmeyer, P. Hough-transform and extended ransac algorithms for automatic detection of 3d building roof planes from lidar data. In Proceedings of the ISPRS Workshop on Laser Scanning 2007 and SilviLaser 2007, Espoo, Finland, 12–14 September 2007; Volume 36, pp. 407–412. [Google Scholar]
- Sampath, A.; Shan, J. Segmentation and reconstruction of polyhedral building roofs from aerial lidar point clouds. IEEE Trans. Geosci. Remote Sens. 2009, 48, 1554–1567. [Google Scholar] [CrossRef]
- Xue, F.; Lu, W.; Webster, C.J.; Chen, K. A derivative-free optimization-based approach for detecting architectural symmetries from 3D point clouds. ISPRS J. Photogramm. Remote Sens. 2019, 148, 32–40. [Google Scholar] [CrossRef]
- Van Leeuwen, M.; Coops, N.C.; Wulder, M.A. Canopy surface reconstruction from a LiDAR point cloud using Hough transform. Remote Sens. Lett. 2010, 1, 125–132. [Google Scholar] [CrossRef]
- Widyaningrum, E.; Gorte, B.; Lindenbergh, R. Automatic building outline extraction from ALS point clouds by ordered points aided hough transform. Remote Sens. 2019, 11, 1727. [Google Scholar] [CrossRef] [Green Version]
- Zhao, K.; Han, Q.; Zhang, C.; Xu, J.; Cheng, M. Deep Hough Transform for Semantic Line Detection. Comput. Vis.-ECCV 2020 2020, 12354, 249–265. [Google Scholar] [CrossRef]
- Qi, C.R.; Litany, O.; He, K.; Guibas, L. Deep hough voting for 3d object detection in point clouds. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Korea, 27 October–2 November 2019; pp. 9276–9285. [Google Scholar]
- Song, W.; Zhang, L.; Tian, Y.; Fong, S.; Liu, J.; Gozho, A. CNN-based 3D Object Classification Using Hough Space of LiDAR Point Clouds. Hum.-Cent. Comput. Inf. Sci. 2020, 10, 19. [Google Scholar] [CrossRef]
- Deuge, M.; Quadros, A.; Hungl, C.; Douillard, B. Unsupervised Feature Learning for Classification of Outdoor 3D Scans. In Proceedings of the Australasian Conference on Robotics and Automation (ACRA), University of New South Wales, Sydney, Australia, 2–4 December 2013; University of New South Wales: Kensington, Australia, 2013. [Google Scholar]
- Wu, Z.; Song, S.; Khosla, A.; Yu, F.; Zhang, L.; Tang, X.; Xiao, J. 3D ShapeNets: A Deep Representation for Volumetric Shapes. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 1912–1920. [Google Scholar]
- Wang, Y.; Sun, Y.; Liu, Z.; Sarma, S.E.; Bronstein, M.M.; Solomon, J.M. Dynamic Graph CNN for Learning on Point Clouds. ACM Trans. Graph. 2019, 38, 1–12. [Google Scholar] [CrossRef] [Green Version]
- Li, Y.; Bu, R.; Sun, M.; Wu, W.; Di, X.; Chen, B. PointCNN: Convolution On X-Transformed Points. In NIPS’18: Proceedings of the 32nd International Conference on Neural Information Processing Systems, Montréal, Canada, 3–8 December 2018; Curran Associates Inc.: Red Hook, NY, USA, 2018; pp. 820–830. [Google Scholar]
Category | Sydney | Our | Training Dataset | Evaluation Dataset | Total |
---|---|---|---|---|---|
pole | 0 | 236 | 160 | 76 | 236 |
pedestrian | 69 | 83 | 92 | 60 | 152 |
tree | 0 | 415 | 286 | 129 | 415 |
bush | 0 | 223 | 141 | 82 | 223 |
building | 0 | 385 | 252 | 133 | 385 |
vehicle | 97 | 93 | 115 | 75 | 190 |
Total | 166 | 1435 | 1046 | 555 | 1601 |
Type | 20 × 20 × 20 | 25 × 25 × 25 | 30 × 30 × 30 | 32 × 32 × 32 | ||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|
P | R | F1 | P | R | F1 | P | R | F1 | P | R | F1 | |
pole | 94.9% | 97.4% | 96.1% | 95% | 100% | 97.4% | 100% | 98.7% | 99.3% | 90.5% | 100% | 95% |
pedestrian | 100% | 78.3% | 86.8% | 96.4% | 90% | 93.1% | 94.2% | 81.7% | 87.5% | 100% | 76.7% | 86.8% |
tree | 87.5% | 98.4% | 92.6% | 94.7% | 96.9% | 95.8% | 93.9% | 96.9% | 95.4% | 93.2% | 95.4% | 94.3% |
bush | 95.2% | 97.6% | 96.4% | 96.5% | 100% | 98.2% | 91.9% | 96.3% | 94.% | 90.1% | 100% | 94.8% |
building | 94.9% | 98.5% | 96.7% | 94.3% | 99.2% | 96.7% | 97% | 98.5% | 97.8% | 95.6% | 98.5% | 97% |
vehicle | 96.8% | 81.3% | 88.4% | 100% | 82.7% | 90.6% | 89.2% | 88% | 88.6% | 95.4% | 82.7% | 88.6% |
Avg | 94.8% | 91.9% | 92.8% | 96.4% | 94.8% | 95.3% | 94.3% | 93.3% | 93.7% | 94.1% | 92.2% | 92.7% |
Accuracy | 93.5% | 95.7% | 94.6 | 93.7 |
Computational Load | 20 × 20 × 20 | 25 × 25 × 25 | 30 × 30 × 30 | 32 × 32 × 32 |
---|---|---|---|---|
3D Hough transformation | 1.7 s | 2.7 s | 3.8 s | 4.3 s |
Average training time (s) | 170 s | 173 s | 176 s | 178 s |
Average testing time (s) | 5 s | 8 s | 11 s | 13 s |
I × J × K | 2D Res | Pole | Pedestrian | Tree | Bush | Building | Vehicle | Avg |
---|---|---|---|---|---|---|---|---|
20 × 20 × 20 | 0.10 | 97.4% | 95.0% | 96.9% | 100% | 100% | 90.7% | 96.67% |
0.15 | 96.1% | 96.7% | 97.7% | 98.8% | 97.7% | 85.3% | 95.38% | |
0.20 | 98.7% | 96.7% | 98.4% | 100% | 99.2% | 86.7% | 96.62% | |
0.25 | 98.7% | 96.7% | 100% | 100% | 97.0% | 88.0% | 96.73% | |
0.30 | 100% | 95.0% | 98.4% | 100% | 100% | 84.0% | 96.23% | |
25 × 25 × 25 | 0.10 | 100% | 95.0% | 95.3% | 100% | 91.0% | 89.3% | 95.10% |
0.15 | 100% | 100% | 97.7% | 100% | 93.2% | 86.7% | 96.27% | |
0.20 | 100% | 95.0% | 95.3% | 100% | 97.7% | 86.7% | 95.78% | |
0.25 | 100% | 96.7% | 99.2% | 100% | 97.7% | 92.0% | 97.60% | |
0.30 | 100% | 93.3% | 95.3% | 98.8% | 95.5% | 89.3% | 95.37% | |
30 × 30 × 30 | 0.10 | 100% | 93.3% | 94.6% | 97.6% | 91.0% | 89.3% | 94.30% |
0.15 | 100% | 98.3% | 96.9% | 100% | 97.7% | 86.7% | 96.60% | |
0.20 | 100% | 90.0% | 89.1% | 100% | 97.7% | 89.3% | 94.35% | |
0.25 | 97.3% | 98.3% | 94.6% | 100% | 97.0% | 88.0% | 95.87% | |
0.30 | 97.4% | 96.7% | 96.1% | 98.8% | 99.2% | 85.3% | 95.58% |
Method | Input | Average Training Time (s) | Average Testing Time (s) | Accuracy | Params |
---|---|---|---|---|---|
VoxNet [21] | Voxel | 100 s | 3.5 s | 91.7% | 11.18M |
3DShapeNet [39] | Voxel | 425 s | 2 s | 93.6% | 15.9M |
MVCNN [16] | Image | 2150 s | 45 s | 95.7% | 0.92M |
DGCNN [40] | Graph | 8400 s | 25 s | 95.3% | 1.35M |
PointCNN [41] | point | 10,750 s | 30 s | 96.2% | 0.3M |
2D&3DHNet (ours, 25 × 25 × 25 voxel) | point | 147 s | 9 s | 97.6% | 7.97M |
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2022 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Song, W.; Li, D.; Sun, S.; Zhang, L.; Xin, Y.; Sung, Y.; Choi, R. 2D&3DHNet for 3D Object Classification in LiDAR Point Cloud. Remote Sens. 2022, 14, 3146. https://doi.org/10.3390/rs14133146
Song W, Li D, Sun S, Zhang L, Xin Y, Sung Y, Choi R. 2D&3DHNet for 3D Object Classification in LiDAR Point Cloud. Remote Sensing. 2022; 14(13):3146. https://doi.org/10.3390/rs14133146
Chicago/Turabian StyleSong, Wei, Dechao Li, Su Sun, Lingfeng Zhang, Yu Xin, Yunsick Sung, and Ryong Choi. 2022. "2D&3DHNet for 3D Object Classification in LiDAR Point Cloud" Remote Sensing 14, no. 13: 3146. https://doi.org/10.3390/rs14133146
APA StyleSong, W., Li, D., Sun, S., Zhang, L., Xin, Y., Sung, Y., & Choi, R. (2022). 2D&3DHNet for 3D Object Classification in LiDAR Point Cloud. Remote Sensing, 14(13), 3146. https://doi.org/10.3390/rs14133146