Learning Representative Features by Deep Attention Network for 3D Point Cloud Registration
Abstract
:1. Introduction
2. Related Work
2.1. Feature Extraction
2.2. Attention Mechanism
3. Methods
3.1. Feature Extraction
3.2. Attentional Layer
3.3. Loss Function
3.3.1. Feature Loss
3.3.2. Overlap Loss
3.3.3. Matchability Loss
4. Experiments
4.1. Dataset
4.2. Comparison on Different Estimators
- prob: The interest points are sampled by probabilistic sampling, where the probability is generated by the product of overlap score and saliency score. Subsequently, RANSAC is utilized to perform feature marching and obtain the transformation matrix.
- topk: Firstly, a similarity matrix is derived based on the Euclidean distance between the features of points from two point clouds. The top k pairs of points with the highest similarity are selected. Next, similar to prob, we conduct feature matching on those selected pairs.
- topk-corres: It is similar to the topk approach. However, instead of conducting feature matching on those points, the transformation matrix is directly computed by applying RANSAC on the top k correspondences found in the similarity matrix.
- topk-kabsch: It is similar to topk. The difference is that the correspondences and their similarity scores are used as the input to the kabsch algorithm, which provides a closed-form solution for computing the transformation matrix.
4.3. Performance on Partial-to-Partial Point Cloud
4.4. Performance on Point Clouds with Gaussian Noise
4.5. Statistical Significance Test
4.6. Performance on Point Clouds from Unseen Categories
4.7. Additional Experiments on Additive Noise
4.8. Runtime Analysis
5. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Li, J.; Saydam, S.; Xu, Y.; Liu, B.; Li, B.; Lin, X.; Zhang, W. Class-Aware Tiny Object Recognition over Large-Scale 3D Point Clouds. Neurocomputing 2023, 529, 166–181. [Google Scholar] [CrossRef]
- Lin, C.H.; Kong, C.; Lucey, S. Learning efficient point cloud generation for dense 3D object reconstruction. In Proceedings of the 32nd AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018; Volume 32. [Google Scholar]
- Liu, M. Robotic online path planning on point cloud. IEEE Trans. Cybern. 2015, 46, 1217–1228. [Google Scholar] [CrossRef] [PubMed]
- Woo, H.; Kang, E.; Wang, S.; Lee, K.H. A new segmentation method for point cloud data. Int. J. Mach. Tools Manuf. 2002, 42, 167–178. [Google Scholar] [CrossRef]
- Huang, X.; Zhang, J.; Wu, Q.; Fan, L.; Yuan, C. A coarse-to-fine algorithm for registration in 3D street-view cross-source point clouds. In Proceedings of the 2016 International Conference on Digital Image Computing: Techniques and Applications (DICTA), Gold Coast, QLD, Australia, 30 November–2 December 2016; pp. 1–6. [Google Scholar]
- Bai, X.; Luo, Z.; Zhou, L.; Fu, H.; Quan, L.; Tai, C.L. D3feat: Joint learning of dense detection and description of 3D local features. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 6359–6367. [Google Scholar]
- Choy, C.; Dong, W.; Koltun, V. Deep global registration. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 2514–2523. [Google Scholar]
- Gojcic, Z.; Zhou, C.; Wegner, J.D.; Guibas, L.J.; Birdal, T. Learning multiview 3D point cloud registration. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 1759–1769. [Google Scholar]
- Choy, C.; Park, J.; Koltun, V. Fully convolutional geometric features. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 8958–8966. [Google Scholar]
- Huang, S.; Gojcic, Z.; Usvyatsov, M.; Wieser, A.; Schindler, K. Predator: Registration of 3D point clouds with low overlap. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 4267–4276. [Google Scholar]
- Rusu, R.B.; Blodow, N.; Beetz, M. Fast point feature histograms (FPFH) for 3D registration. In Proceedings of the 2009 IEEE International Conference on Robotics and Automation, Kobe, Japan, 12–17 May 2009; pp. 3212–3217. [Google Scholar]
- Tombari, F.; Salti, S.; Di Stefano, L. Unique signatures of histograms for local surface description. In Computer Vision—ECCV 2010, Proceedings of the 11th European Conference on Computer Vision, Heraklion, Crete, Greece, 5–11 September 2010; Proceedings, Part III 11; Springe: Berlin/Heidelberg, Germany, 2010; pp. 356–369. [Google Scholar]
- Qi, C.R.; Su, H.; Mo, K.; Guibas, L.J. Pointnet: Deep learning on point sets for 3D classification and segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 652–660. [Google Scholar]
- Qi, C.R.; Yi, L.; Su, H.; Guibas, L.J. Pointnet++: Deep hierarchical feature learning on point sets in a metric space. Adv. Neural Inf. Process. Syst. 2017, 30, 1–10. [Google Scholar]
- Zeng, A.; Song, S.; Nießner, M.; Fisher, M.; Xiao, J.; Funkhouser, T. 3dmatch: Learning local geometric descriptors from rgb-d reconstructions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 1802–1811. [Google Scholar]
- Sarlin, P.E.; DeTone, D.; Malisiewicz, T.; Rabinovich, A. Superglue: Learning feature matching with graph neural networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 4938–4947. [Google Scholar]
- Fischer, K.; Simon, M.; Olsner, F.; Milz, S.; Gross, H.M.; Mader, P. Stickypillars: Robust and efficient feature matching on point clouds using graph neural networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 313–323. [Google Scholar]
- Shi, C.; Chen, X.; Huang, K.; Xiao, J.; Lu, H.; Stachniss, C. Keypoint matching for point cloud registration using multiplex dynamic graph attention networks. IEEE Robot. Autom. Lett. 2021, 6, 8221–8228. [Google Scholar] [CrossRef]
- Qin, Z.; Yu, H.; Wang, C.; Guo, Y.; Peng, Y.; Xu, K. Geometric transformer for fast and robust point cloud registration. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 11143–11152. [Google Scholar]
- Yew, Z.J.; Lee, G.H. Regtr: End-to-end point cloud correspondences with transformers. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 6677–6686. [Google Scholar]
- Thomas, H.; Qi, C.R.; Deschaud, J.E.; Marcotegui, B.; Goulette, F.; Guibas, L.J. Kpconv: Flexible and deformable convolution for point clouds. In Proceedings of the IEEE/CVF international Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 6411–6420. [Google Scholar]
- Wang, Y.; Sun, Y.; Liu, Z.; Sarma, S.E.; Bronstein, M.M.; Solomon, J.M. Dynamic graph cnn for learning on point clouds. ACM Trans. Graph. 2019, 38, 1–12. [Google Scholar] [CrossRef]
- Choy, C.; Gwak, J.; Savarese, S. 4D spatio-temporal convnets: Minkowski convolutional neural networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–17 June 2019; pp. 3075–3084. [Google Scholar]
- Wang, Y.; Solomon, J.M. Deep closest point: Learning representations for point cloud registration. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 3523–3532. [Google Scholar]
- Fu, K.; Liu, S.; Luo, X.; Wang, M. Robust point cloud registration framework based on deep graph matching. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 8893–8902. [Google Scholar]
- Li, J.; Lee, G.H. Usip: Unsupervised stable interest point detection from 3D point clouds. In Proceedings of the IEEE/CVF international Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 361–370. [Google Scholar]
- Sun, Y.; Cheng, C.; Zhang, Y.; Zhang, C.; Zheng, L.; Wang, Z.; Wei, Y. Circle loss: A unified perspective of pair similarity optimization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 6398–6407. [Google Scholar]
- Besl, P.J.; McKay, N.D. Method for registration of 3-D shapes. Proc. Sens. Fusion IV Control Paradig. Data Struct. 1992, 1611, 586–606. [Google Scholar] [CrossRef]
- Zhou, Q.Y.; Park, J.; Koltun, V. Fast global registration. In Computer Vision—ECCV 2016, Proceedings of the 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Proceedings, Part II 14; Springer: Berlin/Heidelberg, Germany, 2016; pp. 766–782. [Google Scholar]
- Yuan, W.; Eckart, B.; Kim, K.; Jampani, V.; Fox, D.; Kautz, J. Deepgmr: Learning latent gaussian mixture models for registration. In Computer Vision—ECCV 2020, Proceedings of the 16th European Conference, Glasgow, UK, 23–28 August 2020; Proceedings, Part V 16; Springer: Berlin/Heidelberg, Germany, 2020; pp. 733–750. [Google Scholar]
- Zhou, Q.Y.; Park, J.; Koltun, V. Open3D: A modern library for 3D data processing. arXiv 2018, arXiv:1801.09847. [Google Scholar]
- Song, Y.; Shen, W.; Peng, K. A novel partial point cloud registration method based on graph attention network. Vis. Comput. 2023, 39, 1109–1120. [Google Scholar] [CrossRef]
- Wu, Z.; Song, S.; Khosla, A.; Yu, F.; Zhang, L.; Tang, X.; Xiao, J. 3D shapenets: A deep representation for volumetric shapes. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1912–1920. [Google Scholar]
#Keypoints | 100 | 200 | 300 | 400 | 500 | 600 |
---|---|---|---|---|---|---|
prob | 2.7544/0.05226 | 2.3391/0.04305 | 2.0229/0.04081 | 2.3391/0.01939 | 4.2302/0.03271 | 10.5938/0.07608 |
topk | 1.6285/0.03398 | 1.5996/0.03365 | 1.6234/0.03499 | 1.6099/0.03423 | 1.6323/0.03537 | 1.6394/0.03498 |
topk-corres | 1.3452/0.03736 | 1.2001/0.03337 | 1.1640/0.03298 | 1.0684/0.03286 | 1.0878/0.03409 | 1.0931/0.03395 |
topk-kabsch | 1.6367/0.03298 | 1.6119/0.03329 | 1.6111/0.03340 | 1.6048/0.03340 | 1.6085/0.03356 | 1.6179/0.03357 |
Method | RMSE (R) | RMSE (t) | MAE (R) | MAE (t) |
---|---|---|---|---|
ICP | 29.0295 | 0.9938 | 22.6922 | 0.8716 |
FGR | 58.3071 | 0.3122 | 36.8645 | 0.2096 |
DeepGMR | 87.6945 | 0.4001 | 66.0553 | 0.2933 |
Predator | 9.1573 | 0.0687 | 3.4375 | 0.0290 |
Our | 2.8160 | 0.0318 | 1.1054 | 0.0102 |
Method | RMSE (R) | RMSE (t) | MAE (R) | MAE (t) |
---|---|---|---|---|
ICP | 29.7609 | 0.9931 | 23.0322 | 0.8704 |
FGR | 40.3478 | 0.1910 | 19.3586 | 0.1025 |
DeepGMR | 45.8762 | 0.1877 | 23.5213 | 0.09801 |
Predator | 4.6180 | 0.0371 | 1.8349 | 0.0144 |
Our | 1.8833 | 0.0138 | 0.6241 | 0.0050 |
Experiment | SE (R) | SE (t) | AE (R) | AE (t) |
---|---|---|---|---|
Partial | ||||
Noisy |
Method | RMSE (R) | RMSE (t) | MAE (R) | MAE (t) |
---|---|---|---|---|
ICP | 29.0295 | 0.9938 | 22.6922 | 0.8716 |
FGR | 58.3623 | 0.3034 | 37.1349 | 0.2086 |
GAN | 5.977 | 0.046 | 4.534 | 0.021 |
Our | 3.1371 | 0.0314 | 1.2608 | 0.0117 |
Method | Our (Prob) | Our (Topk) | Our (Topk-Kabsch) | Our (Topk-Corres) | Predator |
---|---|---|---|---|---|
Time (ms) | 49 | 50 | 41 | 55 | 43 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Xia, X.; Fan, Z.; Xiao, G.; Chen, F.; Liu, Y.; Hu, Y. Learning Representative Features by Deep Attention Network for 3D Point Cloud Registration. Sensors 2023, 23, 4123. https://doi.org/10.3390/s23084123
Xia X, Fan Z, Xiao G, Chen F, Liu Y, Hu Y. Learning Representative Features by Deep Attention Network for 3D Point Cloud Registration. Sensors. 2023; 23(8):4123. https://doi.org/10.3390/s23084123
Chicago/Turabian StyleXia, Xiaokai, Zhiqiang Fan, Gang Xiao, Fangyue Chen, Yu Liu, and Yiheng Hu. 2023. "Learning Representative Features by Deep Attention Network for 3D Point Cloud Registration" Sensors 23, no. 8: 4123. https://doi.org/10.3390/s23084123
APA StyleXia, X., Fan, Z., Xiao, G., Chen, F., Liu, Y., & Hu, Y. (2023). Learning Representative Features by Deep Attention Network for 3D Point Cloud Registration. Sensors, 23(8), 4123. https://doi.org/10.3390/s23084123