Research on Intelligent Robot Point Cloud Grasping in Internet of Things
Abstract
:1. Introduction
- (1)
- This study designs a grasping quality evaluation network based on the PointNet network, which is used to evaluate the quality of the generated candidate grasping positions, and a plug-and-play lightweight attention mechanism for point clouds that can be embedded in the feature extraction phase of the PointNet network to improve the network performance without significantly increasing the computational cost.
- (2)
- Generating a grasp dataset containing object grasp location and quality labels based on the YCB dataset [18] for training our proposed grasp quality evaluation network.
- (3)
- The actual grasping experiments are carried out with the Baxter robot and compared with the existing methods; the results show that our method has higher accuracy and higher grasping success rate.
2. Related Work
2.1. Processing of Point Cloud Data
2.2. Robot Grasping Based on Object Point Cloud
2.3. Attention Mechanism in Computer Vision
3. Principal Analysis
3.1. PointNet Network Structure Analysis
3.2. Analysis of Attention Mechanism in Computer Vision
4. Grasping Quality Classification Network Incorporating Attention Mechanism
4.1. Structure Design of Point Cloud Classification Network
4.2. Point Cloud Attention Mechanism Network Design
4.3. Design of PointNet Grasping Quality Classification Network Incorporating Attention Mechanism
5. Model Training and Actual Grasping Experiments
5.1. Generating Grasping Dataset
5.1.1. Sampling of Candidate Grasp Positions
5.1.2. Generating Training Labels
5.2. Training the Generated Network
5.3. Effect of Verification Module on Network Accuracy
5.4. Actual Grasping Experiments
6. Conclusions and Future Work
Author Contributions
Funding
Conflicts of Interest
References
- Nord, J.H.; Koohang, A.; Paliszkiewicz, J. The Internet of Things: Review and Theoretical Framework. Expert Syst. Appl. 2019, 133, 97–108. [Google Scholar] [CrossRef]
- Shin, M.; Paik, W.; Kim, B.; Hwang, S. An IoT Platform with Monitoring Robot Applying CNN-Based Context-Aware Learning. Sensors 2019, 19, 2525. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Grieco, L.A.; Rizzo, A.; Colucci, S.; Sicari, S.; Piro, G.; di Paola, D.; Boggia, G. IoT-Aided Robotics Applications: Technological Implications, Target Domains and Open Issues. Comput. Commun. 2014, 54, 32–47. [Google Scholar] [CrossRef]
- Bicchi, A.; Kumar, V. Robotic Grasping and Contact: A Review. In Proceedings of the IEEE International Conference on Robotics and Automation, San Francisco, CA, USA, 24–28 April 2000; Volume 1. [Google Scholar] [CrossRef]
- Sahbani, A.; El-Khoury, S.; Bidaud, P. An Overview of 3D Object Grasp Synthesis Algorithms. Rob. Auton. Syst. 2012, 60, 326–336. [Google Scholar] [CrossRef] [Green Version]
- Nguyen, V.D. Constructing Force-Closure Grasps. Int. J. Rob. Res. 1988, 7, 1368–1373. [Google Scholar] [CrossRef]
- Du, G.; Wang, K.; Lian, S.; Zhao, K. Vision-Based Robotic Grasping from Object Localization, Object Pose Estimation to Grasp Estimation for Parallel Grippers: A Review. Artif. Intell. Rev. 2021, 54, 1677–1734. [Google Scholar] [CrossRef]
- Lenz, I.; Lee, H.; Saxena, A. Deep Learning for Detecting Robotic Grasps. Int. J. Robot. Res. 2015, 34, 705–724. [Google Scholar] [CrossRef] [Green Version]
- Redmon, J.; Angelova, A. Real-Time Grasp Detection Using Convolutional Neural Networks. In Proceedings of the IEEE International Conference on Robotics and Automation, Seattle, WA, USA, 26–30 May 2015. [Google Scholar]
- Chu, F.J.; Xu, R.; Vela, P.A. Real-World Multiobject, Multigrasp Detection. IEEE Robot Autom. Lett. 2018, 3, 3355–3362. [Google Scholar] [CrossRef] [Green Version]
- Kumra, S.; Kanan, C. Robotic Grasp Detection Using Deep Convolutional Neural Networks. In Proceedings of the IEEE International Conference on Intelligent Robots and Systems, Vancouver, BC, Canada, 24–28 September 2017. [Google Scholar]
- Mahler, J.; Liang, J.; Niyaz, S.; Laskey, M.; Doan, R.; Liu, X.; Ojea, J.A.; Goldberg, K. Dex-Net 2.0: Deep Learning to Plan Robust Grasps with Synthetic PoInt. Clouds and Analytic Grasp Metrics. In Proceedings of the Robotics: Science and Systems, Cambridge, MA, USA, 12–16 July 2017; Volume 13. [Google Scholar]
- ten Pas, A.; Gualtieri, M.; Saenko, K.; Platt, R. Grasp Pose Detection in PoInt. Clouds. Int. J. Robot. Res. 2017, 36, 1455–1473. [Google Scholar] [CrossRef]
- Liang, H.; Ma, X.; Li, S.; Gorner, M.; Tang, S.; Fang, B.; Sun, F.; Zhang, J. PointNetGPD: Detecting Grasp Configurations from PoInt. Sets. In Proceedings of the IEEE International Conference on Robotics and Automation, Montreal, QC, Canada, 20–24 May 2019. [Google Scholar]
- Qi, C.R.; Su, H.; Mo, K.; Guibas, L.J. PointNet: Deep Learning on PoInt. Sets for 3D Classification and Segmentation. In Proceedings of the 30th IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017, Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
- Mousavian, A.; Eppner, C.; Fox, D. 6-DOF GraspNet: Variational Grasp Generation for Object Manipulation. In Proceedings of the IEEE International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019. [Google Scholar]
- Qi, C.R.; Yi, L.; Su, H.; Guibas, L.J. PointNet++: Deep Hierarchical Feature Learning on PoInt. Sets in a Metric Space. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
- Calli, B.; Singh, A.; Bruce, J.; Walsman, A.; Konolige, K.; Srinivasa, S.; Abbeel, P.; Dollar, A.M. Yale-CMU-Berkeley Dataset for Robotic Manipulation Research. Int. J. Robot. Res. 2017, 36, 261–268. [Google Scholar] [CrossRef]
- Guo, Y.; Wang, H.; Hu, Q.; Liu, H.; Liu, L.; Bennamoun, M. Deep Learning for 3D PoInt. Clouds: A Survey. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 43, 4338–4364. [Google Scholar] [CrossRef] [PubMed]
- Chen, X.; Ma, H.; Wan, J.; Li, B.; Xia, T. Multi-View 3D Object Detection Network for Autonomous Driving. In Proceedings of the 30th IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017, Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
- Ku, J.; Mozifian, M.; Lee, J.; Harakeh, A.; Waslander, S.L. JoInt. 3D Proposal Generation and Object Detection from View Aggregation. In Proceedings of the IEEE International Conference on Intelligent Robots and Systems, Madrid, Spain, 1–5 October 2018. [Google Scholar]
- Zhou, Y.; Tuzel, O. VoxelNet: End-to-End Learning for PoInt. Cloud Based 3D Object Detection. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
- Lang, A.H.; Vora, S.; Caesar, H.; Zhou, L.; Yang, J.; Beijbom, O. PointPillars: Fast Encoders for Object Detection from PoInt. Clouds. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
- Wang, L.; Huang, Y.; Hou, Y.; Zhang, S.; Shan, J. Graph Attention Convolution for PoInt. Cloud Semantic Segmentation. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
- Bohg, J.; Morales, A.; Asfour, T.; Kragic, D. Data-Driven Grasp Synthesis-A Survey. IEEE Trans. Robot. 2014, 30, 289–309. [Google Scholar] [CrossRef] [Green Version]
- Sundermeyer, M.; Mousavian, A.; Triebel, R.; Fox, D. Contact-GraspNet: Efficient 6-DoF Grasp Generation in Cluttered Scenes. In Proceedings of the IEEE International Conference on Robotics and Automation, Xi’an, China, 30 May–5 June 2021. [Google Scholar]
- Mnih, V.; Heess, N.; Graves, A.; Kavukcuoglu, K. Recurrent Models of Visual Attention. In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada, 8–13 December 2014; Volume 3. [Google Scholar]
- Hu, J.; Shen, L.; Albanie, S.; Sun, G.; Wu, E. Squeeze-and-Excitation Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 42, 2011–2023. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Wang, Q.; Wu, B.; Zhu, P.; Li, P.; Zuo, W.; Hu, Q. ECA-Net: Efficient Channel Attention for Deep Convolutional Neural Networks. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020. [Google Scholar]
- Jaderberg, M.; Simonyan, K.; Zisserman, A.; Kavukcuoglu, K. Spatial Transformer Networks. In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada, 7–12 December 2015. [Google Scholar]
- Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. CBAM: Convolutional Block Attention Module. In Computer Vision—ECCV 2018; Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics); Springer: Cham, Switzerland, 2018; Volume 11211. [Google Scholar]
- Phan, A.V.; Nguyen, M.L.; Nguyen, Y.L.H.; Bui, L.T. DGCNN: A Convolutional Neural Network over Large-Scale Labeled Graphs. Neural Netw. 2018, 108, 533–543. [Google Scholar] [CrossRef] [PubMed]
- Guo, M.H.; Cai, J.X.; Liu, Z.N.; Mu, T.J.; Martin, R.R.; Hu, S.M. PCT: PoInt. Cloud Transformer. Comput. Vis. Media 2021, 7, 187–199. [Google Scholar] [CrossRef]
- He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
- Ferrari, C.; Canny, J. Planning Optimal Grasps. In Proceedings of the IEEE International Conference on Robotics and Automation, Nice, France, 12–14 May 1992; Volume 3. [Google Scholar]
- Pokorny, F.T.; Kragic, D. Classical Grasp Quality Evaluation: New Algorithms and Theory. In Proceedings of the IEEE International Conference on Intelligent Robots and Systems, Tokyo, Japan, 3–7 November 2013. [Google Scholar]
- Qin, Y.; Chen, R.; Zhu, H.; Song, M.; Xu, J.; Su, H. S4G: Amodal Single-View Single-Shot SE(3) Grasp Detection in Cluttered Scenes. In Proceedings of the Conference on Robot Learning(CoRL), Osaka, Japan, 30 October–1 November 2019. [Google Scholar]
Model | Module | Accuracy |
---|---|---|
PointNet (vanilla) | none | 86.95% |
PointNet | input | 87.59% |
PointNet | feature | 86.77% |
PointNet | feature + reg. | 88.12% |
PointNet | both | 89.71% |
Model | AM | Accuracy |
---|---|---|
PointNet | - | 89.71% |
PointNet | 91.30% | |
PointNet | 90.41% | |
PointNet | 90.62% | |
PointNet | 90.13% | |
PointNet | 89.53% | |
PointNet | 89.32% |
Model | Location | Accuracy |
---|---|---|
PointNet | - | 89.71% |
PointNet | I | 90.34% |
PointNet | II | 91.30% |
PointNet | III | 90.89% |
PointNet | IV | 90.64% |
PointNet | V | 89.25% |
PointNet | VI | 88.92% |
Method | Input Data | Year | Accuracy |
---|---|---|---|
GPD (3 channels) [13] | point cloud | 2017 | 79.71% |
GPD (12 channels) [13] | point cloud | 2017 | 86.34% |
S4G [37] | point cloud | 2019 | 87.11% |
Contact-GraspNet [26] | point cloud | 2021 | 90.25% |
Ours | point cloud | 2022 | 91.30% |
Method | Banana | Glasses-Case | Medicine Bottle | Packing Box | Orange | Rubber | Rubik’s Cube | Pen | Medicine Box | Tea Bottle | Average |
---|---|---|---|---|---|---|---|---|---|---|---|
GPD (3 channels) | 85.00% | 80.00% | 85.00% | 80.00% | 85.00% | 75.00% | 90.00% | 75.00% | 80.00% | 80.00% | 81.50% |
GPD (12 channels) | 95.00% | 80.00% | 85.00% | 80.00% | 90.00% | 80.00% | 95.00% | 85.00% | 90.00% | 90.00% | 87.00% |
S4G | 100.00% | 85.00% | 85.00% | 90.00% | 90.00% | 80.00% | 100.00% | 85.00% | 95.00% | 90.00% | 90.00% |
Contact-GraspNet | 100.00% | 90.00% | 90.00% | 90.00% | 95.00% | 85.00% | 100.00% | 85.00% | 95.00% | 95.00% | 92.50% |
Ours | 100.00% | 90.00% | 90.00% | 90.00% | 95.00% | 90.00% | 100.00% | 90.00% | 95.00% | 95.00% | 93.50% |
Method | Success Rate | Completion Rate | Time Efficiency |
---|---|---|---|
GPD (3 channels) | 66.00% | 82.00% | 22,697 ms |
GPD (12 channels) | 71.00% | 89.00% | 25,712 ms |
S4G | 78.00% | 91.00% | 8159 ms |
Contact-GraspNet | 81.00% | 94.00% | 12,861 ms |
Ours | 81.00% | 95.00% | 13,296 ms |
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2022 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Wang, Z.; Li, S.; Bai, Q.; Song, Q.; Zhang, X.; Pu, R. Research on Intelligent Robot Point Cloud Grasping in Internet of Things. Micromachines 2022, 13, 1999. https://doi.org/10.3390/mi13111999
Wang Z, Li S, Bai Q, Song Q, Zhang X, Pu R. Research on Intelligent Robot Point Cloud Grasping in Internet of Things. Micromachines. 2022; 13(11):1999. https://doi.org/10.3390/mi13111999
Chicago/Turabian StyleWang, Zhongyu, Shaobo Li, Qiang Bai, Qisong Song, Xingxing Zhang, and Ruiqiang Pu. 2022. "Research on Intelligent Robot Point Cloud Grasping in Internet of Things" Micromachines 13, no. 11: 1999. https://doi.org/10.3390/mi13111999
APA StyleWang, Z., Li, S., Bai, Q., Song, Q., Zhang, X., & Pu, R. (2022). Research on Intelligent Robot Point Cloud Grasping in Internet of Things. Micromachines, 13(11), 1999. https://doi.org/10.3390/mi13111999