An Optimized Convolutional Neural Network for the 3D Point-Cloud Compression
Abstract
:1. Introduction
- First, we have evaluated the compression effect by altering the parameters of depth, stride, and activation function of the neural networks. The experimental results show that the Sigmoid function outperforms the other activation functions.
- Second, we have proposed an optimized point-cloud compression scheme to enhance the effectiveness of the point-cloud compression.
2. Related Work
3. Methods
3.1. Model Training
3.1.1. Neural Network Structure
3.1.2. Activation Functions
- (1)
- Sigmoid.The Sigmoid function is also known as the logistic function, which is used for the output of hidden layer neurons and the binary classification. The value range is (0, 1), which can map a real number to the interval of (0, 1). The effect is more preferable when the feature difference is complex or not large.
- (2)
- Tanh.The Tanh activation function is also named as the hyperbolic tangent activation function. Similar to the Sigmoid function, the Tanh function also uses the truth values, and the Tanh function converts it to the range of −1 to 1. Therefore, the output of the Tanh function is zero-centered, which solves the problem of slow convergence of Sigmoid function and improves the convergence speed compared with Sigmoid.
- (3)
- Relu.The Relu function converts the output of some neurons to be 0, which causes the sparsity of the network. Thus, the Relu function has the advantages of reducing the parameter’s interdependence and alleviating the occurrence of over-fitting problems.
- (4)
- Leaky Relu.Compared to the Relu function, the output of this function is no longer 0 under the condition that x is negative. Here, a is usually taken as 0.01. Thus, it solves the problem of gradient disappearance in the Relu activation function, which also shows the advantages of efficient computation and faster convergence rates than the Sigmoid/Tanh function.
- (5)
- EluThe exponential linear unit (Elu) activation function is proved to have high noise robustness and can convert the average activation value of neurons that approach zero. Due to the necessity to calculate the index, the calculation cost is high.
- (6)
- GeluThe Gaussian error linear units (Gelu) function introduces the idea of stochastic regularity in activation, a probabilistic description of neuronal input that intuitively complies with our natural understanding.
- (7)
- SeluThe scaled empirical linear units (Selu) function introduces the attribute of self-normalization, which avoids the problem of sudden disappearance or the explosive growth of the gradient. Therefore, this function enables the learning process to be more stable than other functions mentioned above.
- (8)
- R-ReluThe Random Relu (R-Relu) function is also a variant of Leaky Relu, where the slope of negative value is random in training.
- (9)
- HardSwitchThe HardSwitch activation function is used to approximate the Switch activation function. This function introduces subsection calculation, which greatly reduces the amount of calculation under the condition. This allows one to retain the feature, and thus the Switch function can enable the neural network layer to have more expressive ability.
3.1.3. Training
3.2. Network Outputs
3.3. Loss Function
4. Result and Discussion
4.1. Evaluation of the Compression by Different Depth of The Network
4.2. Evaluation of the Compression by Different Number of Strides
4.3. Evaluation of the Compression of Different Activation Functions
5. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
Appendix A
References
- Popişter, F.; Popescu, D.; Păcurar, A.; Păcurar, R. Mathematical Approach in Complex Surfaces Toolpaths. Mathematics 2021, 9, 1360. [Google Scholar] [CrossRef]
- Zhao, J.; Xu, S.; Zhang, B.; Gu, J.; Doermann, D.; Guo, G. Towards compact 1-bit CNNs via Bayesian learning. Int. J. Comput. Vis. 2022, 130, 201–225. [Google Scholar] [CrossRef]
- Li, B.; Wu, B.; Su, J.; Wang, G. Eagleeye: Fast sub-net evaluation for efficient neural network pruning. In Proceedings of the Computer Vision—ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Proceedings, Part II 16. Springer: Berlin/Heildeberg, Germany, 2020; pp. 639–654. [Google Scholar]
- Zhang, B.; Wang, R.; Wang, X.; Han, J.; Ji, R. Modulated convolutional networks. IEEE Trans. Neural Netw. Learn. Syst. 2021, 1–14. [Google Scholar] [CrossRef] [PubMed]
- Yeom, J.; Stan, T.; Hong, S.; Voorhees, P.W. Segmentation of experimental datasets via convolutional neural networks trained on phase field simulations. Acta Mater. 2021, 214, 116990. [Google Scholar] [CrossRef]
- Qayyum, W.; Ehtisham, R.; Bahrami, A.; Camp, C.; Mir, J.; Ahmad, A. Assessment of Convolutional Neural Network Pre-Trained Models for Detection and Orientation of Cracks. Materials 2023, 16, 826. [Google Scholar] [CrossRef]
- Wang, Y.; Sun, Y.; Liu, Z.; Sarma, S.E.; Bronstein, M.M.; Solomon, J.M. Dynamic graph cnn for learning on point clouds. Acm Trans. Graph. 2019, 38, 1–12. [Google Scholar] [CrossRef] [Green Version]
- Liu, Z.; Tang, H.; Lin, Y.; Han, S. Point-voxel cnn for efficient 3d deep learning. Adv. Neural Inf. Process. Syst. 2019, 32. [Google Scholar] [CrossRef]
- Xu, S.; Li, Y.; Zhao, J.; Zhang, B.; Guo, G. Poem: 1-bit point-wise operations based on expectation-maximization for efficient point-cloud processing. arXiv 2021, arXiv:2111.13386. [Google Scholar]
- Alkhouly, A.A.; Mohammed, A.; Hefny, H.A. Improving the performance of deep neural networks using two proposed activation functions. IEEE Access 2021, 9, 82249–82271. [Google Scholar] [CrossRef]
- Liu, J.; Liu, Y.; Zhang, Q. A weight initialization method based on neural network with asymmetric activation function. Neurocomputing 2022, 483, 171–182. [Google Scholar] [CrossRef]
- Kumar, A.; Sodhi, S.S. Neural network with NewSigmoid activation function. J. Intell. Fuzzy Syst. 2022, 43, 545–559. [Google Scholar] [CrossRef]
- Siegel, J.W.; Xu, J. Approximation rates for neural networks with general activation functions. Neural Netw. 2020, 128, 313–321. [Google Scholar] [CrossRef] [PubMed]
- Guede, C.; Andrivon, P.; Marvie, J.E.; Ricard, J.; Redmann, B.; Chevet, J.C. V-PCC: Performance evaluation of the first MPEG Point Cloud Codec. In Proceedings of the SMPTE 2020 Annual Technical Conference and Exhibition, SMPTE, Virtual, 10–12 November 2020; pp. 1–27. [Google Scholar]
- Dumic, E.; da Silva Cruz, L.A. point-cloud coding solutions, subjective assessment and objective measures: A case study. Symmetry 2020, 12, 1955. [Google Scholar] [CrossRef]
- Garcia, D.C.; de Queiroz, R.L. Intra-frame context-based octree coding for point-cloud geometry. In Proceedings of the 2018 25th IEEE International Conference on Image Processing (ICIP), Athens, Greece, 7–10 October 2018; IEEE: Piscatawaj, NJ, USA, 2018; pp. 1807–1811. [Google Scholar]
- de Queiroz, R.L.; Garcia, D.C.; Chou, P.A.; Florencio, D.A. Distance-based probability model for octree coding. IEEE Signal Process. Lett. 2018, 25, 739–742. [Google Scholar] [CrossRef]
- Thanou, D.; Chou, P.A.; Frossard, P. Graph-based compression of dynamic 3D point-cloud sequences. IEEE Trans. Image Process. 2016, 25, 1765–1778. [Google Scholar] [CrossRef] [Green Version]
- de Queiroz, R.L.; Chou, P.A. Motion-compensated compression of point cloud video. In Proceedings of the 2017 IEEE International Conference on Image Processing (ICIP), Beijing, China, 17–20 September 2017; IEEE: Piscatawaj, NJ, USA, 2017; pp. 1417–1421. [Google Scholar]
- Mekuria, R.; Blom, K.; Cesar, P. Design, implementation, and evaluation of a point cloud codec for tele-immersive video. IEEE Trans. Circuits Syst. Video Technol. 2016, 27, 828–842. [Google Scholar] [CrossRef] [Green Version]
- Wu, Z.; Song, S.; Khosla, A.; Yu, F.; Zhang, L.; Tang, X.; Xiao, J. 3d shapenets: A deep representation for volumetric shapes. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–15 June 2015; pp. 1912–1920. [Google Scholar]
- Valenzise, G.; Purica, A.; Hulusic, V.; Cagnazzo, M. Quality assessment of deep-learning-based image compression. In Proceedings of the 2018 IEEE 20th International Workshop on Multimedia Signal Processing (MMSP), Vancouver, BC, Canada, 29–31 August 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 1–6. [Google Scholar]
- Huang, T.; Liu, Y. 3d point cloud geometry compression on deep learning. In Proceedings of the 27th ACM International Conference on Multimedia, Nice, France, 21–25 October 2019; pp. 890–898. [Google Scholar]
- Rusu, R.B.; Cousins, S. 3d is here: Point cloud library (pcl). In Proceedings of the 2011 IEEE International Conference on Robotics and Automation, Shanghai, China, 09–13 May 2011; IEEE: Piscatawaj, NJ, USA, 2011; pp. 1–4. [Google Scholar]
- de Hoog, J.; Ahmed, A.N.; Anwar, A.; Latré, S.; Hellinckx, P. Quality-aware compression of point clouds with google draco. In Proceedings of the 16th International Conference on P2P, Parallel, Grid, Cloud and Internet Computing (3PGCIC-2021), Fukuoka, Japan, 28–30 October 2021; Springer: Cham, Switzerland, 2022; pp. 227–236. [Google Scholar]
- Achlioptas, P.; Diamanti, O.; Mitliagkas, I.; Guibas, L. Learning representations and generative models for 3d point clouds. In Proceedings of the International Conference on Machine Learning, PMLR, Stockholm, Sweden, 10–15 July 2018; pp. 40–49. [Google Scholar]
- Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets Advances in neural information processing systems. arXiv 2014, arXiv:1406.2661. [Google Scholar]
- Ballé, J.; Laparra, V.; Simoncelli, E.P. End-to-end optimized image compression. arXiv 2016, arXiv:1611.01704. [Google Scholar]
- Quach, M.; Valenzise, G.; Dufaux, F. Learning convolutional transforms for lossy point cloud geometry compression. In Proceedings of the 2019 IEEE International Conference on Image Processing (ICIP), Taipei, Taiwan, 22–25 September 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 4320–4324. [Google Scholar]
- Karim, A.M.; Kaya, H.; Alcan, V.; Sen, B.; Hadimlioglu, I.A. New optimized deep learning application for COVID-19 detection in chest X-ray images. Symmetry 2022, 14, 1003. [Google Scholar] [CrossRef]
- Sung, Y.H.; Park, S.J.; Kim, D.Y.; Kim, S. GPS Spoofing Detection Method for Small UAVs Using 1D Convolution Neural Network. Sensors 2022, 22, 9412. [Google Scholar] [CrossRef]
- Arenas, R. Design of a Forest Fire Early Alert System through a Deep 3D-CNN Structure and a WRF-CNN Bias Correction. Sensors 2022, 22, 8790. [Google Scholar]
- Liu, C.; Ding, W.; Chen, P.; Zhuang, B.; Wang, Y.; Zhao, Y.; Zhang, B.; Han, Y. RB-Net: Training highly accurate and efficient binary neural networks with reshaped point-wise convolution and balanced activation. IEEE Trans. Circuits Syst. Video Technol. 2022, 32, 6414–6424. [Google Scholar] [CrossRef]
- Xu, S.; Li, Y.; Wang, T.; Ma, T.; Zhang, B.; Gao, P.; Qiao, Y.; Lü, J.; Guo, G. Recurrent bilinear optimization for binary neural networks. In Proceedings of the Computer Vision–ECCV 2022: 17th European Conference, Tel Aviv, Israel, 23–27 October 2022; Proceedings, Part XXIV. Springer: Cham, Switzerland, 2022; pp. 19–35. [Google Scholar]
- Jin, T.; Dai, H.; Cao, L.; Zhang, B.; Huang, F.; Gao, Y.; Ji, R. Deepwalk-aware graph convolutional networks. Sci. China Inf. Sci. 2022, 65, 152104. [Google Scholar] [CrossRef]
- Zeng, B.; Liu, B.; Li, H.; Liu, X.; Liu, J.; Chen, D.; Peng, W.; Zhang, B. FNeVR: Neural Volume Rendering for Face Animation. arXiv 2022, arXiv:2209.10340. [Google Scholar]
- Sedaghat, N.; Zolfaghari, M.; Amiri, E.; Brox, T. Orientation-boosted voxel nets for 3D object recognition. arXiv 2016, arXiv:1604.03351. [Google Scholar]
- Mekuria, R.; Li, Z.; Tulvan, C.; Chou, P. Evaluation Criteria for PCC (Point-Cloud Compression); DANS: The Hague, The Netherlands, 2016. [Google Scholar]
Accuracy | ||||
---|---|---|---|---|
Function | ||||
Relu | 0.00831488 | 0.0139485 | 0.0181215 | |
Selu | 0.00893118 | 0.0151461 | 0.0208207 | |
Elu | 0.00664042 | 0.0118255 | 0.0128144 | |
Tanh | 0.00688373 | 0.0094999 | 0.0245558 | |
Leaky Relu | 0.00744641 | 0.0097154 | 0.0153705 | |
R-Relu | 0.00840702 | 0.0109112 | 0.0121359 | |
Gelu | 0.00805524 | 0.0096145 | 0.0102793 | |
HardSwitch | 0.00719350 | 0.0176457 | 0.0181156 | |
sigmoid | 0.00431234 | 0.0078694 | 0.0089128 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Luo, G.; He, B.; Xiong, Y.; Wang, L.; Wang, H.; Zhu, Z.; Shi, X. An Optimized Convolutional Neural Network for the 3D Point-Cloud Compression. Sensors 2023, 23, 2250. https://doi.org/10.3390/s23042250
Luo G, He B, Xiong Y, Wang L, Wang H, Zhu Z, Shi X. An Optimized Convolutional Neural Network for the 3D Point-Cloud Compression. Sensors. 2023; 23(4):2250. https://doi.org/10.3390/s23042250
Chicago/Turabian StyleLuo, Guoliang, Bingqin He, Yanbo Xiong, Luqi Wang, Hui Wang, Zhiliang Zhu, and Xiangren Shi. 2023. "An Optimized Convolutional Neural Network for the 3D Point-Cloud Compression" Sensors 23, no. 4: 2250. https://doi.org/10.3390/s23042250
APA StyleLuo, G., He, B., Xiong, Y., Wang, L., Wang, H., Zhu, Z., & Shi, X. (2023). An Optimized Convolutional Neural Network for the 3D Point-Cloud Compression. Sensors, 23(4), 2250. https://doi.org/10.3390/s23042250