Next Article in Journal
New Smart Sensor for Voltage Unbalance Measurements in Electrical Power Systems
Next Article in Special Issue
MASPC_Transform: A Plant Point Cloud Segmentation Network Based on Multi-Head Attention Separation and Position Code
Previous Article in Journal
Physical Load While Using a Tablet at Different Tilt Angles during Sitting and Standing
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Single Stage and Single View 3D Point Cloud Reconstruction Network Based on DetNet

School of Computer Science, Northeast Electric Power University, Jilin 132011, China
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(21), 8235; https://doi.org/10.3390/s22218235
Submission received: 22 September 2022 / Revised: 25 October 2022 / Accepted: 26 October 2022 / Published: 27 October 2022
(This article belongs to the Special Issue Intelligent Point Cloud Processing, Sensing and Understanding)

Abstract

:
It is a challenging problem to infer objects with reasonable shapes and appearance from a single picture. Existing research often pays more attention to the structure of the point cloud generation network, while ignoring the feature extraction of 2D images and reducing the loss in the process of feature propagation in the network. In this paper, a single-stage and single-view 3D point cloud reconstruction network, 3D-SSRecNet, is proposed. The proposed 3D-SSRecNet is a simple single-stage network composed of a 2D image feature extraction network and a point cloud prediction network. The single-stage network structure can reduce the loss of the extracted 2D image features. The 2D image feature extraction network takes DetNet as the backbone. DetNet can extract more details from 2D images. In order to generate point clouds with better shape and appearance, in the point cloud prediction network, the exponential linear unit (ELU) is used as the activation function, and the joint function of chamfer distance (CD) and Earth mover’s distance (EMD) is used as the loss function of 3DSSRecNet. In order to verify the effectiveness of 3D-SSRecNet, we conducted a series of experiments on ShapeNet and Pix3D datasets. The experimental results measured by CD and EMD have shown that 3D-SSRecNet outperforms the state-of-the-art reconstruction methods.

1. Introduction

In many tasks, such as virtual reality [1], experimental assistance [2], and robot navigation [3], a detailed 3D model is required, however, the facilities required to sample 3D models from the real world are costly. Moreover, it is uneconomical to manually reconstruct 3D models from 2D maps on a large scale. Many researchers proposed methods to reconstruct 3D models from a single image [4,5,6].
There are mainly two structures to represent a 3D model, voxel and point cloud. The former is just like 2D pixels but fits 3D objects into grids, sometimes containing other information such as features. It is also a regular data structure so many successful 2D methods can be easily applied. Many approaches [7,8] have focused on the voxel grid as output. However, the computational cost increases cubically to perform better geometric information or apply convolution methods.
Point cloud represents geometric information by a set of data points, each point represented by (x, y, z). Fan et al. [4] firstly applied deep learning methods to generate point clouds and proposed the chamfer distance and Earth mover’s distance, but there are many methods to improve its evaluation results. 3D-LMNet [5] used an autoencoder to design a two-stage point cloud construction network. Most of the existing works are similar to 3D-LMNet, which are multi-stage point cloud generation networks. These multi-stage networks inevitably suffered feature loss and were considered a waste of time. At the same time, the existing work often pays more attention to the structure of the point cloud generation network, while ignoring the feature extraction of 2D images.
The main tasks of point cloud reconstruction are: (1) To retain more details or small targets in the image when extracting 2D image features in order to obtain a better reconstruction effect. (2) To generate a point cloud through a simple network structure to reduce the loss of features in the process of transmission in different network stages. We propose an end-to-end point cloud reconstruction network called 3D-SSRecNet, which applies DetNet [9] as an image feature extractor and gains detailed features.
The key contributions of our work are as follows:
  • We propose a one-stage neural network for 3D reconstruction from a single image, namely, 3D-SSRecNet. 3D-SSRecNet takes an image as input and directly outputs the predicted point cloud without further processing.
  • 3D-SSRecNet includes feature extraction and 3D point cloud generation. The feature extraction network is better at extracting the detailed features of the 2D input. The point cloud generation network has a plain structure and uses a suitable activation function in its multi-layer perceptron, which reduces the loss of features during forwarding propagation to obtain an elaborate output.
  • Experiments on ShapeNet and pix3D dataset have shown that 3D-SSRecNet outperforms the state-of-art reconstruction methods for the task of single-view reconstruction. At the same time, we also proved the validity of the activation function of the point cloud generation network through experiments.

2. Related Work

The technology of reconstructing 3D models from 2D images has many practical applications. Therefore, 2D to 3D reconstruction technologies applicable to different application scenarios will be quite different. For example, in [2], a method is proposed to predict the liquid or solid in transparent vessels to XYZ maps. This method can be applied, for example, to the task of a robot arm taking containers and pouring liquid. Therefore, this study proposed a scale-invariant loss so that the predicted scale of XYZ map can conform to the original 2D input scale. Part-Wise AtlasNet proposed in [10] can output 3D reconstruction with a fine local structure. This is because each neural network of Part-Wise AtlasNet is only responsible for reconstructing a specific part of the 3D object. Part-Wise AtlasNet has achieved a very refined reconstruction effect, but the reconstruction process is very time consuming. Part-Wise AtlasNet is obviously more suitable for tasks such as high-precision reconstruction and the display of cultural relics. In order to obtain 3D data in real time and accurately, hardware assistance is required in addition to consideration of the camera parameters. For example, SLAM [11] constructs a 3D map by positioning the camera in real time. In virtual reality or game modeling tasks, a point cloud reconstruction network may be required to reconstruct point clouds with reasonable contour and shape without considering hardware parameters. The 3D-SSRecNet proposed in this paper is more suitable for virtual reality or game modeling scenarios. Here, we summarize the references similar to the application scenarios of 3D-SSRecNet. Due to the development of deep learning [12] and big data [13], great progress has been made in this field, and numerous valuable research has emerged. These studies can be roughly divided into 2D images to voxels and 2D images to point clouds.
From 2D images to voxels. Several approaches focus on generating voxelized output representations. V3DOR [14] applied autoencoders and variational autoencoders to generate a smoother and high-resolution 3D model. The encoder aims to learn latent representation from the image and the decoder tries to obtain corresponding 3D voxels. Xie et al. [15] applied ResNet [16] as a part of the autoencoder and proposed a multi-scale context-aware fusion module to gain better results with more views of the object. Han et al. [17] proposed a novel shape representation that enabled a tube-by-tube manner via discriminative neural networks. Based on the network, they proposed an RNN-based model to gain the 3D corresponding representation from the input image. TMVNet [18] applied the transformers to the encoder and proposed a 3D feature fusion layer to refine the predictions. Kniaz et al. [19] proposed an image-to-voxel translation model which applied a generative adversarial network. Sym3DNet [20] applied a symmetry fusion step and perceptual loss to apply symmetry prior. Yang et al. [21] designed a memory-based framework to obtain a heavily occluded 3D model to handle challenging situations. However, voxel reconstruction gains sparse space information and has a high costs. It is difficult to both predict higher resolution 3D models and process them efficiently.
From 2D images to point clouds. Fan et al. [1] designed a framework called PSGN, which firstly applied deep learning methods to the point sets generation problem. They proposed chamfer distance and Earth mover’s distance to judge the distance between point sets. 3D-LMNet [2] trained a point cloud autoencoder, then try to map images to corresponding learned embedding and they proposed diversity loss for uncertain reconstruction. 3D-ARNet [22] combined an autoencoder and a point prediction network. After the image is input into the image encoder, a simple point cloud is obtained. Pumarola et al. [23] proposed a conditional flow-based generative model to generate a map from image to point cloud, which is different from other generative models such as VAEs or GAN. Hafiz et al. [24] proposed the SE-MD network. SE-MD uses an autoencoder network as the feature extraction network and multiple decoding networks as point cloud generation networks. The final result can be obtained by fusing all the outputs of all the point cloud generation networks. However, these multi-stage models may suffer more cost of computational resources and feature loss when the feature maps propagate across networks. 3D-ReConstnet [6] applied the residual network to extract the features from the input image and used MLPs to predict point sets and, meanwhile, learned Gaussian probability distribution to refine the self-occluded part of an object and then directly applied MLPs to predict the point cloud. Ping et al. [25] projected the predicted point cloud and tried to fit edge details with ground truth. In order to enhance the features of 2D images, 3D-FEGNet [26] adds an edge extraction module to the feature extraction network. After comparing the reconstruction results of the above networks, we found that the key to a better reconstruction network are: (1) the feature extraction part of the network reflects more detailed 2D image features; (2) the loss of the extracted two-dimensional image features in the network transmission is minimized. This paper designs a single-stage point cloud reconstruction network and uses DetNet, which can retain more detailed features, as the feature extraction network.

3. Approach

3.1. Architecture of 3D-SSRecNet

The architecture of 3D-SSRecNet is shown in Figure 1. 3D-SSRecNet has two main parts: a 2D image feature extraction network and a point cloud prediction network. These two parts constitute a simple single-stage point cloud reconstruction network. The single-stage network structure only transfers the features of 2D pictures within the network of 3D-SSRecNet. Compared with the two-stage reconstruction network that needs to transmit features across the network, this network structure reduces the loss of features.
Given a 2D image, firstly, we obtain a latent representation V by DetNet. Then, we map V to a low dimensionality feature V’ by a full connection (FC) layer. A multi-layer perceptron (MLP) is directly applied to predict a point set afterward. During training, chamfer distance and Earth mover’s distance loss function are computed, and the update of trainable parameters is supervised.

3.2. 2D Image Feature Extraction

Many image feature networks applied downsampling, which brought a higher receptive field, but unavoidably caused the loss of image details. However, for reconstruction, image details are crucial for the recovery of geometric shape. This kind of network is more suitable for image classification tasks, but not for reconstruction tasks that require more detailed features.
DetNet [9] not only retains more details but also retains a large receptive field. Although DetNet was designed for object detection, its novel dilated bottleneck structure provides high-resolution feature maps and a large receptive field. We use DetNet as the backbone of image feature extraction. DetNet follows the same structure as ResNet-50 [16] until state 4, so DetNet also has the advantage of being easy to train and will not fall into gradient disappearance. Table 1 shows the parameters of the last two stages of DetNet, that is, the differences between DetNet and ResNet-50. After stage 4, DetNet keeps the size of the feature map at 16 × 16, which enables more details to be retained. The fifth and sixth stages of DetNet are composed of bottlenecks with dilated convolution, and some bottlenecks have 1 × 1 convolutions on their shortcut connections. Dilated convolution increases the receptive field. However, considering the amount of computation and memory, stage 5 and stage 6 set the same channel number of 256. At the end of the baseline, a fully connected layer is applied.
As shown in Figure 1, after feature extraction of the input image, we obtain a 1000-dimensional latent feature V of the input image. After that, the full connection (FC) layer compresses the dimension of vector V from 1000 to 100 and obtains vector V’.

3.3. Point Could Prediction

We use three-layer MLP to predict the point sets directly. The dimensions of the outputs of the three MLP layers are 512, 1024, and N × 3, respectively. The output of the feature extraction network: vector V’ is fed into MLPs of the point cloud prediction network. On the first two layers, ELU [27] is introduced as an activation function which is defined as:
y = elu x = x ,   x 0 α e x 1 , x < 0
where parameter α is set to 1. The curves of ELU activation function and its derivative are shown in Figure 2a,b, respectively. For common activation functions, such as ReLU, the value corresponding to the negative axis is 0. However, the normalized point cloud coordinate interval is [−1, 1], which indicates that the point cloud coordinates will have negative values. As shown in Figure 2a, the value corresponding to the negative axis of the ELU activation function is non-zero. Therefore, using ELU as the activation function, the negative value information in the reconstructed network will not be lost in the forward propagation process. As shown in Figure 3b, the derivative of ELU is also non-zero on the negative axis. In the backpropagation process of the network, the negative gradient will not be lost, and it can help update the network weight.
In the experimental part, we prove that the reconstruction effect of using ELU as the activation function is better than that of using other activation functions. We directly output the predicted point sets after the last activation layer. It is actualized by the tanh function, whose outputs belong to [−1, 1], which is the same as the required point set data.

3.4. Loss Function

To define the loss function of reconstruction of point clouds. We have to consider two important properties. (1) Point cloud is an unordered point set so we shall obtain the same data despite how we change the order of points. (2) The geometric feature of a real object shall not change significantly regardless of any rotation transformations. However, for instance, the point coordinates seem different as we take a rotation transformation. Fan et al. [4] proposed the use of the chamfer distance (CD) and Earth mover’s distance (EMD) which satisfied the requirement. The chamfer distance is defined below.
  L C D X P * , X P = x X P * m i n y X P | | x y | | 2 2 + y X P * m i n x X P | | x y | | 2 2
It is a differentiable function with respect to point locations which are respectively calculated for each point, and it obtains the distance between it and its nearest point in the ground truth.
The Earth mover’s distance is defined below.
L EMD X P * , X P = m i n φ : X P * X P x X P * | | x φ x | | 2  
It is a measure between two distributions, so it can be considered as a “distance” between two point sets, and it constitutes the mapping between two point sets to guarantee the unique output.
Many research on point cloud reconstruction [5,17] point out the characteristics of CD and EMD: CD is related to the contour of the reconstructed point cloud. Lower CD values result in better point cloud contours. The 3D reconstruction network trained by CD can more easily capture the rough contour of objects in 2D images. However, the reconstructed network trained by CD can easily produce splash point clouds, but the visual effect is not good.
We use Figure 3 to explain the cause of point clouds. In Figure 3, the blue dot represents the ground truth, and the green dot and red dot represent two different reconstruction point clouds, respectively. D1~D6 respectively represent the distance between the six points obtained from the first reconstruction and their corresponding ground truth points. D1’~D6’ respectively represent the distance between the six points obtained from the second reconstruction and their corresponding ground truth points. We can see that the reconstruction result represented by the green point in Figure 3a is better than the reconstruction result represented by the red point in Figure 3a because the distance D1’ between one of the red points and its corresponding ground truth point is obviously greater than D2’~D6‘. From Formula (2), if the sum of D1~D6 is equal to the sum of D1’~D6’, the network trained with CD as the loss function cannot distinguish between the two reconstructions. Therefore, we can say that CD may confuse different reconstructions with similar chamfer distances. However, in the EMD loss function, φ represents the bijection relationship between the ground truth and the reconstructed point cloud, so EMD loss will not cause the above confusion.
EMD is related to the visual quality of the reconstructed point cloud. Lower EMD value always presumes higher visual quality. However, it is inclined to obtain a bad contour of the object. Synthesizing the pros and cons of CD and EMD, the loss function of our network is defined as:
L o s s = L C D + L EMD

4. Experiment

We evaluated the proposed 3D-SSRecNet on ShapeNet [28] and Pix3D [29] datasets, respectively. ShapeNet is a big collection of textured CAD models which consists of 13 classes and 43,809 point cloud models for both training and testing. We used the 80–20% train/test split to perform our experiment. We performed the same experiment on the Pix3D database. The Pix3D database consists of three classes and 7595 point cloud models. This dataset is a CAD model of the real scene. Experiments on Pix3D can better evaluate the practicability of a point cloud reconstruction algorithm.
We used the gradient optimization algorithm Adam to optimize the proposed 3D-SSRecNet. In training, we set the learning rate to 0.0005 and the epoch to 50. The training environment is as follows: Ubuntu 18.04.6, CUDA 10.1, and the model of GPU is NVIDIA Tesla T4 × 4. We used the CD and EMD values calculated on 1024 sampling points to evaluate the quality of the reconstructed point cloud.

4.1. Experiment on ShapeNet

To verify the advantage of ELU activation function, we fixed the structure of 3D-SSRecNet and replaced ELU with other activation functions such as Leaky ReLU, softsign, and softplus. Table 2 shows the reconstruction results using different activation functions. Among all the 13 classes, ELU performed better in 12 categories.
Figure 4 shows the point cloud reconstruction results obtained using different activation functions on the ShapeNet dataset. We can see that the network trained by ELU considers details better and preserves finer geometric information.
We compared our 3D-SSRecNet with PSGN [4], 3D-LMNet [5], SE-MD [24], 3D-VENet [25], 3D-ARNet [22], 3D-ReConstnet [6], and 3D-FEGNet [26]. All the experiments followed the same train/test split and used the same loss functions, CD and EMD. Table 3 and Table 4 show the reconstruction results of different reconstruction methods on the ShapeNet dataset. The smaller the values of CD and EMD of a method, the better the reconstruction quality of the method. From the results shown in Table 3, the CD values of 3D-SSRecNet are slightly lower than that of 3D-Renstnet in the two categories of cabinet and monitor. In other categories, the reconstruction effect of 3D-SSRecNet is the best. In Table 3, PSGN, 3D-LMNet, SE-MD, and 3D-ARNet are two-stage networks, while other networks are single-stage networks. It can be seen from the CD values shown in Table 3 that the performance of single-stage networks in most categories is better than that of two-stage networks. This shows that the propagation of features between different stages of the network will cause feature loss. It can be seen from the EMD values shown in Table 4 that the reconstruction results of SSRecNet in most categories are better than those of other networks.
Figure 5 shows the reconstruction effect of 13 categories of 3D-SSRecNet on the ShapeNet dataset. The reconstructed point cloud resolution shown in Figure 5 is 2048.

4.2. Experiment on Pix3D

The Pix3D dataset consists of a large number of real indoor 2D images and their corresponding metadata (such as masks, ground truth CAD models, and attitudes). It can be seen from Figure 6 that the background of 2D images in Pix3D dataset is very complex, which poses a greater challenge to the feature extraction part of the reconstruction network. Therefore, the Pix3D dataset can be used to evaluate the generalization ability of the reconstructed network to real scenes.
Table 5 shows that ELU also provides better evaluation values on Pix3D. We exhibit the prediction outputs with different activation functions in Figure 6 to visualize how the activation function affects the prediction. Figure 6 shows that ELU also generates better qualitative results.
We also compared our network with PSGN [4], 3D-LMNet [5], 3D-ARNet [22], 3D-Reconstnet [6], and 3D-FEGNet [26]. Table 6 and Table 7 show the reconstruction results of different reconstruction methods on the Pix3D dataset. In all three categories, we can see that our 3D-SSRecNet has the lowest evaluation value (except the EMD value on chair category), which indicates that 3D-SSRecNet has a strong generalization ability for real scenes.
Figure 7 shows the reconstruction effect of 13 categories of 3D-SSRecNet on the Pix3D dataset. The reconstructed point cloud resolution shown in Figure 7 is 2048.

5. Conclusions

In this paper, we proposed an efficient 3D point cloud reconstruction called 3D-SSRecNet. Given an image, it learns the latent representation and after dimensionality reduction we apply MLPs to predict the correspondent point cloud directly. We conducted several experiments on ShapeNet and Pix3D data sets. We proved that the reconstruction effect of using activation function ELU in the generation network is better than that of using other activation functions. That is, the CD and EMD values of the point cloud generated using the ELU are lower than those of the point cloud generated using other activation functions. Figure 4 and Figure 6 also show that the shape and contour of the point cloud generated using ELU are better than those of the point cloud generated using other activation functions. By comparing CD and EMD values, it is proved that 3D-SSRecNet can outperform the state-of-the-art methods. The proposed 3D-SSRecNet is more suitable for tasks that need to obtain the contour and shape of a point cloud at a low cost. In future work, we will try to improve the local effect of the output point cloud with the Part-Wise AtlasNet on the premise of maintaining the existing low cost.

Author Contributions

Conceptualization, methodology, resources, funding acquisition, B.L.; software, validation, visualization, S.Z.; and writing—original draft preparation, writing—review and editing, Y.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Science and Technology Development Plan Project of Jilin Province under grant 20200201165JC.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Garrido, D.; Rodrigues, R.; Augusto Sousa, A.; Jacob, J.; Castro Silva, D. Point Cloud Interaction and Manipulation in Virtual Reality. In Proceedings of the 2021 5th International Conference on Artificial Intelligence and Virtual Reality (AIVR), Kumamoto, Japan, 23–25 July 2021; pp. 15–20. [Google Scholar]
  2. Eppel, S.; Xu, H.; Wang, Y.R.; Aspuru-Guzik, A. Predicting 3D shapes, masks, and properties of materials inside transparent containers, using the TransProteus CGI dataset. Digit. Discov. 2021, 1, 45–60. [Google Scholar] [CrossRef]
  3. Xu, T.X.; Guo, Y.C.; Lai, Y.K.; Zhang, S.H. TransLoc3D: Point Cloud based Large-scale Place Recognition using Adaptive Receptive Fields. arXiv 2021, arXiv:2105.11605v3. [Google Scholar]
  4. Fan, H.; Su, H.; Guibas, L. A Point Set Generation Network for 3D Object Reconstruction from a Single Image. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition 2017, Honolulu, HI, USA, 21–26 July 2017; pp. 2463–2471. [Google Scholar] [CrossRef] [Green Version]
  5. Mandikal, P.; Navaneet, K.L.; Agarwal, M.; Babu, R.V. 3D-LMNet: Latent embedding matching for accurate and diverse 3D point cloud reconstruction from a single image. arXiv 2018, arXiv:1807.07796. [Google Scholar]
  6. Li, B.; Zhang, Y.; Zhao, B.; Shao, H. 3D-ReConstnet: A Single-View 3D-Object Point Cloud Reconstruction Network. IEEE Access 2020, 1, 99. [Google Scholar] [CrossRef]
  7. Gwak, J.; Choy, C.B.; Chandraker, M.; Garg, A.; Savarese, S. Weakly supervised 3D reconstruction with adversarial con-straint. In Proceedings of the 2017 International Conference on 3D Vision (3DV), Qingdao, China, 10–12 October 2017; pp. 263–272. [Google Scholar]
  8. Yang, B.; Wen, H.; Wang, S.; Clark, R.; Markham, A.; Trigoni, N. 3D Object Reconstruction from a Single Depth View with Adversarial Learning. In Proceedings of the IEEE International Conference on Computer Vision Workshops, Venice, Italy, 22–29 October 2017; pp. 679–688. [Google Scholar]
  9. Li, Z.; Peng, C.; Yu, G.; Zhang, X.; Deng, Y.; Sun, J. DetNet: A Backbone network for Object Detection. arXiv 2018, arXiv:1804.06215. [Google Scholar]
  10. Yu, Q.; Yang, C.; Wei, H. Part-Wise AtlasNet for 3D point cloud reconstruction from a single image. Knowl. Based Syst. 2022, 242, 108395. [Google Scholar] [CrossRef]
  11. Fuentes-Pacheco, J.; Ruiz-Ascencio, J.; Rendon-Mancha, J.M. Visual simultaneous localization and mapping: A survey. Artif. Intell. Rev. 2012, 43, 55–81. [Google Scholar] [CrossRef]
  12. Cheikhrouhou, O.; Mahmud, R.; Zouari, R.; Ibrahim, M.; Zaguia, A.; Gia, T.N. One-Dimensional CNN Approach for ECG Arrhythmia Analysis in Fog-Cloud Environments. IEEE Access 2021, 9, 103513–103523. [Google Scholar] [CrossRef]
  13. Kimothi, S.; Thapliyal, A.; Akram, S.V.; Singh, R.; Gehlot, A.; Mohamed, H.G.; Anand, D.; Ibrahim, M.; Noya, I.D. Big Data Analysis Framework for Water Quality Indicators with Assimilation of IoT and ML. Electronics 2022, 11, 1927. [Google Scholar] [CrossRef]
  14. Tahir, R.; Sargano, A.B.; Habib, Z. Voxel-Based 3D Object Reconstruction from Single 2D Image Using Variational Autoencoders. Mathematics 2021, 9, 2288. [Google Scholar] [CrossRef]
  15. Xie, H.; Yao, H.; Sun, X.; Zhou, S.; Zhang, S. Pix2Vox: Context-Aware 3D Reconstruction from Single and Multi-View Images. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019. [Google Scholar]
  16. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  17. Han, Z.; Qiao, G.; Liu, Y.-S.; Zwicker, M. SeqXY2SeqZ: Structure Learning for 3D Shapes by Sequentially Predicting 1D Occupancy Segments from 2D Coordinates. In European Conference on Computer Vision; Springer: Cham, Switzerland, 2020; pp. 607–625. [Google Scholar]
  18. Peng, K.; Islam, R.; Quarles, J.; Desai, K. TMVNet: Using Transformers for Multi-View Voxel-Based 3D Reconstruction. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–20 June 2022; pp. 222–230. [Google Scholar]
  19. Kniaz, V.V.; Knyaz, V.A.; Remondino, F.; Bordodymov, A.; Moshkantsev, P. Image-to-voxel model translation for 3d scene reconstruction and segmentation. In European Conference on Computer Vision; Springer: Cham, Switzerland, 2020; pp. 105–124. [Google Scholar]
  20. Siddique, A.; Lee, S. Sym3DNet: Symmetric 3D Prior Network for Single-View 3D Reconstruction. Sensors 2022, 22, 518. [Google Scholar] [CrossRef] [PubMed]
  21. Yang, S.; Xu, M.; Xie, H.; Perry, S.; Xia, J. Single-View 3D Object Reconstruction from Shape Priors in Memory. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 3151–3160. [Google Scholar] [CrossRef]
  22. Chen, H.; Zuo, Y. 3D-ARNet: An accurate 3D point cloud reconstruction network from a single-image. Multimedia Tools Appl. 2021, 81, 12127–12140. [Google Scholar] [CrossRef]
  23. Pumarola, A.; Popov, S.; Moreno-Noguer, F.; Ferrari, V. C-flow: Conditional generative flow models for images and 3d point clouds. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 7949–7958. [Google Scholar]
  24. Hafiz, A.M.; Bhat, R.U.A.; Parah, S.A.; Hassaballah, M. SE-MD: A Single-encoder multiple-decoder deep network for point cloud generation from 2D images. arXiv 2021, arXiv:2106.15325. [Google Scholar]
  25. Ping, G.; Esfahani, M.A.; Chen, J.; Wang, H. Visual enhancement of single-view 3D point cloud reconstruction. Comput. Graph. 2022, 102, 112–119. [Google Scholar] [CrossRef]
  26. Wang, E.; Sun, H.; Wang, B.; Cao, Z.; Liu, Z. 3D-FEGNet: A feature enhanced point cloud generation network from a single image. IET Comput. Vis. 2022. [Google Scholar] [CrossRef]
  27. Clevert, D.A.; Unterthiner, T.; Hochreiter, S. Fast and Accurate Deep Network Learning by Exponential Linear Units (ELUs). arXiv 2015, arXiv:1511.07289. [Google Scholar]
  28. Chang, A.X.; Funkhouser, T.; Guibas, L.; Hanrahan, P.; Huang, Q.; Li, Z.; Savarese, S.; Savva, M.; Song, S.; Su, H.; et al. ShapeNet: An information-rich 3D model repository. arXiv 2015, arXiv:1512.03012. [Google Scholar]
  29. Sun, X.; Wu, J.; Zhang, X.; Zhang, Z.; Zhang, C.; Xue, T.; Tenenbaum, J.B.; Freeman, W.T. Pix3D: Dataset and Methods for Single-Image 3D Shape Modeling. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition 2018, Salt Lake City, UT, USA, 18–23 June 2018; pp. 2974–2983. [Google Scholar]
Figure 1. Architecture of 3D-SSRecNet.
Figure 1. Architecture of 3D-SSRecNet.
Sensors 22 08235 g001
Figure 2. ELU activation function and its derivative.
Figure 2. ELU activation function and its derivative.
Sensors 22 08235 g002
Figure 3. Different reconstructions with the same CD loss value.
Figure 3. Different reconstructions with the same CD loss value.
Sensors 22 08235 g003
Figure 4. Visualization of 3D-SSRecNet’s output with different activation functions on ShapeNet.
Figure 4. Visualization of 3D-SSRecNet’s output with different activation functions on ShapeNet.
Sensors 22 08235 g004
Figure 5. Visualization of our predictions on ShapeNet.
Figure 5. Visualization of our predictions on ShapeNet.
Sensors 22 08235 g005
Figure 6. Visualization of 3D-SSRecNet’s predictions with different activation functions on Pix3D.
Figure 6. Visualization of 3D-SSRecNet’s predictions with different activation functions on Pix3D.
Sensors 22 08235 g006
Figure 7. Visualization of our predictions on Pix3D.
Figure 7. Visualization of our predictions on Pix3D.
Sensors 22 08235 g007
Table 1. The parameters of last two stages DetNet.
Table 1. The parameters of last two stages DetNet.
StageFeature Map SizeParameters
(Convolution Kernel Size, Number of Output Channels)
Stage 516 × 16 1 × 1 kernel 256 c h a n e l s 3 × 3 kernel , dilate   2 256 c h a n e l s 1 × 1 kernel 256 c h a n e l s + 1 × 1 c o n v
1 × 1 kernel 256 c h a n e l s 3 × 3 kernel , dilate   2 256 c h a n e l s 1 × 1 kernel 256 c h a n e l s × 2
Stage 616 × 16 1 × 1 kernel 256 c h a n e l s 3 × 3 kernel , dilate   2 256 c h a n e l s 1 × 1 kernel 256 c h a n e l s + 1 × 1 c o n v
1 × 1 kernel 256 c h a n e l s 3 × 3 kernel , dilate   2 256 c h a n e l s 1 × 1 kernel 256 c h a n e l s × 2
1 × 116 × 16 average pool, 1000-d fully connected layer
Table 2. Reconstruction results of different activation functions on ShapeNet dataset.
Table 2. Reconstruction results of different activation functions on ShapeNet dataset.
ClassSoftsignSoftplusLeaky ReLUELUSoftsignSoftplusLeaky ReLUELU
C D × 10 2 E M D × 10 2
Airplane2.602.382.372.383.052.722.682.72
Bench3.693.583.553.513.343.273.183.15
Cabinet5.195.094.94.774.284.214.033.91
Car3.583.573.573.562.852.852.852.84
Chair4.454.484.404.354.304.364.234.17
Lamp5.045.214.974.996.366.756.236.19
Monitor5.004.844.734.724.964.774.504.49
Rifle2.522.582.492.453.693.983.483.48
Sofa4.574.624.524.443.823.893.733.71
Speaker6.266.265.995.945.585.655.275.23
Table4.864.834.404.354.754.744.214.16
Telephone3.753.713.573.523.413.523.063.05
Vessel3.853.773.773.724.204.263.983.96
Table 3. Reconstruction results on ShapeNet evaluated by CD.
Table 3. Reconstruction results on ShapeNet evaluated by CD.
ClassPSGN3D-LMNetSE-MD3D-VENet3D-ARNet3D-Reconstnet3D-FEGNetOurs
C D × 10 2
Airplane3.743.343.113.092.982.422.362.38
Bench4.634.554.344.264.443.573.603.51
Cabinet6.986.095.895.496.014.664.844.77
Car5.24.554.524.304.273.593.573.56
Chair6.396.416.475.765.944.414.354.35
Lamp6.337.107.086.076.475.035.134.99
Monitor6.156.406.365.766.084.614.674.72
Rifle2.912.752.812.672.652.512.452.45
Sofa6.985.855.695.345.544.584.564.44
Speaker8.758.107.927.287.655.946.005.94
Table6.006.055.625.465.684.414.424.35
Telephone4.564.634.514.204.103.593.503.52
Vessel4.384.374.244.224.153.813.753.72
Mean5.625.45.274.925.074.094.094.05
Table 4. Reconstruction results on ShapeNet evaluated by EMD.
Table 4. Reconstruction results on ShapeNet evaluated by EMD.
ClassPSGN3D-LMNetSE-MD3D-VENet3D-ARNet3D-Reconstnet3D-FEGNetOurs
E M D × 10 2
Airplane6.384.774.783.563.122.802.672.72
Bench5.884.994.614.093.933.223.753.15
Cabinet6.046.356.374.694.813.844.753.91
Car4.874.104.113.573.382.873.402.84
Chair9.638.026.536.115.454.244.524.17
Lamp16.1715.8012.119.977.606.406.116.19
Monitor7.597.136.745.635.584.384.884.49
Rifle8.486.085.894.063.393.632.913.48
Sofa7.425.655.214.804.493.834.563.71
Speaker8.709.157.866.786.595.266.245.23
Table8.407.826.146.105.234.264.624.16
Telephone5.075.435.113.613.253.063.393.05
Vessel6.185.685.254.594.053.994.093.96
Mean7.757.006.215.204.683.984.303.93
Table 5. Reconstruction results of different activation functions on Pix3D dataset.
Table 5. Reconstruction results of different activation functions on Pix3D dataset.
ClassSoftsignSoftplusLeaky ReLUELUSoftsignSoftplusLeaky ReLUELU
C D × 10 2 E M D × 10 2
Chair5.705.575.585.526.326.136.086.05
Sofa6.226.266.096.045.125.244.924.90
Table7.737.327.036.888.668.207.607.59
Table 6. Reconstruction results on Pix3D evaluated by CD.
Table 6. Reconstruction results on Pix3D evaluated by CD.
ClassPSGN3D-LMNet3D-ARNet3D-ReconstNet3D-FEGNetOurs
C D × 10 2
Chair8.057.357.225.595.665.52
Sofa8.458.188.136.146.236.04
Table10.8511.210.317.047.586.88
Mean9.128.918.556.266.496.15
Table 7. Reconstruction results on Pix3D evaluated by EMD.
Table 7. Reconstruction results on Pix3D evaluated by EMD.
ClassPSGN3D-LMNet3D-ARNet3D-ReconstNet3D-FEGNetOurs
E M D × 10 2
Chair12.559.147.945.998.246.05
Sofa9.167.226.695.026.774.90
Table15.1612.7310.427.6011.407.59
Mean12.299.708.356.208.806.18
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Li, B.; Zhu, S.; Lu, Y. A Single Stage and Single View 3D Point Cloud Reconstruction Network Based on DetNet. Sensors 2022, 22, 8235. https://doi.org/10.3390/s22218235

AMA Style

Li B, Zhu S, Lu Y. A Single Stage and Single View 3D Point Cloud Reconstruction Network Based on DetNet. Sensors. 2022; 22(21):8235. https://doi.org/10.3390/s22218235

Chicago/Turabian Style

Li, Bin, Shiao Zhu, and Yi Lu. 2022. "A Single Stage and Single View 3D Point Cloud Reconstruction Network Based on DetNet" Sensors 22, no. 21: 8235. https://doi.org/10.3390/s22218235

APA Style

Li, B., Zhu, S., & Lu, Y. (2022). A Single Stage and Single View 3D Point Cloud Reconstruction Network Based on DetNet. Sensors, 22(21), 8235. https://doi.org/10.3390/s22218235

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop