Next Article in Journal
A Study of Forest Phenology Prediction Based on GRU Models
Previous Article in Journal
Battery-Free Pork Freshness Estimation Based on Colorimetric Sensors and Machine Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Efficient Bidirectional Point Pyramid Attention Network for 3D Point Cloud Completion

School of Digital Arts and Design, Dalian Neusoft University of Information, Dalian 116023, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2023, 13(8), 4897; https://doi.org/10.3390/app13084897
Submission received: 18 November 2022 / Revised: 17 December 2022 / Accepted: 11 January 2023 / Published: 13 April 2023

Abstract

:
Point cloud completion is a necessary task in real-world applications of recovering a complete geometry from missing regions of 3D objects. Furthermore, model efficiency is of vital importance in computer vision. In this paper, we present an efficient encoder–decoder network that predicts missing point clouds on the basis of incomplete point clouds. There are several advantages to this approach. First, a Mixed Attention Module (MAM) was implemented to obtain the correlational information of points. Second, the proposed Bidirectional Point Pyramid Attention Network (BiPPAN) can achieve simple and fast multiscale feature fusion to capture important features. Lastly, the designed encoder–decoder framework comprises skip connections to capture long-distance dependencies and structural information. We can conclude from the results of the experiments that the proposed network is an efficient and effective method to accomplish point cloud completion tasks.

1. Introduction

Three-dimensional vision is a major research area in computer vision. Recently, due to the rapid development of 3D sensing technology, 3D vision has been one of the research hotspots. Three-dimensional data can emerge in different forms, including depth images, point clouds, meshes, and volumetric grids. The point cloud is a common display form because it preserves the original geometric information in 3D space and requires less memory to store. However, point clouds obtained from laser scanners or other devices are usually incomplete, which brings some difficulties to the subsequent processing of point clouds. Therefore, as a point cloud preprocessing method, completing point clouds from missing and sparse raw data has become an important task [1,2].
Deep learning technology has dominated numerous research areas and shown significant advantages in image recognition [3], speech recognition [4], natural language processing [5] and other fields [6]. With the widespread application of 3D data, many available large 3D datasets have emerged, such as ModelNet [7] and ShapeNet [8]. At the same time, deep learning on 3D shape classification, 3D object detection and tracking, and 3D point cloud completion is receiving increasing attention [9,10].
As a pioneering work, PointNet [11] developed by Qi et al. directly takes point clouds as input and output that respect the permutation invariance of points. The major contribution is that PointNet uses a max pooling layer as a symmetric function to extract global features from all the points, and multi-layer perceptron (MLP) layers to learn pointwise features independently. It is able to carry out object classification, part segmentation, and scene semantic parsing. After PointNet, many other methods sprang up to generate complete point clouds [12,13]. The challenge of point cloud completion is to extract structural information from unordered and unstructured point cloud data. Obviously, learned features directly identify the quality of completion. Therefore, it is very important to effectively extract and exploit the features of point clouds to raise the accuracy of prediction. We set the goal of capturing the correlational information of points and selectively convey geometric information from the local regions of incomplete point clouds, so that we could reconstruct complete point clouds. The designed modules enable our model to learn structural features better, and preserve detailed information for point cloud completion to improve prediction accuracy.
In this work, we demonstrate a method to only generate the missing part of point clouds. Our main contributions are reflected in three aspects and are summed up in Figure 1.
(1)
We offer an encoder–decoder network for point cloud completion. A set of experiments proved the effectiveness and robustness of the proposed method on several challenging datasets.
(2)
We applied a Channelwise Attention Module (CAM) and a Mixed Attention Module (MAM) to introduce a reasonable weight to learn the importance of distinct features. While fusing different input features, the contributions of the fused output features are not equal. The attention mechanism allows for the network to infer the missing regions from incomplete point clouds, exploiting more effective geometric information.
(3)
We designed a simple and highly effective decoder that consists of skip connections and a Bidirectional Point Pyramid Attention Network (BiPPAN). BiPPAN applies top–down and bottom–up multiscale feature fusion. In fact, a multilayer pyramid network model was used on PF-Net [14]. However, a point cloud generation method propagating low-level features to improve the entire feature hierarchy has not yet been invented.
To verify whether the proposed network acted effectively, we evaluated our network with extensive experiments on public ShapeNet datasets of three different sizes (ShapeNet-13, ShapeNet-16, ShapeNet-55). We can conclude from the experimental results that our proposed method outperformed PF-Net as the baseline method, whether it was on large or small datasets.
The remainder of the paper is organized as follows. Section 2 demonstrates the related works. Section 3 presents our network and loss function. Section 4 describes our experiments on the ShapeNet dataset to perform completion in point clouds. Lastly, Section 5 summarizes our paper.

2. Related Work

2.1. 3D Shape Completion

In this section, we review recent work applying deep learning technology to shape completion, and then introduce the feature pyramid network and its application to point cloud completion.
In the field of shape completion, PCN [15] is the first deep learning network using point clouds without any voxelization. It presents an encoder–decoder architecture and FoldingNet to generate a dense point cloud in a coarse-to-fine fashion. Subsequently, many models have been developed to achieve high-resolution completion and strong generalization performance. By combining reinforcement learning with a generative adversarial network (GAN), RL-GAN-Net [16] and Render4Completion [17] are a novel set of architectures for 3D point cloud representation learning and generation whose reconstruction results had better robustness. Huang et al. [18] proposed a new recurrent forward network (RFNet) for point cloud completion and achieved the most advanced performance, which consisted of three separate modules: recurrent feature extraction (RFE), forward dense completion (FDC), and raw shape protection (RSP).

2.2. Multiscale Feature Representations

The feature pyramid network (FPN) is a top–down pathway to combine multiscale features. It can effectively represent multiscale features and is widely used in various tasks [19,20]. Huang et al. [14] designed a point fractal network (PF-Net) to hierarchically generate a missing point cloud that utilizes a multilayer point pyramid decoder (PPD) to predict the completion results of different layers.

3. Methods

In this part, we state the detailed information of our network. Figure 2 shows the full architecture of our method, which only predicts the missing parts from incomplete inputs. It consists of an encoder with Attention Combined Multi-Layer Perceptron (ACMLP) and a decoder with a Bidirectional Point Pyramid Attention Network (BiPPAN). The attention module was applied to the encoder and decoder as detailed in Section 3.1. The encoder shows its talent not only with excellent feature extraction, but also in the presentation of feature aggregation. We detail the point feature encoder in Section 3.2. Between the input and the decoder, skip connections connect the geometric information of the point cloud with the point features in the decoder. BiPPAN performs the prediction of missing point clouds with different resolutions. We describe the decoder in detail in Section 3.3. To optimize the network parameters, multistage completion loss was applied. The multistage completion loss was set as the Chamfer Distance (CD) error between the predicted point cloud and ground truth, which was composed of three items. We detail the loss function in Section 3.4.

3.1. Mixed Attention Module

Because of its ability to select features, the attention mechanism has been applied in a series of tasks, such as machine translation, object recognition, visual question answering and so on [21,22]. Recent studies showed that the attention mechanism has been widely used in 3D shape completion. Wen et al. [23] proposed a skip-attention network (SA-NET) for point cloud completion tasks, and designed a skipping-attention mechanism to effectively utilize the local structural details of a point cloud. Yu et al. [24] presented a novel model called PoinTr for point cloud completion that fully uses an encoder–decoder architecture with the adapted transformer blocks. Inspired by squeeze-and-excitation networks (SENets) [25] that adaptively recalibrate channelwise feature responses, we adopted channelwise and pointwise attention in our encoder–decoder architecture to conquer the problem of point cloud completion. The structure of the mixed attention module is illustrated in Figure 3.
Channelwise attention: Given a point feature matrix z with the size of C × L , we applied channelwise average pooling to obtain a global feature vector with the size of C × 1 . Then, we operated two fully connected (FC) layers around ReLU activation units. The layers were able to fully capture channelwise dependencies.
C A = F ( z , W ) = W 2 δ ( W 1 p o o l ( z ) )
where the p o o l denotes the global average pooling, W 1 R C / 16 × C and W 2 R C × C / 16 denote parameters in the two FC layers, and δ refers to the ReLU activation units. C A R C × 1 is the channelwise attention of z .
Pointwise attention: Similar to the strategy for predicting the channelwise attention for a point feature matrix z R C × L , a pointwise average pooling operation was performed to aggregate the pointwise feature vector with the size of 1 × L. Then, we computed P A = F ( z , W ) = W 2 δ ( W 1 p o o l ( z ) ) . The weight parameters of two fully connected layers are W 1 R L / 16 × L and W 2 R L × L / 16 . P A R 1 × L is the pointwise attention of z .
Lastly, the output feature vector z ~ of the mixed attention module was obtained by weighting the feature z with CA and PA as follows. Using the sigmoid function σ , the attention matrix was normalized to [0,1].
z ~ = σ ( C A × P A ) × z + z
Through the operations above, the feature representation z ~ could learn discrimination and robust representation for point clouds.

3.2. Point Feature Encoder

The size of the incomplete point cloud input was N = 2048 with its 3-dimensional coordinates. The encoder network was utilized to extract features from the incomplete input. Motivated by PointNet, we designed an Attention Combined Multi-Layer Perceptron (ACMLP) as our point cloud feature encoder. As shown in Figure 4, the structure of the encoder comprised a shared multi-layer perceptron (MLP) with LeakyReLU activation, which consisted of five layers with neuron sizes 64, 128, 256, 512, and 1024. Different dimensions of MLP can extract low, mid-, and high-level features, each of which contains rich point cloud information. In order to utilize these features effectively, we used max pooling to obtain the global latent representation f i of the output of the last four layers, whose size was f i : = 128 , 256 , 512 , 1024 , for i = 1,...,4, and concatenate f i to produce a 1920-dimensional feature vector. Then, we adopted a channelwise attention module for the 1920-dimensional features to obtain the weighted multilevel features f ~ that contained low-, mid-, and high-level feature information, and could help in generating effective features for point cloud prediction. Since we used max pooling for the 1920-dimensional feature vector, we did not take the pooling layer into the channelwise attention module. Eventually, we generated the final feature vector F 1 through a fully connected layer.

3.3. Point Pyramid Decoder

The decoder structure is composed of two modules: skip connections and Bidirectional Point Pyramid Attention Network (BiPPAN).
As shown in Figure 2, we concatenated the input incomplete point clouds with the F 1 F 3 via the long-range skip connections. We obtained F 2 and F 3 by passing F 1 through fully connected layers. We also employed the mixed attention module to fuse these features and generate F 1 F 3 as the input of BiPPAN. There are two benefits to such skip connections. One is to provide long-range information compensation, so the raw incomplete point cloud geometry information is still available in the decoder architecture. The other is that residual learning can facilitate gradient backpropagation.
BiPPAN is based on the feature pyramid network (FPN) that contains a top–down pathway to aggregate multiscale features from Level 1 to 3 ( F 1 = 512 × 3 , F 2 = 256 × 3 and F 3 = 128 × 3 ), as shown in Figure 5a. Levels 1–3 represent low-to-high levels. The conventional FPN fuses multiscale features in a top–down manner. Formally,
F 3 o u t = C o n v ( F 3 )
F 2 o u t = C o n v ( F 2 + R e s i z e ( F 3 o u t ) )
F 1 o u t = C o n v ( F 1 + R e s i z e ( F 2 o u t ) )
where R e s i z e is an upsampling/downsampling or reshape operation for size matching, and C o n v is a convolutional operation. F 1 o u t , F 2 o u t and F 3 o u t ( F 1 o u t = 512 × 3 , F 2 o u t = 128 × 3 and F 3 o u t = 64 × 3 ) are the predicted missing regions of the incomplete point cloud in different resolutions.
Following this idea, we designed a bidirectional point pyramid attention network to optimize multiscale feature fusion on top of FPN, as depicted in Figure 5b. First, we added not only a bottom–up path aggregation network, but also cross-level connections. By fully fusing high-level features with rich global information and low-level features with many fine details of local geometry information, we could predict the detailed structure of the missing point cloud. Second, we introduced the mixed attention module to learn the importance of different input features whose purpose is to selectively aggregate the features.

3.4. Loss Function

The loss function in our method is multistage completion loss L c o m . Because we set the pyramid level N = 3 in our network, the multistage completion loss L c o m was composed of three parts ( d C D 1 , d C D 2 , d C D 3 ):
L c o m = d C D 1 ( F 1 o u t , F g t ) + α × d C D 2 ( F 2 o u t , F g t ) + 2 α × d C D 3 ( F 3 o u t , F g t )
where α is the hyperparameter, and d C D is the Chamfer Distance proposed by Fan [26]. The mean Chamfer Distance can measure the average nearest squared distance between prediction point cloud F o u t and ground truth F g t , which is calculated with:
d C D ( F o u t , F g t ) = 1 F o u t x F o u t min y F gt x y 2 2 + 1 F g t y F g t min x F out y x 2 2
In detail, d C D 1 calculates the squared distance between predicted detailed points F 1 o u t and the ground truth of the missing region F g t . Items d C D 2 and d C D 3 calculate the squared distance among secondary central points F 2 o u t , predicted primary central points, F 3 o u t and ground truth F g t , F g t .

4. Experimental

In this part, we present the three benchmarks for point cloud completion and the implementation details of our network. Then, we show the results of our method and some other baseline methods on the benchmark. Lastly, we provide an ablation study and robustness test of our network. We evaluate how our model works both quantitatively and qualitatively.

4.1. Data Generation and Model Training

We tested our method on a commonly used benchmark for 3D point cloud completion, i.e., ShapeNet. We used 13 and 16 categories of different objects in the benchmark dataset ShapeNet-Part [27]. Since a more diverse dataset can comprehensively test the capabilities of the model, we also conducted experiments using all 55 categories in ShapeNet. Similar to the dataset of PoinTr, we randomly sampled 80% of the objects from each category of ShapeNet-55 to form the training set, and 20% for evaluation, resulting in 41,952 models for training and 10,518 models for testing. Complete point clouds were generated by uniformly sampling 2048 points from 5 randomly distributed viewpoints. During training and testing, we sampled 512 points as the ground truth F g t of the incomplete point cloud, that is, 25% of the original data were missing. We applied iterative farthest point sampling (IFPS) [28] to extract ground truths F g t and F g t with 128 and 64 points, respectively.
We implemented all of our models with the PyTorch deep learning framework, and used the Adam optimizer to update the parameters of the network during training. The learning rate was initially set to 0.0001, the model was trained for 200 epochs, and the continuous learning rate decreased by 0.2 every 40 epochs. We defined the batch size to be 24. The models were trained and tested on a NVIDIA GeForce RTX 3090 GPU with CUDA 11.7. The mixed attention module can be used flexibly in convolutional neural networks (CNNs) to capture feature correlations effectively, and the bidirectional point pyramid attention network (BiPPAN) can be used as a decoder in the network.

4.2. Completion Results on ShapeNet

We applied the mean Chamfer Distance as the evaluation metric, which contained two items in our experiments: P r e d G T and G T P r e d .
The Chamfer Distance from a predicted point cloud to the ground truth ( P r e d G T ) measures the difference between the predicted and real point clouds, which is calculated with:
d C D ( F P r e d , F G T ) = 1 F P r e d x F P r e d min y F GT x y 2 2
The Chamfer Distance from the ground truth to a predicted point cloud ( G T P r e d ) manifests the extent to which the shape of the prediction covers the real point cloud, which is calculated with:
d C D ( F G T , F P r e d ) = 1 F G T x F G T min y F Pred x y 2 2
Results on ShapeNet-13: We compared the results of our model with those of L-GAN [29], PCN [15], 3D-point Capsule Networks [30] and PF-Net [14] on standard metrics. In Table 1, the results of the above four methods are cited from [14]. Our method surpassed the other ones on 13 categories in terms of P r e d G T and G T P r e d errors, and achieved the best average Chamfer Distance in the missing region. This suggests that our network was able to reconstruct the missing point cloud with higher precision.
Results on ShapeNet-16: We experimented with our method and the most advanced PF-Net method on ShapeNet-16. Since PF-Net was only trained on the ShapeNet-13 dataset, we trained and tested it again using the same dataset according to their open-source code for fair comparison and quantitative analysis. Table 2 shows that our method resulted in a lower average Chamfer Distance error when increasing the dataset from 13 to 16 categories. The three new categories, namely, Earphone, Knife, Rocket, on both P r e d G T and G T P r e d errors were all smaller than those of PF-Net. These results justify that the proposed network could handle a larger dataset and effectively reconstruct the full shape. Figure 6 shows the qualitative results of our method and PF-Net on ShapeNet-16. Our method preserved the original input geometry while better computing and refining the missing parts of the completion for the shown examples.
Furthermore, we compared the number of trainable parameters in PF-Net and our method. As shown in Table 2, our method performed significantly better with fewer parameters than those of PF-Net.
Results on ShapeNet-55: To further investigate how PF-Net and our method performed with more dataset categories, we conducted experiments on ShapeNet-55. Due to the imbalanced number of shapes in the dataset, we classified the 55 categories into simple, moderate, and hard classes on the basis of their number. Specifically, more than 2500 training shapes were considered to be of the simple class, such as table, chair, and airplane. Fewer than 80 samples were classified as of the hard class, such as birdhouse, bag, and keyboard. Bed, camera, and rifle were classified as of the moderate class. We chose 3 categories from each class as examples to demonstrate the results. The average CD results for the three classes are shown in Table 3 with details. The category errors and average errors of our method were smaller than those of PF-Net. The results of our method on ShapeNet-55 were especially better than those on ShapeNet-13 and ShapeNet-16. Due to the imbalanced number of training samples, the average CD errors of the moderate and hard classes with insufficient samples for PF-Net were higher than those of the simple class. For our BiPPAN, there were no significant differences in the average errors of the three classes. In Figure 7, we present the qualitative results for 9 categories on Shapenet-55. As shown in the example, our method was able to complete the missing point cloud with higher accuracy and more details for the various incomplete categories.
Regardless of quantitative comparison or qualitative analysis, our method could improve the quality of the completed point cloud on a diverse dataset and achieve good results in the point cloud completion task.

4.3. Ablation Study

To test the effectiveness and necessity of the three designed modules of the Attention Mechanism (AM, including the Channelwise Attention Module of the encoder and the Mixed Attention Module of the decoder), the Bidirectional Point Pyramid Network (BiPPN), and skip connections (Skip-C), we exhaustively compared and tested our method BiPPAN with its four variants on ShapeNet-16. The results of the ablation study are shown in Table 4.
Model A is a very basic baseline that includes the standard CMLP and FPN as the encoder and decoder architectures, respectively. Models B, C, D had the same structure as that of BiPPAN (our entire network) except for the removed/replaced module. Model B was the variant where AM was removed from the BiPPAN model. Adding AM reduced P r e d G T and G T P r e d errors. This confirmed the effectiveness of AM. Model C focused on demonstrating the effectiveness of introducing BiPPN into network design. After changing the decoder from FPN into BiPPN, P r e d G T and G T P r e d errors were smaller, as expected. When removing Skip-C from the BiPPAN model, the G T P r e d error of Model D increased markedly. This clearly shows that Skip-C alone can bring significant performance improvement. Notably, our BiPPAN method performed better than its variants. A visual comparison is presented in Figure 8. For the shown examples, Baseline Model A and Models B, C, D suffered from some obvious fairness coarsening and exhibited impotence for most of the details. This further confirmed the effectiveness of the BiPPAN design.

4.4. Robustness Test

In order to analyze the robustness of our method, we evaluated the point cloud completion of missing regions in different degrees on ShapeNet-16. We varied the number of incomplete point clouds from 512 to 1024 and 1536. In other words, we extended the missing part from 25% to 50% and 75% of the original shape. Table 5 and Figure 9 show the experimental results of the network on the Airplane and Car categories. Table 5 shows that the errors between the predicted point cloud and the ground truth did not differ much when 25%, 50%, and 75% of the points were missing. Figure 9 displays that, although the partial data were quite tough, our predictions were sensible. In the case of the 75% missing area, the complete car shape could be defined and completed only on the basis of the front of the car. This experiment verified the strong completion robustness of our method when dealing with input point clouds with distinct missing degrees.

5. Conclusions

In this paper, we recommended a novel Bidirectional Point Pyramid Attention Network framework (BiPPAN) for point cloud completion. Our method produces missing point clouds via an encoder–decoder structure. The attention module was introduced into the encoder and decoder to raise the correlational information of point clouds, which extracts the local details of object features better [31]. The point pyramid attention decoder was designed not only to fully use hierarchical features from the bidirectional connection, but also to capture long-distance structural information by employing skip connections as well [32]. Experimental results on ShapeNet-13, ShapeNet-16, and ShapeNet-55 demonstrate that, no matter the size of the dataset, our proposed method was effective and efficient in comparison with other methods. An ablation and robustness test validated our claim that the attention module, bidirectional connection, and skip connections could more effectively extract and exploit features of point clouds to improve prediction accuracy. Our method tended to reduce the error in challenging classes.
Hence, this method can be widely applied to the task of generating target point clouds with both rich semantic profiles and detailed features while preserving existing contours. The good results of BiPPAN on many datasets demonstrate its ability to repair and complete 3D shapes, and improve the accuracy of downstream tasks such as 3D recognition, and 3D object detection and tracking. As a point cloud preprocessing method, our method has tremendous potential in autonomous vehicles, 3D reconstruction, and remote sensing research.

Author Contributions

Methodology and writing—original draft preparation, Y.L.; data curation and visualization, Y.X.; investigation and supervision, J.G.; conceptualization and project administration, Q.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Young Science and Technology Star Project of Dalian, China [grant number 2022RQ092], and the Technology Innovation Fund Project of Dalian Neusoft University of Information [grant number TIFP202303].

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Publicly available datasets were analyzed in this study. The datasets can be found here: https://www.shapenet.org/ (accessed on 17 November 2022).

Acknowledgments

The authors acknowledge all editors and reviewers for their suggestions. We would also like to thank the Dalian Ascend AI Computing Center and Dalian Ascend AI Ecosystem Innovation Center for providing computing power and technical support.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wen, X.; Han, Z.; Cao, Y.-P.; Wan, P.; Zheng, W.; Liu, Y. Cycle4completion: Unpaired point cloud completion using cycle transformation with missing region coding. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 13080–13089. [Google Scholar]
  2. Sun, Y.; Wang, Y.; Liu, Z.; Siegel, J.E.; Sarma, S.E. PointGrow: Autoregressively Learned Point Cloud Generation with Self-Attention. In Proceedings of the IEEE Winter Conference on Applications of Computer Vision, Snowmass Village, CO, USA, 1–5 March 2020. [Google Scholar]
  3. Chen, C.; Li, O.; Tao, D.; Barnett, A.; Rudin, C.; Su, J.K. This looks like that: Deep learning for interpretable image recognition. In Advances in Neural Information Processing Systems; NeurIPS: San Diego, CA, USA, 2019; Volume 32. [Google Scholar]
  4. Santhanavijayan, A.; Kumar, D.N.; Deepak, G. A semantic-aware strategy for automatic speech recognition incorporating deep learning models. In Intelligent System Design; Springer: Singapore, 2021; pp. 247–254. [Google Scholar]
  5. Pandey, B.; Pandey, D.K.; Mishra, B.P.; Rhmann, W. A comprehensive survey of deep learning in the field of medical imaging and medical natural language processing: Challenges and research directions. J. King Saud Univ. Comput. Inf. Sci. 2021, 34, 5083–5099. [Google Scholar] [CrossRef]
  6. Zhang, W.; Li, H.; Li, Y.; Liu, H.; Chen, Y.; Ding, X. Application of deep learning algorithms in geotechnical engineering: A short critical review. Artif. Intell. Rev. 2021, 54, 5633–5673. [Google Scholar] [CrossRef]
  7. Wu, Z.; Song, S.; Khosla, A.; Yu, F.; Zhang, L.; Tang, X.; Xiao, J. 3d shapenets: A deep representation for volumetric shapes. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1912–1920. [Google Scholar]
  8. Chang, A.X.; Funkhouser, T.; Guibas, L.; Hanrahan, P.; Huang, Q.; Li, Z.; Savarese, S.; Savva, M.; Song, S.; Su, H.; et al. Shapenet: An information-rich 3d model repository. Comput. Sci. 2015. [Google Scholar] [CrossRef]
  9. Cui, Y.; Chen, R.; Chu, W.; Chen, L.; Tian, D.; Li, Y.; Cao, D. Deep learning for image and point cloud fusion in autonomous driving: A review. IEEE Trans. Intell. Transp. Syst. 2021, 23, 722–739. [Google Scholar] [CrossRef]
  10. Guo, Y.; Wang, H.; Hu, Q.; Liu, H.; Liu, L.; Bennamoun, M. Deep learning for 3d point clouds: A survey. In IEEE Transactions on Pattern Analysis and Machine Intelligence; IEEE: Piscataway Township, NJ, USA, 2020; Volume 43, pp. 4338–4364. [Google Scholar]
  11. Qi, C.R.; Su, H.; Mo, K.; Guibas, L.J. PointNet: Deep Learning on Point Sets for 3D Classification and Segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  12. Alliegro, A.; Valsesia, D.; Fracastoro, G.; Magli, E.; Tommasi, T. Denoise and contrast for category agnostic shape completion. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 4629–4638. [Google Scholar]
  13. Xiang, P.; Wen, X.; Liu, Y.-S.; Cao, Y.-P.; Wan, P.; Zheng, W.; Han, Z. Snowflakenet: Point cloud completion by snowflake point deconvolution with skip-transformer. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 5499–5509. [Google Scholar]
  14. Huang, Z.; Yu, Y.; Xu, J.; Ni, F.; Le, X. PF-Net: Point Fractal Network for 3D Point Cloud Completion. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020. [Google Scholar]
  15. Yuan, W.; Khot, T.; Held, D.; Mertz, C.; Hebert, M. PCN: Point Completion Network. In Proceedings of the International Conference on 3D Vision, Verona, Italy, 5–8 September 2018. [Google Scholar]
  16. Sarmad, M.; Lee, H.J.; Kim, Y.M. RL-GAN-Net: A Reinforcement Learning Agent Controlled GAN Network for Real-Time Point Cloud Shape Completion. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
  17. Hu, T.; Han, Z.; Shrivastava, A.; Zwicker, M. Render4Completion: Synthesizing Multi-View Depth Maps for 3D Shape Completion. In Proceedings of the IEEE/CVF International Conference on Computer Vision Workshop, Seoul, Korea, 27–28 October 2019. [Google Scholar]
  18. Huang, T.; Zou, H.; Cui, J.; Yang, X.; Wang, M.; Zhao, X.; Zhang, J.; Yuan, Y.; Xu, Y.; Liu, Y. Rfnet: Recurrent forward network for dense point cloud completion. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 12508–12517. [Google Scholar]
  19. Tan, M.; Pang, R.; Le, Q.V. EfficientDet: Scalable and Efficient Object Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020. [Google Scholar]
  20. Liu, S.; Qi, L.; Qin, H.; Shi, J.; Jia, J. Path Aggregation Network for Instance Segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  21. Miculicich, L.; Ram, D.; Pappas, N.; Henderson, J. Document-level neural machine translation with hierarchical attention networks. In Empirical Methods in Natural Language Processing; Association for Computational Linguistics: Brussels, Belgium, 2018. [Google Scholar]
  22. Guo, M.-H.; Xu, T.-X.; Liu, J.-J.; Liu, Z.-N.; Jiang, P.-T.; Mu, T.-J.; Zhang, S.H.; Martin, R.R.; Cheng, M.M.; Hu, S.M. Attention mechanisms in computer vision: A survey. Comput. Vis. Media 2022, 8, 331–368. [Google Scholar] [CrossRef]
  23. Wen, X.; Li, T.; Han, Z.; Liu, Y.S. Point Cloud Completion by Skip-attention Network with Hierarchical Folding. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020. [Google Scholar]
  24. Yu, X.; Rao, Y.; Wang, Z.; Liu, Z.; Lu, J.; Zhou, J. Pointr: Diverse point cloud completion with geometry-aware transformers. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Montreal, QC, Canada, 10–17 October 2021; pp. 12498–12507. [Google Scholar]
  25. Hu, J.; Shen, L.; Albanie, S.; Sun, G.; Wu, E. Squeeze-and-Excitation Networks. In IEEE Transactions on Pattern Analysis and Machine Intelligence; IEEE: Piscataway Township, NJ, USA, 2018. [Google Scholar]
  26. Fan, H.; Su, H.; Guibas, L. A Point Set Generation Network for 3D Object Reconstruction from a Single Image. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  27. Yi, L.; Kim, V.G.; Ceylan, D.; Shen, I.; Yan, M.; Su, H.; Lu, C.; Huang, Q.; Sheffer, A.; Guibas, L. A scalable active framework for region annotation in 3D shape collections. ACM Trans. Graph. 2016, 35, 1–12. [Google Scholar] [CrossRef]
  28. Qi, C.R.; Yi, L.; Su, H.; Guibas, L.J. PointNet++: Deep Hierarchical Feature Learning on Point Sets in a Metric Space. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  29. Achlioptas, P.; Diamanti, O.; Mitliagkas, I.; Guibas, L. Learning Representations and Generative Models for 3D Point Clouds. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  30. Zhao, Y.; Birdal, T.; Deng, H.; Tombari, F. 3D Point Capsule Networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
  31. Wang, J.; Cui, Y.; Guo, D.; Li, J.; Liu, Q.; Shen, C. Pointattn: You only need attention for point cloud completion. arXiv 2022, arXiv:2203.08485. [Google Scholar]
  32. Bai, Y.; Wang, X.; Ang, M.H., Jr.; Rus, D. BIMS-PU: Bi-Directional and Multi-Scale Point Cloud Upsampling. In IEEE Robotics and Automation Letters; IEEE: Piscataway Township, NJ, USA, 2022; pp. 7447–7454. [Google Scholar]
Figure 1. Our method directly predicts missing point clouds (blue points) with incomplete point clouds (gray points) as input. We designed an encoder–decoder architecture with a Channelwise Attention Module (CAM) and a Mixed Attention Module (MAM). The decoder utilizes a bidirectional point pyramid attention network and skip connections to generate multiscale predictions (primary center points, secondary center points, detailed points).
Figure 1. Our method directly predicts missing point clouds (blue points) with incomplete point clouds (gray points) as input. We designed an encoder–decoder architecture with a Channelwise Attention Module (CAM) and a Mixed Attention Module (MAM). The decoder utilizes a bidirectional point pyramid attention network and skip connections to generate multiscale predictions (primary center points, secondary center points, detailed points).
Applsci 13 04897 g001
Figure 2. Illustration of our framework. The encoder (yellow) adopts ACMLP to extract features from input point clouds. The decoder (green) adopts BiPPAN with skip connections to predict the multiscale missing parts. Then, the Chamfer Distance errors of the prediction point cloud and ground truth are calculated.
Figure 2. Illustration of our framework. The encoder (yellow) adopts ACMLP to extract features from input point clouds. The decoder (green) adopts BiPPAN with skip connections to predict the multiscale missing parts. Then, the Chamfer Distance errors of the prediction point cloud and ground truth are calculated.
Applsci 13 04897 g002
Figure 3. (a) Diagram and (b) schema of the mixed attention module.
Figure 3. (a) Diagram and (b) schema of the mixed attention module.
Applsci 13 04897 g003
Figure 4. Structure of the proposed ACMLP.
Figure 4. Structure of the proposed ACMLP.
Applsci 13 04897 g004
Figure 5. Comparison between (a) FPN and (b) the proposed BiPPAN decoder.
Figure 5. Comparison between (a) FPN and (b) the proposed BiPPAN decoder.
Applsci 13 04897 g005
Figure 6. Qualitative results on ShapeNet-16 showing the input point cloud (Input) and the ground truth (G.T.), and the predictions of PF-Net and our model.
Figure 6. Qualitative results on ShapeNet-16 showing the input point cloud (Input) and the ground truth (G.T.), and the predictions of PF-Net and our model.
Applsci 13 04897 g006
Figure 7. Qualitative results on ShapeNet-55. Point cloud completion results of our method and PF-Net on some objects from the simple, moderate, and hard classes.
Figure 7. Qualitative results on ShapeNet-55. Point cloud completion results of our method and PF-Net on some objects from the simple, moderate, and hard classes.
Applsci 13 04897 g007
Figure 8. Comparison of point cloud completion results of the proposed network, baseline Model A, and Models B, C, D.
Figure 8. Comparison of point cloud completion results of the proposed network, baseline Model A, and Models B, C, D.
Applsci 13 04897 g008
Figure 9. Examples of point cloud completion results when the input point cloud misses (a) 25%, (b) 50%, and (c) 75% of the original point cloud. Gray and blue denote the input incomplete point cloud and predicted missing point cloud, respectively. Yellow represents the real point cloud.
Figure 9. Examples of point cloud completion results when the input point cloud misses (a) 25%, (b) 50%, and (c) 75% of the original point cloud. Gray and blue denote the input incomplete point cloud and predicted missing point cloud, respectively. Yellow represents the real point cloud.
Applsci 13 04897 g009
Table 1. Results of a comparison between our method and most advanced methods on the ShapeNet-13 dataset. The numbers in pairs are Pred → GT and GT → Pred Chamfer Distance × 103 (lower is better).
Table 1. Results of a comparison between our method and most advanced methods on the ShapeNet-13 dataset. The numbers in pairs are Pred → GT and GT → Pred Chamfer Distance × 103 (lower is better).
CategoryLGAN-AEPCN3D CapsulePF-NetBiPPAN
Airplane3.357/1.1305.060/1.2432.676/1.4011.091/1.0700.964/0.875
Bag5.707/5.3033.251/4.3145.228/4.2023.929/3.7682.864/2.823
Cap8.968/4.6087.015/4.24011.04/4.7395.290/4.8004.371/3.489
Car4.531/2.5182.741/2.1235.944/3.5082.489/1.8392.28/1.681
Chair7.359/2.3393.952/2.3013.049/2.2072.074/1.8241.758/1.442
Guitar0.838/0.5361.419/0.6890.625/0.6620.456/0.4290.377/0.376
Lamp8.464/3.62711.61/7.1399.912/5.8475.122/3.4603.731/2.513
Laptop7.649/1.4133.070/1.4222.129/1.7331.247/0.9971.118/0.864
Motorbike4.914/2.0364.962/1.9228.617/2.7082.206/1.7751.931/1.562
Mug6.139/4.7353.590/3.5915.155/5.1683.138/3.2383.036/2.914
Pistol3.944/1.4244.484/1.4145.980/1.7821.122/1.0550.929/0.831
Skateboard5.613/1.6833.025/1.74011.49/2.0441.136/1.3371.007/1.029
Table2.658/2.4842.503/2.4523.929/3.0982.235/1.9341.744/1.626
Mean5.395/2.6034.360/2.6615.829/3.0082.426/2.1172.008/1.694
Table 2. Results of a comparison between our method and PF-Net on ShapeNet-16 dataset. We also provide the number of parameter (Params) in the last rows.
Table 2. Results of a comparison between our method and PF-Net on ShapeNet-16 dataset. We also provide the number of parameter (Params) in the last rows.
CategoryPF-NetBiPPAN
Pred → GTGT → PredPred → GTGT → Pred
Airplane1.0841.1190.9630.872
Bag3.9794.6682.832.624
Cap5.2544.8974.0043.614
Car2.5481.9142.3161.634
Chair2.1542.0191.7381.428
Earphone6.0038.0583.1943.690
Guitar0.4640.5460.3770.382
Knife0.5550.5630.4640.465
Lamp4.9433.8833.8632.557
Laptop1.3091.0721.1350.888
Motorbike2.3281.8361.9441.581
Mug3.0803.5803.1472.893
Pistol1.2841.0530.9200.811
Rocket1.0520.7620.8530.620
Skateboard1.1961.3621.0251.004
Table2.3052.1231.7601.629
Mean2.4712.4661.9081.668
Params (×106)76.57111.763
Table 3. Results of a comparison between our method and PF-Net on the ShapeNet-55 dataset. The numbers in pairs are Pred → GT and GT → Pred Chamfer Distance × 103 (lower is better). CD-S, CD-M, and CD-H were used to represent CD errors in the simple, moderate, and hard classes.
Table 3. Results of a comparison between our method and PF-Net on the ShapeNet-55 dataset. The numbers in pairs are Pred → GT and GT → Pred Chamfer Distance × 103 (lower is better). CD-S, CD-M, and CD-H were used to represent CD errors in the simple, moderate, and hard classes.
SimpleModerateHardCD-SCD-MCD-HCD-
Ave
TableChairAirplaneBedCameraRifleBirdhouseBagKeyboard
PF-Net1.963
/1.882
2.101
/1.975
1.162
/1.414
3.327
/4.742
4.827
/5.613
1.235
/1.029
3.689
/3.650
3.896
/3.107
1.911
/1.006
1.845
/1.737
2.464
/2.142
2.744
/2.362
2.463
/2.150
BiPPAN1.386
/1.156
1.591
/1.177
0.884
/0.874
2.741
/1.987
3.258
/2.321
0.541
/0.488
2.015
/2.770
1.889
/1.289
0.766
/0.594
1.449
/1.134
1.801
/1.301
1.647
/1.245
1.738
/1.274
Table 4. Ablation study on different components of our proposed encoder–decoder network framework, including Attention Module (AM), Bidirectional Point Pyramid Network (BiPPN), and skip connections (Skip-C).
Table 4. Ablation study on different components of our proposed encoder–decoder network framework, including Attention Module (AM), Bidirectional Point Pyramid Network (BiPPN), and skip connections (Skip-C).
ModelAMBiPPNSkip-CPred → GT (×103)GT → Pred (×103)
A 2.4982.300
B 2.1251.869
C 2.2421.837
D 2.0042.011
BiPPAN1.9081.668
Table 5. Robustness results of different extent of incomplete input point cloud using our method.
Table 5. Robustness results of different extent of incomplete input point cloud using our method.
Missing Ratio25%50%75%
CD (×103)Pred → GTGT → PredPred → GTGT → PredPred → GTGT → Pred
Airplane0.9630.8720.9710.8090.9680.896
Car2.3161.6342.4941.7862.4791.841
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, Y.; Xiao, Y.; Gang, J.; Yu, Q. An Efficient Bidirectional Point Pyramid Attention Network for 3D Point Cloud Completion. Appl. Sci. 2023, 13, 4897. https://doi.org/10.3390/app13084897

AMA Style

Li Y, Xiao Y, Gang J, Yu Q. An Efficient Bidirectional Point Pyramid Attention Network for 3D Point Cloud Completion. Applied Sciences. 2023; 13(8):4897. https://doi.org/10.3390/app13084897

Chicago/Turabian Style

Li, Yang, Yao Xiao, Jialin Gang, and Qingjun Yu. 2023. "An Efficient Bidirectional Point Pyramid Attention Network for 3D Point Cloud Completion" Applied Sciences 13, no. 8: 4897. https://doi.org/10.3390/app13084897

APA Style

Li, Y., Xiao, Y., Gang, J., & Yu, Q. (2023). An Efficient Bidirectional Point Pyramid Attention Network for 3D Point Cloud Completion. Applied Sciences, 13(8), 4897. https://doi.org/10.3390/app13084897

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop