Next Article in Journal
A Multi-Strategy Improved Sparrow Search Algorithm for Coverage Optimization in a WSN
Previous Article in Journal
A Trade-Off between Complexity and Interaction Quality for Upper Limb Exoskeleton Interfaces
Previous Article in Special Issue
Use-Case-Oriented Evaluation of Wireless Communication Technologies for Advanced Underground Mining Operations
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Learning Representative Features by Deep Attention Network for 3D Point Cloud Registration

1
Beijing Institute of System Engineering, Beijing 100101, China
2
Artificial Intelligence Institute of China Electronics Technology Group Corporation, Beijing 100041, China
3
State Key Laboratory of Software Development Environment, Beihang University, Beijing 100191, China
4
School of Computer Science and Engineering, University of New South Wales, Sydney, NSW 2052, Australia
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(8), 4123; https://doi.org/10.3390/s23084123
Submission received: 5 March 2023 / Revised: 4 April 2023 / Accepted: 13 April 2023 / Published: 20 April 2023
(This article belongs to the Special Issue Sensors, Robotics and Networks in Mining)

Abstract

:
Three-dimensional point cloud registration, which aims to find the transformation that best aligns two point clouds, is a widely studied problem in computer vision with a wide spectrum of applications, such as underground mining. Many learning-based approaches have been developed and have demonstrated their effectiveness for point cloud registration. Particularly, attention-based models have achieved outstanding performance due to the extra contextual information captured by attention mechanisms. To avoid the high computation cost brought by attention mechanisms, an encoder–decoder framework is often employed to hierarchically extract the features where the attention module is only applied in the middle. This leads to the compromised effectiveness of the attention module. To tackle this issue, we propose a novel model with the attention layers embedded in both the encoder and decoder stages. In our model, the self-attentional layers are applied in the encoder to consider the relationship between points inside each point cloud, while the decoder utilizes cross-attentional layers to enrich features with contextual information. Extensive experiments conducted on public datasets prove that our model is able to achieve quality results on a registration task.

1. Introduction

A point cloud is a widely used data structure for 3D object description due to its precision and easy accessibility. Its applications range from autonomous driving to underground mining [1,2,3,4]. Figure 1 depicts a practical example, where points cloud data are adopted in the detection of bolts in the mining industry [1]. Point cloud registration is an important task over point cloud data and aims to locate the optimal transformation that aligns two different point cloud datasets. It is fundamental to many downstream applications, including simultaneous localization and mapping (SLAM), robotics, and 3D reconstruction. For instance, registration can be a useful tool in aligning street views during the construction of smart cities [5]. Due to the various practical applications, the exploration of point cloud registration holds great importance. Traditional point cloud registration methods rely heavily on initialization pose and struggle to handle complex cases with partial or noisy point clouds. Recent advances in deep learning have led to a growing number of learning-based approaches applied to registration tasks.
Existing learning-based techniques for point cloud registration generally employ correspondence-based methods [6,7,8,9,10], where typically three main steps are involved. First, a network is developed to extract features on a pointwise basis. Next, correspondences between two point clouds are identified by assessing feature similarity. Last, a transformation matrix is produced using the acquired correspondences. Of the three steps, the feature extraction phase is especially vital in obtaining high-quality results. In the earlier stage for investigating the point cloud registration problem, handcrafted features such as FPFH [11] and SHOT [12] are commonly utilized to derive pointwise features. On the other hand, more recent studies have focused on deep-learning-based approaches. Most methods aim to learn the spatial encoding of points or the geometric structure of local areas [13,14,15]. While these features have proved useful in identifying the correspondences, they struggle to distinguish points in nonsalient regions, where the point features are not sufficiently discriminative. Consequently, performances are less satisfactory without the availability of shared contextual information across point clouds.
Recently, the attention mechanism has gained popularity as a tool for feature matching. The cross- and self-attention mode, which was proposed by [16], has been shown to be effective in identifying two-view correspondences. This technique ensembles the intuition that when humans attempt to match two images, they frequently compare details by looking back and forth. Similarly, the attention mechanism enables the exchange of information between two items, thereby improving the accuracy of potential matching. The attention approach has been incorporated into numerous computer vision tasks, including registration. However, applying this module directly to raw point clouds can result in computational issues due to the necessity of stacking multiple attentional layers to learn representative features with sufficient contextual information. Unlike 2D images, which have a limited number of pixels, point clouds often contain a large number of points, making it prohibitively expensive to handle the linear computation brought by the attentional layer. In light of this, recent studies attempt to use specific sampling strategies to initially reduce the number of points [17,18]. However, the use of a sampling strategy can significantly impact feature quality since a downsampled point cloud cannot well preserve the original geometric structure. Since most feature extraction models depend on the structure or detail of a local patch to describe a point, point reduction can distort the shape and lose useful information. Additionally, the choice of sampling strategy presents another challenge for the registration task. Therefore, directly downsampling points may not be the optimal solution for addressing the expensive computation costs in attention approaches for the point cloud registration task.
The aforementioned issues have motivated studies to integrate the attention module with a hierarchical feature extraction model to alleviate the high computation costs [10,19,20]. In contrast to prior methods, which compute features directly on the downsampled point cloud, this approach allows each layer to gradually sparsify the points and deepen features based on the previous layer to ensure the preservation of information. However, the efficacy of the attention module may be restricted, as it is usually applied at the last subsampled layer, where only the sparse points at the highest level are utilized in detecting potential matches. This may result in a situation where features propagated to lower levels cannot benefit fully from the advantages of the attention mechanism. Therefore, it is critical to maximize the value of the attention module without incurring expensive computation costs for 3D point clouds.
To address this issue, we propose a model called DAN based on an encoder–decoder framework, which fully capitalizes on the advantages of a multplex graph attention module at a reasonable computational cost and hierarchically derives keypoints and associated features. Instead of solely employing the attention module at the last subsampled points, our method embeds self-attentional layers in the encoder and cross-attentional layers in the decoder. As a result, the encoder is responsible for extracting individual features within the point cloud, while the decoder emphasizes contextual information across point clouds, which can be beneficial in finding potential matches. This approach avoids the problem of contextual information derived from the abstract level not being propagated in the right direction. Furthermore, with the assistance of graph attentional layers, the topological structure of the point cloud can be thoroughly explored, enhancing feature robustness.

2. Related Work

In this section, we mainly review the deep learning registration models that adopt correspondence-based methods.

2.1. Feature Extraction

Correspondence-based methods usually involve the extraction of pointwise features and then establish correspondences based on feature similarity. Thus, the goal is to find representative features that can identify correspondences from points that share similar characteristics. A revolutionary work, PointNet [13], is proposed to directly process unordered point sets, which is comprised of several shared multilayer perceptron (MLP) layers. It computes the spatial encoding of individual points, which makes it highly efficient. However, it does not take the local structure of point sets into account. To make up such deficiency, PointNet++ [14] is designed to capture the hierarchical structure. The set abstraction operation in PointNet++ provides a systematic way to summarize the local pattern around keypoints by grouping the features of points from the neighborhood.
Inspired by the success of convolutional neural networks (CNNs) in 2D computer vision tasks, many models utilize the idea of CNNs in 3D point cloud feature extraction. Different from rigid formats such as 2D pixels, point clouds are unordered and uneven. Because of these properties, directly adopting the concept of CNNs on 3D point clouds is challenging. One of the solution is to construct a 3D voxel analogous to the grid pattern of 2D pixels. However, the additional dimensionality could induce huge computation cost, which is unfeasible to process large-scale data. Thus, some studies seek alternative ways to define the notion of convolution for 3D point clouds. KPConv [21] is another widely used model for point-cloud-related tasks. It introduces the concept of kernel points, which mimics the functionality of kernel pixels in an image. Ref. [22] devised Edge-Conv, which utilizes graph structure to present point clouds. It gives the flexibility to dynamically capture the topological structure in each layer. Generally, these works can be used in any point-cloud-related task, such as segmentation and classification. However, the discrepancies between registration and other tasks result in different demands on feature extraction.
As the registration task involves the matching of points between two or more point clouds, extracting discriminative features that can help to identify correspondences is usually the main focus. Based on these factors, many works are proposed specifically to address the registration problem. Ref. [15] leveraged a 3D convolutional network to learn the representation of a point from its local region features. The 3D convolutional layer is constructed by the TDF voxel grid around a point to capture the local structure. FCGF [9] constructed a fully convolutional network, where the generalized sparse convolutional layer is defined by [23] along with sparse tensors. It is capable of processing high-dimensional data effectively, since it can reuse intermediate activations on overlapping areas. Ref. [6] proposes a method to utilize joint learning on point clouds. The model produces a descriptor and a detection score for each point. The detection score is capable of finding the keypoints that are more discriminative, which makes it easier to identify correspondences. In contrast with previous methods which are more generalized, these methods often produces more distinctive features that can aid in the search for correspondences. They often use contrastive loss to train pointwise features, as it can guide the model to minimize the similarity between correspondences and maximize the similarity between points that cannot be matched together.

2.2. Attention Mechanism

Inspired by the success of transformers in NLP, many studies began to adopt attention mechanisms in computer vision tasks, including point cloud registration. The usage of attention mechanisms can help to enrich the features from a different perspective. As the first work that embeds a transformer module into the model, DCP [24] provided the functionality of contextual aggregation that considers the relations between two input point clouds, instead of embedding them independently. Another model, ref. [25], proposes a method that utilizes the transformer in a different way. It firstly constructs a graph from the point cloud by seeing points as nodes. Then, it employs the transformer into the edge generator to explore the relationship between points with the aid of contextual information.
To thoroughly utilize the advantage of attention mechanisms in computer vision tasks, SuperGlue [16] developed a multiplex graph neural network to solve the two-view correspondences problem, where the keypoints are defined as nodes and the connections are defined as edges. The message passing is carried out via alternating cross- and self-attention. As SuperGlue has achieved remarkable results, some studies attempt to adapt this module onto point cloud registration tasks. Since directly using attention mechanisms on raw point clouds tends to be infeasible for large-scale datasets due to the computation cost, refs. [17,18] seeks different sampling strategies to mitigate this issue. Fischer, K. et al. [17] introduced a pillar layer, which encodes a set of dense points by selecting a sparse subset of keypoints. Shi, C. et al. [18] make use of an existing pretrained model [26] to detect keypoints first.
Another form of using attention modules without a reduction in points was firstly proposed by [10]. It devises a method that embeds the attention module between encoder and decoder. It provides the functionality of detecting overlapping points and their matchability. Following this work, ref. [20] proposed a model that predicts rigid transformation via a closed-form solution where the correspondence loss is able to be taken into account, as it provides a differentiable formula to estimate the distance between the transformed point and the target point. These methods use an attention module on superpoints which has been downsampled and then propagate to original dense points. However, the computed features of superpoints might not be able to contain enough geometric information. To this end, ref. [19] aims to devise a model that improves the accuracy of matchings between superpoints by considering the geometric structure, rather than only making use of abstract embeddings.
Attention modules have been proved to be a useful tool in registration tasks, as they have the capability of obtaining contextual information which other methods are not able to provide. However, for the aforementioned methods, the result relies on the superpoint operations, where the power of the attention mechanism might be diminished during propagation.

3. Methods

Given two point clouds P = { p i R 3 | i = 1 , , N } and Q = { q j R 3 | j = 1 , , M } , the registration task aims to find the optimal transformation T that could align point cloud P and Q. T can be represented by
T = R t 0 1 ,
where R S O ( 3 ) and t R 3 . For correspondence-based methods, the goal is to solve the following:
min R , t ( p i , q j ) C | | R p i + t q j | | ,
where C is a set of true correspondences. Our model aims at identifying the correspondences based on the similarity of feature points.
Inspired by [21], our model adopts an encoder–decoder framework which extracts points and associated features in a hierarchical manner. In this section, we first describe the basic operations within the encoder and decoder in our model. Next, we briefly explain the attentional layers and how they are embedded in the model.

3.1. Feature Extraction

As shown in Figure 2, the encoder is established by alternating KPConv layer and the self-attention layer. Through this approach, we can retrieve the basic multilevel features. The embeddings of point cloud P at the l t h level, denoted by E l p , is derived by
E l p = C o n v ( E l 1 p , P l 1 ) ,
E l p = S e l f A t t n ( E l p ) .
where E l p is the intermediate results of the embeddings; P l 1 is the subsampled point cloud at level l 1 ; C o n v ( · ) and S e l f A t t n ( · ) represent the KPConv layer and self-attention layer, respectively.
For the KPConv layer, following [21], the convolution operation is defined on a set of kernel points. We denote a point as x. The point convolution F with the kernel g on the point x can be presented by
( F g ) ( x ) = x i N g ( x i x ) f i ,
where N = { | | x i x | | r } is a set of neighbors of x that lies in the radius r. Say we introduce a set of kernel points { x ˜ k | k < K } and their weight parameter W k that maps input dimension to output dimension. For y i = x i x , the kernel function can be further defined as
g ( y i ) = k < K h ( y i , x ˜ k ) W k .
h ( · ) is the function that describes the correlation between y i and x ˜ k as following:
h ( y i , x ˜ k ) = max ( 0 , 1 | | y i x ˜ k | | σ ) .
In the decoder stage, the model aims to enrich the feature by contextual information. At each level, it adopts the cross-attentional layer between two point clouds and then propagates the features back to the lower level. During the propagation, we apply skip connection from embeddings in the encoder stage at the corresponding level. We denote the features of point cloud P at level l as F l P . The computation of features is as following:
F l P , F l Q = C r o s s A t t n ( F l + 1 P , F l + 1 Q ) ,
F l P = U p s a m p l e ( F l P , E l P , P l ) ,
F l Q = U p s a m p l e ( F l Q , E l Q , Q l ) .
where F l P , F l Q are the intermediate results; C r o s s A t t n ( · ) and U p s a m p l e ( · ) are the cross-attention layer and propagation layer, respectively.

3.2. Attentional Layer

The main process of multiplex graph attention network is shown in Figure 2. Firstly, the input would be processed by an MLP layer which maps the features into graph embeddings. We denote the embeddigs of point cloud P and Q at layer l as F l p and F l q . In the network, the embeddings are updated as following:
F l + 1 = F l + M L P ( F l , m l ) .
Next, we introduce the derivation of message m k by attentional aggregation. For simplicity, we omit the label for layers in the following. Firstly, the query, key, and value are retrieved via linear projection:
q = W q X + b q ,
k v = W k W v Y + b k b v ,
where W and b are the parameter matrix and bias term at l t h layer; X and Y denote the query point cloud and the source point cloud. To support the multiplex graph attentional module, two modes are available here. The first is self-attention, where X and Y represent the embedding of the same point cloud. This allows the model to find the internal correlation between points. The other mode is cross-attention, where X and Y are different point clouds. Through the cross-attention mode, it is capable of detecting the relationship among the points across two point clouds and possibly searching for potential matchings. After the computation of query, key, and value, the scaled dot-product attention value between point i and j can be computed by
a i j = s o f t m a x ( q i k j T d k ) ,
where d k is the dimension of queries and keys. Finally, the message of a point i can be derived:
m i = j a i j v j .
In addition to the intermediate attention module, we apply the self-attention layer in the encoder and the cross-attention layer in the decoder.
Following [10], two extra dimensions are computed by projection on derived features. The first one is the overlap score O, which is to predict whether a point lies in the overlapping region. The other one, saliency score S, evaluates the probability of finding a matching point. They are computed by the following:
O = M L P ( F )
S = s o f t m a x ( F P , F Q t ) ,
where t is the weight to scale the effect of softmax.

3.3. Loss Function

In this study, we adopt feature loss, overlap loss, and saliency loss to supervise the learning of the model.

3.3.1. Feature Loss

Our model utilizes similarity optimization to learn distinctive features by minimizing the similarity between points that are close to each other and maximizing the similarity between points that are far away from each other. Circle loss [27] is adopted here:
L f e a t = log [ 1 + j = 1 L e γ α n j ( s n j Δ n ) · i = 1 k e γ α p i ( s p i Δ p ) ]
where s p is the similarity between features of close points and s n is for nonclose points; Δ n and Δ p are the margins for correspondences and noncorrespondences; γ is the scale factor; and weights are defined as α n j = s n j Δ n and α p i = s p i Δ p . In contrast with other similarity losses, circle loss can be more flexible, as it gives different penalty strengths for between-class similarity and within-class similarity.

3.3.2. Overlap Loss

Our model also computes the overlap loss to learn the probability of a point lying in the overlapping area. The overlap loss function is defined as
L o v e r l a p = 1 N i = 1 N o i log o ^ i + ( 1 o i ) log ( 1 o ^ i )
where i is the index of the correspondence; o i is the ground truth label that is computed by
o i = 1 if | | T P Q ( p i ) q j | | 2 r o , 0 otherwise
where q j is the nearest point of transformed point p i in point cloud Q; r is the threshold to determine the maximal distance between matching points.

3.3.3. Matchability Loss

To identify whether a point is able to find its correspondences from learned features, the matchaility loss should also be taken into consideration.
L m a t c h a b i l i t y = 1 N i = 1 N s i log s ^ i + ( 1 s i ) log ( 1 s ^ i )
The ground truth label s i is defined as
o i = 1 if | | T P Q ( p i ) q j | | 2 r m , 0 otherwise
where q j is a point in Q that has the closest features with p i .

4. Experiments

In this section, we display and analyze the results of different experiments. We select ICP [28], FGR [29], DeepGMR [30], and Predator [10] for comparison on partial-to-partial and noisy point clouds. The first two methods are traditional mainstream registration methods, where we adopt the implementation provided by Open3D [31]. DeepGMR is a registration method that leverages probability distributions over points to achieve robustness against noise. As the proposed method relies on the establishment of correspondences, we chose to include deepGMR as a baseline model to provide a comparison with a correspondence-free method. Additonally, we include Predator, which is one of the most outstanding attention-based models. The key difference between the proposed model and Predator lies in the incorporation of attentional networks. Specifically, Predator employs an intermediate attention module at the highest level of sampled points and features, while our method embeds attentional layers within both the encoder and decoder. In the experiments, we adhere to the code and the pretrained models they provided on ModelNet40. As the training categories were not specified in deepGMR and Predator, we utilized the results from another graph-attention-based network, GAN [32] to make comparisons on unseen categories as it employed the sample settings and protocols.
To assess the transformation result, we use the mean abolute error (MAE) and root mean squared error (RMSE) for rotation (°) and translation. The formulas for MAE and RMSE are as follows:
MAE = 1 N i = 1 N | y i y ^ i | ,
RMSE = 1 N i = 1 N ( y i y ^ i ) 2 ,
where y is the ground truth value and y ^ is the predicted value. Better performance is indicated by lower values of MAE and RMSE. For each experiment, we take the average of MAE or RMSE over all point clouds in the test dataset as the result.
In the figures displaying qualitative outcomes, the source and target point clouds are represented by yellow and blue points, respectively.

4.1. Dataset

In this study, we conduct experiments on ModelNet40 [33], which is a benchmark dataset that includes the CAD models of 12,311 objects from 40 categories. For the experiments, we firstly downsample each point cloud to 1024 points. The target point cloud is obtained by applying transformation to the source point cloud. For the transformation, rotation and translation perturbation is set to 45° and 0.5 , respectively.

4.2. Comparison on Different Estimators

Firstly, to examine which estimator yields the best performance in the registration task, we conduct experiments on the following different estimation methods:
  • prob: The interest points are sampled by probabilistic sampling, where the probability is generated by the product of overlap score and saliency score. Subsequently, RANSAC is utilized to perform feature marching and obtain the transformation matrix.
  • topk: Firstly, a similarity matrix is derived based on the Euclidean distance between the features of points from two point clouds. The top k pairs of points with the highest similarity are selected. Next, similar to prob, we conduct feature matching on those selected pairs.
  • topk-corres: It is similar to the topk approach. However, instead of conducting feature matching on those points, the transformation matrix is directly computed by applying RANSAC on the top k correspondences found in the similarity matrix.
  • topk-kabsch: It is similar to topk. The difference is that the correspondences and their similarity scores are used as the input to the kabsch algorithm, which provides a closed-form solution for computing the transformation matrix.
We evaluate the performance of different estimators with a varying number of points. The mean absolute errors are displayed in Table 1. The experimental results reveal that the best performance for most of the estimators is achieved when the number of keypoints is around 400. Compared with other estimators, the performance of probabilistic sampling is more sensitive to the number of keypoints, especially around 500 to 600. Unlike other methods that mainly rely on similarity, probabilistic sampling relates to the accuracy of the saliency score. The outcomes suggest that when more keypoints are involved, it can be increasingly difficult to identify accurate correspondences among the sampled points given the probability that is relevant to the overlap score and saliency score. In contrast, the increase in keypoints does not have a significant impact on topk-related methods. Notably, the topk estimator with RANSAC based on correspondences outperforms all other estimators. It suggests that utilizing the correspondences generated by our proposed method directly could potentially result in higher robustness.

4.3. Performance on Partial-to-Partial Point Cloud

The experiment on partial-to-partial point clouds are crucial, as it represents one of the most common scenarios in registration tasks. Table 2 presents the experimental results, where in the experiment settings, 70 % of the points are kept for both the source and target point cloud. It is evident that ICP, FGR, and deepGMR struggle to produce quality results when processing cropped point clouds. In contrast, Predator and our method exhibit better performance, as the registration errors shown in Table 2 are much smaller. It is due to the utilization of overlapping scores, which enables both methods to identify and particularly align overlapping regions. Here, our method yields lower registration errors compared with Predator, indicating its better performance in handling cropped point clouds. The qualitative results are demonstrated in Figure 3. Even if partial point clouds are presented, our method is still capable of achieving accurate alignment between two point clouds.

4.4. Performance on Point Clouds with Gaussian Noise

The results of point clouds with Gaussian noise are shown in Table 3, where the noise is sampled from Gaussian distribution, where the mean is 0 and the standard deviation is 0.01 . In contrast to the results on partial-to-partial point clouds, all models except ICP achieve better performance in this scenario. This finding suggests that the presented methods are more effective in processing complete point clouds and are relatively robust to noise. Our method shows the highest performance across all evaluation metrics. Figure 4 shows the qualitative results. Despite the addition of noise increasing the difficulty of finding accurate correspondences, our method is still able to achieve high-quality registration results.

4.5. Statistical Significance Test

The results in Section 4.3 and Section 4.4 demonstrate that Predator and our model achieved the most satisfying results. Therefore, we conducted a statistical significance test on the registration errors of point clouds in the testing set from Predator and our model. Table 4 reports the p-value of different metrics. Here, SE and AE represent squared error and absolute error, respectively, for rotation and translation. For all metrics, p-value is smaller than 1 % . This indicates that there exists statistical difference between the results derived from our model and Predator.

4.6. Performance on Point Clouds from Unseen Categories

We conduct experiments on unseen categories, where the categories of point clouds in the test set differ from those in the training set. For learning-based models, the first 20 categories are used for training, and the test is for testing. The result is presented in Table 5, which demonstrates that our model can achieve good performance for all evaluation metrics, similar to the performance achieved in other experiments. It is noteworthy that our model exhibits low variance across difference experiments. Specifically, the variation of rotation MAE is less than 1 , and the translation MAE is less than 0.01 , which indicates the ability of our model to handle different registration scenarios. The qualitative result is demonstrated in Figure 5. Even if the categories of source and target point clouds do not appear in the training dataset, our method is still able to accurately align them.

4.7. Additional Experiments on Additive Noise

In this section, to evaluate the robustness of our model, we conduct further experiments with additive noise. Given that ICP, FGR, and deepGMR yielded larger errors, we limited the investigation to Predator and our model with different estimators for better comparison. Figure 6 demonstrates the MAE of rotation and translation for different standard deviations of noise. All of our methods achieve better results than Predator under different amounts of noise. We noticed that most of the estimators on our model have similar performances, where the MAE of rotation is less than 3° and translation is smaller than 0.03 . Another observation is that although in Section 4.2 topk-corres achieves the best results, its performance is slightly worse than topk-kabsch when the amount of noise increases.

4.8. Runtime Analysis

As Predator and our method showed superior performance among all the baseline models, we conducted a runtime analysis to compare their efficiency. The test dataset was run under the same conditions, and the overall runtime was divided by the size of the dataset to calculate the runtime of each model. The result is displayed in Table 6. Most of our methods are slightly higher than the computation time of Predator, while the one with topk-kabsch as the estimator had lower runtime than Predator. The essential reason is that the kabsch algorithm provides a closed-form solution, while RANSAC requires an iterative process to sample and estimate parameters. As the results shown in Figure 6 indicate topk-kabsch achieved the lowest MAE most of the time, we argue that our model is able to produce quality and efficient results with the kabsch algorithm as the estimator.

5. Conclusions

In this paper, we propose a model to address the point cloud registration problem. To optimize the advantage of the attention mechanism while keeping computation costs reasonable, our model adopts an encoder–decoder framework with an embedded attentional layer in both the encoder and decoder. The encoder employs self-attention to capture the internal topological structure of individual point clouds, while the decoder uses cross-attention to enhance features by leveraging co-contextual information across point clouds, thereby emphasizing potential match detection. From the experiment results, our model achieved great performance on ModelNet40. However, there is still room for further improvement when it comes to handling registration tasks with other data or scenarios. At the current stage, we only focused on a synthetic dataset of objects to validate the effectiveness of our model. To examine the ability of our method in real cases, we intend to expand the scope to large-scale data, such as the real data of indoor or outdoor scenes captured by laser scanning. Given that KPConv [21] has been validated in datasets with real scenes, it is reasonable to expect that our method, which employs KPConv as its backbone model, will have the potential to handle real data as well. Moreover, as our method utilizes multiple attention layers, its applicability to large-scale data can be challenging due to the significant linear computation involved. Currently, real-time registration is critical for many downstream tasks, such as robotics and autonomous driving. However, achieving real-time performance with high-quality results can be challenging. In the future, we aim to investigate methods for reducing the computational overhead of our approach or allowing parallel computation on large-scale data while retaining its performance.

Author Contributions

Methodology, G.X.; Formal analysis, X.X.; Investigation, F.C.; Data curation, Y.L.; Writing—original draft, Z.F.; Visualization, Y.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data is openly available in [33].

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Li, J.; Saydam, S.; Xu, Y.; Liu, B.; Li, B.; Lin, X.; Zhang, W. Class-Aware Tiny Object Recognition over Large-Scale 3D Point Clouds. Neurocomputing 2023, 529, 166–181. [Google Scholar] [CrossRef]
  2. Lin, C.H.; Kong, C.; Lucey, S. Learning efficient point cloud generation for dense 3D object reconstruction. In Proceedings of the 32nd AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018; Volume 32. [Google Scholar]
  3. Liu, M. Robotic online path planning on point cloud. IEEE Trans. Cybern. 2015, 46, 1217–1228. [Google Scholar] [CrossRef] [PubMed]
  4. Woo, H.; Kang, E.; Wang, S.; Lee, K.H. A new segmentation method for point cloud data. Int. J. Mach. Tools Manuf. 2002, 42, 167–178. [Google Scholar] [CrossRef]
  5. Huang, X.; Zhang, J.; Wu, Q.; Fan, L.; Yuan, C. A coarse-to-fine algorithm for registration in 3D street-view cross-source point clouds. In Proceedings of the 2016 International Conference on Digital Image Computing: Techniques and Applications (DICTA), Gold Coast, QLD, Australia, 30 November–2 December 2016; pp. 1–6. [Google Scholar]
  6. Bai, X.; Luo, Z.; Zhou, L.; Fu, H.; Quan, L.; Tai, C.L. D3feat: Joint learning of dense detection and description of 3D local features. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 6359–6367. [Google Scholar]
  7. Choy, C.; Dong, W.; Koltun, V. Deep global registration. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 2514–2523. [Google Scholar]
  8. Gojcic, Z.; Zhou, C.; Wegner, J.D.; Guibas, L.J.; Birdal, T. Learning multiview 3D point cloud registration. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 1759–1769. [Google Scholar]
  9. Choy, C.; Park, J.; Koltun, V. Fully convolutional geometric features. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 8958–8966. [Google Scholar]
  10. Huang, S.; Gojcic, Z.; Usvyatsov, M.; Wieser, A.; Schindler, K. Predator: Registration of 3D point clouds with low overlap. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 4267–4276. [Google Scholar]
  11. Rusu, R.B.; Blodow, N.; Beetz, M. Fast point feature histograms (FPFH) for 3D registration. In Proceedings of the 2009 IEEE International Conference on Robotics and Automation, Kobe, Japan, 12–17 May 2009; pp. 3212–3217. [Google Scholar]
  12. Tombari, F.; Salti, S.; Di Stefano, L. Unique signatures of histograms for local surface description. In Computer Vision—ECCV 2010, Proceedings of the 11th European Conference on Computer Vision, Heraklion, Crete, Greece, 5–11 September 2010; Proceedings, Part III 11; Springe: Berlin/Heidelberg, Germany, 2010; pp. 356–369. [Google Scholar]
  13. Qi, C.R.; Su, H.; Mo, K.; Guibas, L.J. Pointnet: Deep learning on point sets for 3D classification and segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 652–660. [Google Scholar]
  14. Qi, C.R.; Yi, L.; Su, H.; Guibas, L.J. Pointnet++: Deep hierarchical feature learning on point sets in a metric space. Adv. Neural Inf. Process. Syst. 2017, 30, 1–10. [Google Scholar]
  15. Zeng, A.; Song, S.; Nießner, M.; Fisher, M.; Xiao, J.; Funkhouser, T. 3dmatch: Learning local geometric descriptors from rgb-d reconstructions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 1802–1811. [Google Scholar]
  16. Sarlin, P.E.; DeTone, D.; Malisiewicz, T.; Rabinovich, A. Superglue: Learning feature matching with graph neural networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 4938–4947. [Google Scholar]
  17. Fischer, K.; Simon, M.; Olsner, F.; Milz, S.; Gross, H.M.; Mader, P. Stickypillars: Robust and efficient feature matching on point clouds using graph neural networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 313–323. [Google Scholar]
  18. Shi, C.; Chen, X.; Huang, K.; Xiao, J.; Lu, H.; Stachniss, C. Keypoint matching for point cloud registration using multiplex dynamic graph attention networks. IEEE Robot. Autom. Lett. 2021, 6, 8221–8228. [Google Scholar] [CrossRef]
  19. Qin, Z.; Yu, H.; Wang, C.; Guo, Y.; Peng, Y.; Xu, K. Geometric transformer for fast and robust point cloud registration. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 11143–11152. [Google Scholar]
  20. Yew, Z.J.; Lee, G.H. Regtr: End-to-end point cloud correspondences with transformers. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 6677–6686. [Google Scholar]
  21. Thomas, H.; Qi, C.R.; Deschaud, J.E.; Marcotegui, B.; Goulette, F.; Guibas, L.J. Kpconv: Flexible and deformable convolution for point clouds. In Proceedings of the IEEE/CVF international Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 6411–6420. [Google Scholar]
  22. Wang, Y.; Sun, Y.; Liu, Z.; Sarma, S.E.; Bronstein, M.M.; Solomon, J.M. Dynamic graph cnn for learning on point clouds. ACM Trans. Graph. 2019, 38, 1–12. [Google Scholar] [CrossRef]
  23. Choy, C.; Gwak, J.; Savarese, S. 4D spatio-temporal convnets: Minkowski convolutional neural networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–17 June 2019; pp. 3075–3084. [Google Scholar]
  24. Wang, Y.; Solomon, J.M. Deep closest point: Learning representations for point cloud registration. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 3523–3532. [Google Scholar]
  25. Fu, K.; Liu, S.; Luo, X.; Wang, M. Robust point cloud registration framework based on deep graph matching. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 8893–8902. [Google Scholar]
  26. Li, J.; Lee, G.H. Usip: Unsupervised stable interest point detection from 3D point clouds. In Proceedings of the IEEE/CVF international Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 361–370. [Google Scholar]
  27. Sun, Y.; Cheng, C.; Zhang, Y.; Zhang, C.; Zheng, L.; Wang, Z.; Wei, Y. Circle loss: A unified perspective of pair similarity optimization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 6398–6407. [Google Scholar]
  28. Besl, P.J.; McKay, N.D. Method for registration of 3-D shapes. Proc. Sens. Fusion IV Control Paradig. Data Struct. 1992, 1611, 586–606. [Google Scholar] [CrossRef]
  29. Zhou, Q.Y.; Park, J.; Koltun, V. Fast global registration. In Computer Vision—ECCV 2016, Proceedings of the 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Proceedings, Part II 14; Springer: Berlin/Heidelberg, Germany, 2016; pp. 766–782. [Google Scholar]
  30. Yuan, W.; Eckart, B.; Kim, K.; Jampani, V.; Fox, D.; Kautz, J. Deepgmr: Learning latent gaussian mixture models for registration. In Computer Vision—ECCV 2020, Proceedings of the 16th European Conference, Glasgow, UK, 23–28 August 2020; Proceedings, Part V 16; Springer: Berlin/Heidelberg, Germany, 2020; pp. 733–750. [Google Scholar]
  31. Zhou, Q.Y.; Park, J.; Koltun, V. Open3D: A modern library for 3D data processing. arXiv 2018, arXiv:1801.09847. [Google Scholar]
  32. Song, Y.; Shen, W.; Peng, K. A novel partial point cloud registration method based on graph attention network. Vis. Comput. 2023, 39, 1109–1120. [Google Scholar] [CrossRef]
  33. Wu, Z.; Song, S.; Khosla, A.; Yu, F.; Zhang, L.; Tang, X.; Xiao, J. 3D shapenets: A deep representation for volumetric shapes. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1912–1920. [Google Scholar]
Figure 1. Application of point cloud on bolt detection in the mining industry.
Figure 1. Application of point cloud on bolt detection in the mining industry.
Sensors 23 04123 g001
Figure 2. Architecture of the proposed model.
Figure 2. Architecture of the proposed model.
Sensors 23 04123 g002
Figure 3. Qualitative results of partial-to-partial point clouds.
Figure 3. Qualitative results of partial-to-partial point clouds.
Sensors 23 04123 g003
Figure 4. Qualitative results of point clouds with Gaussian noise.
Figure 4. Qualitative results of point clouds with Gaussian noise.
Sensors 23 04123 g004
Figure 5. Qualitative results of point clouds from unseen categories.
Figure 5. Qualitative results of point clouds from unseen categories.
Sensors 23 04123 g005
Figure 6. Results on additive noise and cropped data. It displays the MAE of rotation (°) and translation versus the standard deviation of noise.
Figure 6. Results on additive noise and cropped data. It displays the MAE of rotation (°) and translation versus the standard deviation of noise.
Sensors 23 04123 g006
Table 1. Mean absolute error of rotation (°) and translation from different estimators with various number of keypoints.
Table 1. Mean absolute error of rotation (°) and translation from different estimators with various number of keypoints.
#Keypoints100200300400500600
prob2.7544/0.052262.3391/0.043052.0229/0.040812.3391/0.019394.2302/0.0327110.5938/0.07608
topk1.6285/0.033981.5996/0.033651.6234/0.034991.6099/0.034231.6323/0.035371.6394/0.03498
topk-corres1.3452/0.037361.2001/0.033371.1640/0.032981.0684/0.032861.0878/0.034091.0931/0.03395
topk-kabsch1.6367/0.032981.6119/0.033291.6111/0.033401.6048/0.033401.6085/0.033561.6179/0.03357
Table 2. The results of partial-to-partial point clouds.
Table 2. The results of partial-to-partial point clouds.
MethodRMSE (R)RMSE (t)MAE (R)MAE (t)
ICP29.02950.993822.69220.8716
FGR58.30710.312236.86450.2096
DeepGMR87.69450.400166.05530.2933
Predator9.15730.06873.43750.0290
Our2.81600.03181.10540.0102
Table 3. The results of point clouds with Gaussian noise.
Table 3. The results of point clouds with Gaussian noise.
MethodRMSE (R)RMSE (t)MAE (R)MAE (t)
ICP29.76090.993123.03220.8704
FGR40.34780.191019.35860.1025
DeepGMR45.87620.187723.52130.09801
Predator4.61800.03711.83490.0144
Our1.88330.01380.62410.0050
Table 4. The p-value of squared error and absolute error for rotation and translation. The first row presents the results for partial-to-partial registration, while the second row presents the registration results for point clouds with Gaussian noise.
Table 4. The p-value of squared error and absolute error for rotation and translation. The first row presents the results for partial-to-partial registration, while the second row presents the registration results for point clouds with Gaussian noise.
ExperimentSE (R)SE (t)AE (R)AE (t)
Partial 1.24 × 10 9 2.25 × 10 8 4.32 × 10 30 8.74 × 10 33
Noisy 9.44 × 10 7 4.87 × 10 8 4.07 × 10 24 6.14 × 10 30
Table 5. The results of point clouds with unseen categories.
Table 5. The results of point clouds with unseen categories.
MethodRMSE (R)RMSE (t)MAE (R)MAE (t)
ICP29.02950.993822.69220.8716
FGR58.36230.303437.13490.2086
GAN5.9770.0464.5340.021
Our3.13710.03141.26080.0117
Table 6. The runtime of our model with different estimators and Predator.
Table 6. The runtime of our model with different estimators and Predator.
MethodOur (Prob)Our (Topk)Our (Topk-Kabsch)Our (Topk-Corres)Predator
Time (ms)4950415543
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Xia, X.; Fan, Z.; Xiao, G.; Chen, F.; Liu, Y.; Hu, Y. Learning Representative Features by Deep Attention Network for 3D Point Cloud Registration. Sensors 2023, 23, 4123. https://doi.org/10.3390/s23084123

AMA Style

Xia X, Fan Z, Xiao G, Chen F, Liu Y, Hu Y. Learning Representative Features by Deep Attention Network for 3D Point Cloud Registration. Sensors. 2023; 23(8):4123. https://doi.org/10.3390/s23084123

Chicago/Turabian Style

Xia, Xiaokai, Zhiqiang Fan, Gang Xiao, Fangyue Chen, Yu Liu, and Yiheng Hu. 2023. "Learning Representative Features by Deep Attention Network for 3D Point Cloud Registration" Sensors 23, no. 8: 4123. https://doi.org/10.3390/s23084123

APA Style

Xia, X., Fan, Z., Xiao, G., Chen, F., Liu, Y., & Hu, Y. (2023). Learning Representative Features by Deep Attention Network for 3D Point Cloud Registration. Sensors, 23(8), 4123. https://doi.org/10.3390/s23084123

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop