Next Article in Journal
Blind Image Deblurring Based on Local Edges Selection
Next Article in Special Issue
Target Points Tracking Control for Autonomous Cleaning Vehicle Based on the LSTM Network
Previous Article in Journal
Comparison of Water Sampling between Environmental DNA Metabarcoding and Conventional Microscopic Identification: A Case Study in Gwangyang Bay, South Korea
Previous Article in Special Issue
A Low Overhead Mapping Scheme for Exploration and Representation in the Unknown Area
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

3-D Point Cloud Registration Using Convolutional Neural Networks

1
Department of Electrical Engineering, National Taipei University of Technology, Taipei Tech Box 2125, Taipei 10608, Taiwan
2
Department of Electrical Engineering, National Taipei University of Technology, Taipei 106, Taiwan
*
Author to whom correspondence should be addressed.
Appl. Sci. 2019, 9(16), 3273; https://doi.org/10.3390/app9163273
Submission received: 3 July 2019 / Revised: 30 July 2019 / Accepted: 5 August 2019 / Published: 9 August 2019
(This article belongs to the Special Issue Intelligent Robotics)

Abstract

:
This paper develops a registration architecture for the purpose of estimating relative pose including the rotation and the translation of an object in terms of a model in 3-D space based on 3-D point clouds captured by a 3-D camera. Particularly, this paper addresses the time-consuming problem of 3-D point cloud registration which is essential for the closed-loop industrial automated assembly systems that demand fixed time for accurate pose estimation. Firstly, two different descriptors are developed in order to extract coarse and detailed features of these point cloud data sets for the purpose of creating training data sets according to diversified orientations. Secondly, in order to guarantee fast pose estimation in fixed time, a seemingly novel registration architecture by employing two consecutive convolutional neural network (CNN) models is proposed. After training, the proposed CNN architecture can estimate the rotation between the model point cloud and a data point cloud, followed by the translation estimation based on computing average values. By covering a smaller range of uncertainty of the orientation compared with a full range of uncertainty covered by the first CNN model, the second CNN model can precisely estimate the orientation of the 3-D point cloud. Finally, the performance of the algorithm proposed in this paper has been validated by experiments in comparison with baseline methods. Based on these results, the proposed algorithm significantly reduces the estimation time while maintaining high precision.

Graphical Abstract

1. Introduction

Automated assembly systems capable of repeatedly performing assembly tasks with high-precision and high-speed play an indispensable role in modern manufacturing lines. Intuitively, to perform these automated assembly tasks, cameras are employed for the purpose of providing vision feedback to estimate the pose (position and orientation) of objects to the system. Based on the feedback, the system controls manipulators to perform object picking and assembly tasks in order to interact with the environment. In recent years, much of the work related to the vision-based automated assembly was published. S. Huang et al. [1,2] proposed the dynamic compensation to realize a fast peg-in-hole alignment in the case of a large position and attitude uncertainty based on vision information provided by high-speed cameras. In References [3,4], 3-D scenes used to estimate the pose of objects were reconstructed from multiple images acquired by several cameras. Nonetheless, these algorithms encountered some problems when extracting features as well as reconstructing 3-D feature points from images in the case of general objects that have no distinctly geometric and color features. Thus, in order to address these problems, in Reference [2], color features were added to objects which are not real in general automated assembly systems. Besides, in Reference [3], B. Wang et al. reconstructed 3-D point clouds of the object surface by employing two Charge-Coupled Device (CCD) cameras integrated with a laser beam to sweep over objects which increased the complexity in the configuration of these systems. Chang [4] proposed a vision-based control algorithm for automatic back shell assembly related to smartphone assembly based on 3-D reconstruction by two uniocular fields of view of two cameras. In Reference [5], T. R. Savarimuthu et al. developed a three-level cognitive system to perform a complex assembly task, Cranfield benchmark, based on learning by demonstration using 3-D point clouds extracted from BumbleBee2 stereo camera for the purpose of detecting the object and estimating its pose and a Kinect sensor with the aim of tracking the object in real-time, respectively.
With the aforementioned explanations, 3-D point cloud data brings reliability and capability applied for most categories of objects without demanding special features. Moreover, based on great breakthroughs in manufacturing technology, 3-D cameras or 3-D scanners become increasingly common in automation fields because of redeeming features such as the precision, speed and reliability provided by these 3-D cameras. As a result, applications such as 3-D pose estimation, 3-D object detection and segmentation and 3-D point cloud matching based 3-D spatial information were focused in recent years. In particular, regarding 3-D pose estimation, K. Chan et al. [6] developed a 3-D point cloud system with the aim of estimating human pose by using a depth sensor whilst a fusion framework for the purpose of 3-D modeling, object detection and pose estimation from a given set of point clouds in the case of occlusion and clutter is proposed in Reference [7]. Besides, a generic 3-D detection VoxelNet network was developed in Reference [8] to accurately detect objects in 3-D point clouds whereas R. Q. Charles et al. [9] proposed a unified architecture, PointNet, that directly employs 3-D point clouds as the input to carry out object classification, part segmentation and scene semantic parsing. Moreover, in Reference [10], 3-D point clouds and 2-D images were fused to segment complicated 3-D urban scenes in the case of a large scale by using deep learning. When it comes to point cloud matching, in our previous research [11], we proposed to apply a learning-based algorithm to resolve 2-D point cloud matching by using a deep neural network. Therefore, a very challenging task, 3-D point cloud matching, also known as 3-D point cloud registration, that is really meaningful in estimating the pose of objects using convolutional neural networks in automated assembly systems, has been addressed in this paper.

2. Related Work

Basically, point cloud registration consists of two stages including rough registration and precise registration stage for algorithms based on Iterative Closest Point (ICP) approach and its variants. In particular, the rough registration is performed to give an initial estimation about the transformation between two point clouds, followed by the precise registration to refine this estimation to best align these two point clouds. On the one hand, to address the rough registration problem, a well-known algorithm, RANdom Sample Consensus (RANSAC) approach, was proposed in Reference [12] to roughly estimate the rotation and translation between two point clouds. Because of slow computational speed, several variants of the RANSAC algorithm were developed to address this problem such as IMPSAC [13], Preemptive RANSAC [14], Distributed RANSAC [15] and Recursive-RANSAC [16]. Besides, Han et al. [17] proposed an improved RANSAC algorithm based on 3-D Region Covariance (RC) descriptor, RC–RANSAC method, to enhance the successful percentage of point cloud registration in a single repetition when describing region covariance of each point. Additionally, other global registration algorithms in the case of arbitrary initial poses between two point clouds were also published in recent years. In Reference [18], Aiger et al. developed a robust and fast alignment principle, 4-Points Congruent Sets (4PCS) algorithm, employing wide bases robust to outliers and noise for 3-D point clouds based on extracting all sets of coplanar four-points from a 3-D point cloud. Moreover, similarity and affine transform problems were taken into account in their research. However, a drawback existed in their method because 4PCS approach took approximately a quadratic time complexity according to the quantity of data points. Hence, they presented an improvement of 4PCS method, called Super 4PCS approach [19], that degraded the quadratic complexity to linear complexity and confined the volume of candidate conjugate pairs created to tackle this negative point. Super 4PCS approach appeared to be an effective solution to handle 3-D point cloud registration when existing a low overlap and outliers. The other existing approach, the Iterative Closest Point method integrated with the Normal Distribution Transform approach (NDT-ICP), was proposed in Reference [20] with the aim of accelerating the registration speed. In addition, Liu et al. [21] developed a point cloud registration algorithm based on improved digital image correlation approach, followed by the ICP method with a high accuracy, anti-noise capability and efficiency. Moreover, Chang et al. proposed the candidate-based axially switching (CBAS) computed closer point (CCP) algorithm [22] to effectively and robustly address rough registration problem in the case of an arbitrarily initial pose between two point clouds.
On the other hand, as regards the precise registration fulfilled after the rough registration, the most common registration algorithm is the ICP approach proposed by Besl et al. in Reference [23] that minimizes the distance between two point clouds. In addition, many variants of the ICP method were presented in recent years with the aim of improving the original method in some aspects such as high accuracy, high speed, robust registration, scale registration, affine registration and globally-optimal solution. Particularly, dual interpolating point-to-surface algorithm adopting surface fitting by using B-spline to establish correspondences between two point clouds was proposed in Reference [24] to address the precise registration for two point clouds with a partial overlap. Chen et al. [25] developed the Hong-Tan based ICP method (HT-ICP) for the purpose of improving the speed and precision of registration for partially overlapping point clouds. Additionally, a registration algorithm by using the hand and soft assignments to improve the robustness and accuracy for partially overlapping point cloud registration was proposed in Reference [26]. To extend the robustness to noise and/or outliers, the Trimmed ICP approach was presented in Reference [27] that applied the Least Trimmed Squares algorithm in all stages of the operation. In Reference [28], Phillips et al. presented the Fraction ICP method that can robustly identify and discard outliers and guarantee the convergence to a locally optimal solution. The other existing approach, Efficient Sparse ICP method, using sparsity-inducing norms to considerably improve robust registration caused by noise and outliers whilst employing a Simulated Annealing search in order to efficiently resolve the underlying optimization problem was proposed in Reference [29]. Yang et al. [30] developed the ICP-based algorithm integrating with a branch-and-bound (BnB) principle to figure out the globally optimal solution for Euclidean registration between two 3-D surfaces or 3-D point clouds starting from any initialization. Scale-ICP approach integrating a scale factor with the original ICP method was proposed in Reference [31] so as to accurately perform the registration between two 3-D point clouds in the case of large-scale stretches and noises. Wang et al. [32] developed a multi-directional affine registration (MDAR) algorithm by using the statistical characteristics and shape features of point clouds for affine registration problem. In addition to the aforementioned distance-minimization-based registration algorithms, another registration approach using distinct features extracted from 3-D point cloud data sets was also presented in References [33,34,35,36], to robustly and efficiently enhance the registration problem. Particularly, many feature descriptors including local-based descriptor, global-based descriptor and hybrid-based descriptor were presented in recent years and a comprehensive review of the existing algorithms can be sought in Reference [37].
Nevertheless, the aforementioned existing algorithms all are based on randomly selecting different points in two 3-D point clouds to form a pair of corresponding bases in rough registration phase whilst iteratively minimizing the distance between two 3-D point clouds until the convergence within the desired tolerance of ICP-based methods in the precise registration phase. These are the reasons why those algorithms cannot be employed in real-time closed-loop object picking and assembly systems due to their slow computation. In addition to the aforementioned algorithms, there are some approaches related to point cloud registration using deep learning which were developed in recent years. In Reference [38], a point cloud registration method using a deep neural network auto-encoder for outdoor localization between a close-proximity scanned point cloud and a large-scale point cloud was proposed. Specially, Aoki et al. [39] developed an efficient and robust point cloud registration architecture, PointNetLK architecture, based on an iterative image registration approach [40] and PointNet architecture [9]. However, the PointNetLK architecture involved iterative computation with the number of iterations in accordance with a minimum threshold of the optimal twist parameters. In this paper, a seemingly novel efficient algorithm using deep learning is proposed to accelerate the estimation time while maintaining the required accuracy.
The key contributions of this research can be summarized as follows:
  • We propose a seemingly novel registration architecture including 3-D convolutional neural networks (3-D CNNs) for the rough registration stage and 2-D CNNs for the precise registration stage that transforms the 3-D point cloud registration problem to a regression problem with estimated outputs as angular disparities between the model point cloud and a data point cloud.
  • We propose two descriptors in order to extract distinct features such as Simplified Spatial Point Distribution and 8-Corner-Based Closest Points from a given 3-D point cloud as data sets for the purpose of training the proposed 3-D convolutional neural network (3-D CNN) model and the proposed 2-D CNN model, respectively. After training, the trained CNN models can quickly estimate the rotation between two 3-D point clouds, followed by the translation estimation based on computing average values of these two point cloud data sets in fixed time with high precision.
The remainder of this paper is organized as follows. Section 3 shows the problem formulation and the strategy of 3-D point cloud registration using CNNs. In Section 4, two categories of descriptors employed to extract coarse and detailed features in terms of the orientation of a 3-D point cloud are proposed with the aim of generating training data sets. Section 5 develops the proposed registration architecture which consists of two stages including the rough and precise registration based on 3-D CNN and 2-D CNN architecture, respectively. Finally, Section 6 provides experimental results whilst conclusions of this paper are drawn in Section 7.

3. Problem Formulation

Given the model point cloud P and a data point cloud Q, point cloud registration is to estimate the transformation including the rotation matrix R and the translation vector t between these two point clouds.
P = { p 1 , p 2 , , p n } , Q = { q 1 , q 2 , , q m }
where p i and q j stand for the i t h point and j t h point in the model and data point cloud, respectively. To fulfill this task, ICP-based approaches minimize an objective function, the mean squared error, based on corresponding closest point pairs ( p i , q j ) in two point clouds as follows.
ξ ( R , t ) = 1 m j = 1 m | | p i R q j t | | 2
where m is the number of points in the data point cloud. Because the ICP algorithm can only successfully match if the uncertainty between two point clouds is small enough, an initial registration stage is necessary to address a large uncertainty problem based on the well-known RANSAC approach. Thus, inspired by this two-stage registration process, we propose a deep CNN framework with two consecutive CNN models including the first model for the rough registration and the second model for the precise registration. In the proposed architecture, only rotation angles are estimated instead of estimating both rotation and translation because the translation can be computed by subtracting coordinate values of two corresponding center points of two point clouds. In addition, because of arbitrary translations between two point clouds in 3-D space, it is not reasonable to use the translation vector as ground truth for the purpose of training. Moreover, the limitation of rotations is confined in the range from 180 to + 180 that make rotation angles potential ground truths for the convergence of the learning process. Figure 1 shows a block diagram of the proposed architecture for 3-D point cloud registration.
Hence, point clouds of an object with different rotation angles in the x y z coordinate system and these corresponding rotation angles are collected with the aim of creating a data set. As regards the CNN architecture, the first CNN model in the rough registration stage is to cover an arbitrary initial rotation in a full range of uncertainty from 180 to 180 whilst the second CNN model is used in order to refine the rotation estimation in covering a smaller range with a higher angular resolution in the precise registration stage. In particular, in this precise registration stage, a trade-off about selecting the range of the uncertainty of the data set should be satisfied in a such way that this uncertainty is small enough to make the CNN model learn deep characteristics and large enough to cover errors existing after the rough registration stage. In this research, the range of this uncertainty is selected from 15 to 15 based on experiments. Ideally, only one CNN model can be considered to handle the registration problem if the number of samples for training is large enough to cover the full range of uncertainty with a high resolution of rotation angles. Nonetheless, it is impossible to realize this idea as a huge data set is required in this case that leads to a difficulty in the convergence of solutions and the time-consuming problem during the training process due to using too much training data. Thus, the precision of the rotation estimation is reasonably guaranteed by using this two-stage CNN model. To demonstrate this estimation process, a simple representation of the model and data point cloud in a plane is visualized in Figure 2. Initially, the arbitrarily angular disparity α is witnessed between model and data which is covered in the full range of uncertainty from 180 to 180 . After the first stage, data can be transformed to d a t a 1 based on a rough estimation α 1 that exists an error α α 1 compared with the ground truth α . Then, the second CNN model seemingly precise estimates the disparity α 2 between model and d a t a 1 to finalize the whole angular disparity α 1 + α 2 between model and data.
In addition, because points in a general 3-D point cloud are not in order and the number of points is not fixed according to different viewpoints for the same object, a descriptor extracting distinct features from a point cloud is employed to tackle these problems. In this case, a powerful descriptor must be able to capture the geometric structure of a data point cloud that can precisely describe the orientation of this point cloud in the coordinate frame and be invariant to translation. Thereby, instead of directly inputting 3-D point cloud data sets, features extracted from 3-D point clouds based on the descriptors are employed as the input of these CNN models. By using features, every sample in the data set has the same format such as the same order and the same finite size, which is good for the learning process. In accordance with using these descriptors, a comprehensive architecture of the proposed algorithm is illustrated in Figure 3.

4. The Proposed Descriptors for 3-D Point Clouds

In this section, two types of descriptors used to extract features in the rough and precise registration stage, respectively, with the aim of generating the training data sets are given. Intuitively, the first descriptor should extract coarse features roughly describing the orientation of an object to ensure the fast convergence in the learning process whilst detailed features precisely expressing the orientation of the object should be captured by the second descriptor. Before showing procedures to compute these two descriptors, two corresponding data sets in the case of a large uncertainty of rotation angles from 180 to 180 and a small uncertainty from 15 to 15 , need to be created in advance.

4.1. Point Cloud Data Set Preparation

In order to generate the data set, several 3-D point clouds with different orientations of an object need to be collected. Nevertheless, it is not suitable to collect a large number of 3-D point clouds of an object when using a 3-D scanner to capture these point cloud data sets. Therefore, the data set can be obtained by randomly rotating and downsampling the model point cloud of the object. Particularly, the box grid filter algorithm provided by Matlab is used to downsample the rotated model point cloud to low-resolution point clouds with the different number of points and distribution characteristics. Based on this approach, the created point clouds can simulate disparities among point clouds that would exist in real 3-D point clouds captured by a 3-D camera. The principal difference between two kinds of data sets of an object is that each rotation angle with respect to an axis is randomly initialized with 80 different values in the range from 180 to 180 for the first data set whereas the second data set is generated based on 80 random values of each rotation angle in the range from 15 to 15 . Totally, there are 3 rotation angles with respect to x, y, and z axis that results in generating 80 3 = 512,000 point clouds with different orientations. Each data set is divided into 3 subsets including training set (392,000 samples), validation set (60,000 samples) and test set (60,000 samples). Algorithm 1 shows procedures step by step in order to generate these data sets.
Algorithm 1 Data set generation
Applsci 09 03273 i001

4.2. Simplified Spatial Point Distribution Descriptor

A 3-D data point cloud Q employed as an input of the descriptors can be given as follows
Q = q 1 q 2 q m = x 1 y 1 z 1 x 2 y 2 z 2 x m y m z m
where q j denotes the j t h data point. In the first stage, the rough registration stage, features extracted from the point set Q should generally describe the orientation of Q such that a successful and fast convergence in the learning process can be guaranteed. Because the rough registration covers a full range of uncertainty from 180 to 180 that leads to significant disparities in characteristics of these features, it is hard to obtain optimal weights with a small enough loss for a CNN model after training. Therefore, a Simplified Spatial Point Distribution (SSPD) descriptor is proposed in this paper to describe coarse characteristics of the orientation of a 3-D data point cloud in 3-D space. The key point is to divide a spatial region involving the point set into several sub-regions and count the number of points in each sub-region, followed by a normalization with the total quantity of points of the point cloud. To begin with, a cubic bounding box with the length of an edge l c u b e is created that consists of the point set.
l c u b e = max norm 2 q i , q j | i j
where n o r m 2 denotes the Euclidean distance between two points q i and q j . Secondly, this cube is divided into several sub-cubes as shown in Figure 4 with the size of s x s x s where the length of an edge of a sub-cube can be computed as follows.
l s u b c u b e = l c u b e s
Finally, the quantity of points in each sub-cube is counted and then divided by the total number of points of the point cloud. Because each sub-cube has one value, SSPD is a 3-D matrix with the size of s × s × s where the parameter s = 15 is selected based on experiments. Algorithm 2 illustrates steps in details to compute SSPD based on a given 3-D point cloud.
Algorithm 2 Offline computations of SSPD
Applsci 09 03273 i002

4.3. 8-Corner-Based Closest Points Descriptor

In comparison with the SSPD descriptor for the rough registration, features employed for training a CNN model to precisely estimate the spatial orientation of a given 3-D point cloud should be able to thoroughly describe the characteristics of this point set in details. Thereby, 8-Corner-Based Closest Points (8CBCP) descriptor is proposed in this paper for the purpose of extracting oriented characteristics of a 3-D point cloud in 3-D space. As shown in the name of this descriptor, the main point is to find d closest points according to each corner of the Axis-Aligned Bounding Box (AABB) [41] of a data point cloud. Based on considering each corner of AABB as a viewpoint when observing the point cloud, the 8CBCP descriptor possesses the robustness in describing oriented features of this point cloud and a simple format because only key points can be collected. In addition, by numbering corners and arranging closest points with a suitably predetermined order, the problem without a specific order of a general point cloud can be effectively addressed. Moreover, to tackle the problem about the disparity in the point distribution among point clouds, the coordinate values of these closest points are normalized with the total number of points of the corresponding point cloud. Hence, the result of 8CBCP is a 2-D matrix with the size of 8 d × 3 because 8 d corresponding closest points can be obtained based on 8 corners. In this research, d is selected based on a trade-off between the total number of points of a point cloud at approximately 1000 points after downsampling and the necessary number of points according to each corner for the purpose of describing orientation features of a point set. Thus, by testing in experiments, d is chosen at 40 points. Although this paper focuses on whole point cloud registration, 8CBCP can be a potential descriptor when resolving partial point cloud registration problem due to multiple different observations applied for one object. Algorithm 3 shows computation procedures in details to generate 8CBCP features of a given 3-D point cloud.
Algorithm 3 Offline computations of 8CBCP
Input: Data point cloud Q ;
1: Compute A A B B and use these parameters to make a cube that bounds of the data point cloud;
   w = max x q i | i = 1 , 2 , , m min x q i | i = 1 , 2 , , m ;
   d = max y q i | i = 1 , 2 , , m min y q i | i = 1 , 2 , , m ;
   h = max z q i | i = 1 , 2 , , m min z q i | i = 1 , 2 , , m ;
2: Shift all points to the first quadrant;
   x q i = x q i min x q i | i = 1 , 2 , , m ; y q i = y q i min y q i | i = 1 , 2 , , m ;
   z q i = z q i min z q i | i = 1 , 2 , , m ;
3: Divide the cube into 8 sub-cubes numbered from 1 to 8 and bin points into the according sub-cube as shown in Figure 5;
4: Based on 8 corner points, red points, of the cube, scan d closest points in the corresponding sub-cube of each corner as illustrated in Figure 6;
5: According to the closest points in terms of each corner in the corresponding sub-cube, consider each corner as a local origin O k ( k = 1 , 2 , , 8 ) and transfer coordinate values of these closest points with respect to the corresponding local frame (Figure 7), followed by a normalization by the total number of points m of the data point cloud;
{ k } α i { k } β i { k } γ i = 1 m x q i y q i z q i x O k y O k z O k ;
where { k } α i , { k } β i , { k } γ i stand for three coordinates according to three axes of point q i with respect to the local frame { k } ;
6: Compute 8CBCP, a 2-D matrix with the size of 8 d x 3;
8 CBCP = { 1 } α 1 { 1 } α d { 2 } α 1 { 2 } α d { 8 } α 1 { 8 } α d { 1 } β 1 { 1 } β d { 2 } β 1 { 2 } β d { 8 } β 1 { 8 } β d { 1 } γ 1 { 1 } γ d { 2 } γ 1 { 2 } γ d { 8 } γ 1 { 8 } γ d T ;
Output: 8CBCP of a data point cloud.

5. The Proposed CNN Architectures

Based on two kinds of features including SSPD and 8CBCP shown in Section 4, the corresponding CNN architectures are developed in this section with the aim of estimating the orientation of a 3-D point cloud on the generated point cloud data set. Related to the first stage, the rough registration, because the input of CNN model is SSPD, a 3-D matrix with the size of s × s × s, a 3-D CNN model is selected. When it comes to the second stage, the precise registration, a 2-D CNN model is employed in accordance with the size of the input 8CBCP in the case of a 2-D matrix. Figure 8 shows the proposed 3-D architecture for the rough registration whilst the proposed 2-D architecture for the precise registration is demonstrated in Figure 9.

5.1. 3-D CNN Architecture for Rough Registration

As shown in Figure 8, to start with, SSPD with the size of s × s × s extracted from an input point cloud is convoluted with each of 16 kernels with the size of 2 × 2 × 2 that results in 16 feature maps in the layer C1. Then, to continuously increase the number of feature maps, we apply 3-D convolutions with the same kernel size 2 × 2 × 2 as the first convolutional layer for the layer C2 and C3. Consequently, the layer C2 has 32 feature maps while there are 64 feature maps in the layer C3. Additionally, in these layers C1, C2 and C3, zero padding is used in convolutional operations such that the output of feature maps has the same size as the original input. Moreover, after the convolutional operation, the result is added with a bias and then passed through a hyperbolic tangent activation function t a n h ( . ) to generate the output of each feature map as follows.
χ k = t a n h ( b k + i w k i x i )
where χ k stands for the output of the k t h feature map, b k is a bias of the k t h feature map, w k i denotes the weight of kernel connected from the i t h input map, x i means the i t h input map and * denotes the convolutional operation. In the subsequent layer, the 3-D max-pooling layer S1 with a window size of 2 × 2 × 2 is applied to each of the feature maps in the layer C3 that results in the same quantity of feature maps but a reduced spatial dimension, which is two times smaller. Next, the output of the max-pooling layer S1 is flattened to become a feature vector such that each element in this vector is fully connected to each of 2048 neurons in the layer H1. Similarly, the output of the layer H1 is fully connected to the layer H2 with 1024 neurons whilst the fully connected layer H3 has 512 neurons connected from the output of the layer H2. Finally, neurons of the last fully connected layer H3 are connected to 3 neurons representative for the estimation of Euler angles in the output layer. The activation function used in each neuron in these fully connected layers is t a n h ( . ) function.

5.2. 2-D CNN Architecture for Precise Registration

With the explanations in the previous section, because the 8CBCP descriptor collects d closest points according to each corner of AABB to generate one sample, there are totally 8 closest point groups that play an equal role in describing the orientation of a 3-D point cloud. Inspired by this idea, the proposed architecture of the 2-D CNN model employs 8 sub-models in parallel in order to independently estimate Euler angles, followed by an average computation of these estimated Euler angles. As shown in Figure 9, as regards the first sub-model, the first part with the size of d × 3 of 8CBCP extracted from an input point cloud is convoluted with each of 8 kernels with the size of 5 × 1 that results in 8 feature maps in the layer C1. Then, the quantity of feature maps is continuously grown in the sequential layers by applying 2-D convolutional operations with a kernel size of 5 × 1 similar to the previous layer C1. As a result, the layer C2 and C3 have 16 feature maps and 32 feature maps, respectively. Similar to 3-D CNN architecture used in the rough registration, zero padding is applied in convolutional operations so as to make no modification about the size of the output of feature maps compared with the original input. Additionally, before being connected to the next layers, the output of the convolutional operation is added by a bias and then delivered to the activation function t a n h ( . ) to create the output of feature maps. Next, the output of feature maps in the layer C3 is passed through the 2-D max-pooling layer S1 with a window size of 2 × 1 that leads to a reduced spatial dimension, which is two times smaller but the same number of feature maps. Then, the output of the max-pooling layer S1 is flattened as a feature vector and elements in this vector is fully connected to each of 1024 neurons in the layer H1. In the same way, the output of the layer H1 is fully connected to 512 neurons in the layer H2 and the outputs of the layer H2 are fully connected to the layer H3 with 256 neurons. Eventually, neurons in the layer H3 are connected to 3 neurons in the output layer O u t p u t 1 which figures out a more precise estimation of Euler angles ψ 1 , θ 1 , ϕ 1 in comparison with the estimation in the rough registration stage. Besides, the activation function applied for neurons in fully connected layers is t a n h ( . ) function. Similar to the first sub-model, 7 sub-models left from the second sub-model to the eighth sub-model possess the same 2-D CNN architecture with the output from ψ 2 , θ 2 , ϕ 2 to ψ 8 , θ 8 , ϕ 8 , respectively. After all estimations of 8 sub-models, an average computation of these Euler angles is required to obtain the mean estimation of the orientation of a data point cloud as the output of the whole 2-D CNN model as follows.
ψ θ ϕ T = 1 8 k = 1 8 ψ k k = 1 8 θ k k = 1 8 ϕ k T

6. Experiments

In this section, experimental results for 3-D point cloud registration are provided to validate the proposed registration algorithm based on CNN architectures. In particular, when it comes to point cloud data, open data including “Bunny”, “Dragon” & “Buddha”, “Armadillo” and “Horse” from the Stanford 3-D Scanning Repository [42,43,44] and NTU CSIE [45], respectively, are employed. Besides, two algorithms such as the “RANSAC then ICP” (RANSAC+ICP) algorithm using Fast Point Feature Histograms (FPFH) [46] to describe features for point clouds and the “Fast Global Registration” (FGR) algorithm [47] provided in Open3D [48,49], a modern library for 3-D data processing, are applied to evaluate the performance of the proposed algorithm. Moreover, in order to give the quantitative evaluation, in addition to the computation time, the mean squared error (MSE) value used to compare the matching rate among these methods is given as follows [22].
M S E = 1 m j = 1 m | | p i q j | |
where m is the number of points in the data point cloud and p i and q j stand for the corresponding closest points in the model point cloud and the data point cloud, respectively. All programs performing these approaches are programmed in Python on an Intel Core i9-9900K CPU @3.6GHz PC equipped with a GeForce RTX 2080 Ti SEA HAWK X Graphics Processing Unit (GPU). In the training process, for each of point cloud data, it takes about 6 hours and 3 hours to train a 3-D CNN model for the rough registration and a 2-D CNN model for the precise registration, respectively. Nevertheless, transfer learning can be employed to reduce the training time of CNN models for a new 3-D point cloud data.
After training, 20 unforeseen samples which are generated by randomly rotating and translating, followed by downsampling the model point cloud, according to each of these five data sets including Bunny, Horse, Dragon, Buddha and Armadillo are employed to validate the performance of the trained CNN models as well as two other approaches such as RANSAC+ICP and FGR approach. Table 1 shows a comparison on the MSE and computation time for test samples from 1 to 10 of Bunny data whilst the similar comparison for test samples from 11 to 20 is given Table 2 where n and m stand for the number of points in the model and data point cloud, respectively. When it comes to the MSE, the smaller MSE is, the better registration result is. Generally, it is noticeable that from all the three algorithms, FGR algorithm gives the worst result in terms of MSE value whereas the results of the proposed approach are a little bit worse than RANSAC+ICP method. Particularly, although FGR method shows the smaller computation time compared with RANSAC+ICP method, MSE values according to this method are quite large with three imprecise estimations observed in samples 3, 14 and 15. Besides, the maximum MSE value at approximately 3.418 mm in sample 3 and the minimum MSE at around 2.851 mm in sample 8 experienced in the proposed algorithm shows the lowest disparity, nearly 0.567 mm, in all estimations of 20 test samples while these differences are witnessed at around 1.063 mm and 11.442 mm in accordance with RANSAC+ICP and FGR approach, respectively. This shows that the developed registration architecture possesses the best stable result in estimating the transformation between two 3-D point clouds in the case of Bunny data. In addition to this advantage, obviously, the smallest computation time is also experienced in the proposed algorithm with the largest value at around 0.027 s whilst it takes approximately 0.140 s and 0.028 s (an imprecise estimation) for the fastest estimation in the case of RANSAC+ICP and FGR approach, respectively. Similar to Bunny data, comparable results on two benchmarks including the MSE and computation time for these three algorithms are shown in Table 3, Table 4, Table 5, Table 6, Table 7, Table 8, Table 9 and Table 10 for Horse, Dragon, Buddha and Armadillo data, respectively.
In order to draw thorough comparisons and reinforce an efficient improvement on reducing the computation time but maintaining the high precision when estimating the pose of a 3-D data point cloud of the proposed algorithm, average values in terms of the MSE and computation time of all tested samples for 5 objects are figured out and visualized into two bar charts as shown in Figure 10 and Figure 11. In fact, it is obvious that FGR method witnesses the strongest fluctuation as well as the highest error of MSE values that lead to the worst performance in terms of the precision of this method. Particularly, the average MSE value of 5 tested objects of this approach is approximately two times as high as the average MSE value of the proposed and RANSAC+ICP approach as illustrated in the last column group in Figure 10. In addition, the lowest discrepancy among the average MSE values of 5 tested objects at roughly 1.130 mm, followed by a larger disparity at approximately 1.840 mm according to RANSAC+ICP approach, draws that our CNNs-based algorithm achieves the robust performance for a variety of objects. Moreover, the proposed approach matches the comparable precision with a slightly smaller average MSE value at around 0.108 mm in comparison with RANSAC+ICP algorithm as given in the column group “Average” in Figure 10. As regards the computation time, the presented algorithm significantly outperforms with approximately 15 times and 2 times as fast as RANSAC+ICP and FGR method, respectively, as shown in the last column group in Figure 11.
To conclude these comparisons, registration results of these three algorithms with specific samples are visualized in Figure 12, Figure 13, Figure 14, Figure 15 and Figure 16 for Bunny, Horse, Dragon, Buddha and Armadillo data, respectively. It is noticeable that data point clouds transformed by the proposed and RANSAC+ICP algorithm are closer to the model point clouds than FGR approach. Additionally, as shown in Figure 12b, Figure 13b, Figure 14b, Figure 15b and Figure 16b, the registration result in the precise stage (blue point cloud) is better than the result in the rough stage (green point cloud) because the 2-D CNN model covers a smaller uncertainty from 15 to 15 than the 3-D CNN model with the full range of uncertainty from 180 to 180 . Therefore, based on 3-D point clouds of an object captured by a 3-D camera, the pose of the object can be precisely estimated in fixed time that is sufficient for the application of real-time closed-loop object picking and assembly systems because of two redeeming features such as the high accuracy and fast estimation.

7. Conclusions

This paper resolves the time-consuming problem of 3-D point cloud registration. The estimation time is significantly shortened by using the proposed two-stage CNN architecture, 3-D CNN model and 2-D CNN model. To generate the training data sets, firstly, the model point cloud of a specific object is randomly rotated in the full range from 180 to 180 and the small range from 15 to 15 and then downsampled several times. Then, these two point cloud data sets, according to the full range and the small range of rotation angles, are passed through two descriptors, SSPD and 8CBCP, respectively. Based on these training data sets, CNN models are independently trained for each specific object to update optimal weights with the aim of estimating the relative orientation between the 3-D model point cloud and a 3-D data point cloud. Specifically, to guarantee comparable accuracy with existing algorithms such as RANSAC then ICP approach, the 2-D CNN model is trained on the labeled data set created by the 8CBCP descriptor. Hence, in the training process, this 2-D CNN model can thoroughly learn distinct characteristics of several point clouds with different orientations in the case of the small uncertainty. After training, the proposed two-stage CNN architecture can quickly estimate the transformation between the model point cloud and a data point cloud in which the translation is estimated based on the average values of these two point cloud data sets. The experimental results show that the proposed algorithm can precisely and quickly estimate the relative transformation between the model point cloud and a data point cloud in fixed time. This leads to potential significance and applicability of the proposed algorithm for the purpose of estimating the pose of an object based on 3-D point clouds in automated assembly systems. Additionally, according to the results of this research, 3-D partial point cloud registration employing CNNs appears to be an important direction of future research.

Author Contributions

Conceptualization, W.C.C.; methodology, W.C.C.; writing–original draft preparation, W.C.C. and V.T.P.; writing–review and editing, W.C.C. and V.T.P.

Funding

This research was supported by Ministry of Science and Technology, Taiwan, R.O.C. under grants MOST 107-2221-E-027-112 and MOST 108-2221-E-027-109-MY2.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Huang, S.; Yamakawa, Y.; Senoo, T.; Ishikawa, M. Realizing peg-and-hole alignment with one eye-in-hand high-speed camera. In Proceedings of the 2013 IEEE/ASME International Conference on Advanced Intelligent Mechatronics, Wollongong, Australia, 9–12 July 2013; pp. 1127–1132. [Google Scholar]
  2. Huang, S.; Yamakawa, Y.; Senoo, T.; Ishikawa, M. Dynamic compensation by fusing a high-speed actuator and high-speed visual feedback with its application to fast peg-and-hole alignment. Adv. Robot. 2014, 28, 613–624. [Google Scholar] [CrossRef]
  3. Wang, B.; Jiang, L.; Li, J.; Cai, H.; Liu, H. Grasping unknown objects based on 3d model reconstruction. In Proceedings of the 2005 IEEE/ASME International Conference on Advanced Intelligent Mechatronics, Monterey, CA, USA, 24–28 July 2005; pp. 461–466. [Google Scholar]
  4. Chang, W.C. Robotic assembly of smartphone back shells with eye-in-hand visual servoing. Robot. Comput.-Integr. Manuf. 2018, 50, 102–113. [Google Scholar] [CrossRef]
  5. Savarimuthu, T.R.; Buch, A.G.; Schlette, C.; Wantia, N.; Roßmann, J.; Martínez, D.; Alenyà, G.; Torras, C.; Ude, A.; Nemec, B.; et al. Teaching a robot the semantics of assembly tasks. IEEE Trans. Syst. Man Cybern. Syst. 2018, 48, 670–692. [Google Scholar] [CrossRef]
  6. Chan, K.C.; Koh, C.K.; Lee, C.G. A 3-D-point-cloud system for human-pose estimation. IEEE Trans. Syst. Man Cybern. Syst. 2014, 44, 1486–1497. [Google Scholar] [CrossRef]
  7. Guo, Y.; Bennamoun, M.; Sohel, F.; Lu, M.; Wan, J. An integrated framework for 3-D modeling, object detection, and pose estimation from point-clouds. IEEE Trans. Instrum. Meas. 2015, 64, 683–693. [Google Scholar]
  8. Zhou, Y.; Tuzel, O. Voxelnet: End-to-end learning for point cloud based 3d object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 4490–4499. [Google Scholar]
  9. Qi, C.R.; Su, H.; Mo, K.; Guibas, L.J. Pointnet: Deep learning on point sets for 3d classification and segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 652–660. [Google Scholar]
  10. Zhang, R.; Li, G.; Li, M.; Wang, L. Fusion of images and point clouds for the semantic segmentation of large-scale 3D scenes based on deep learning. ISPRS J. Photogramm. Remote Sens. 2018, 143, 85–96. [Google Scholar] [CrossRef]
  11. Chang, W.C.; Pham, V.T. An efficient neural network with performance-based switching of candidate optimizers for point cloud matching. In Proceedings of the 6th International Conference on Control, Mechatronics and Automation, Tokyo, Japan, 12–14 October 2018; pp. 159–164. [Google Scholar]
  12. Fischler, M.A.; Bolles, R.C. Random sample consensus: A paradigm for model fitting with applications to image analysis and automated cartography. Commun. ACM 1981, 24, 381–395. [Google Scholar] [CrossRef]
  13. Torr, P.H.S.; Davidson, C. IMPSAC: Synthesis of importance sampling and random sample consensus. IEEE Trans. Pattern Anal. Mach. Intell. 2003, 25, 354–364. [Google Scholar] [CrossRef]
  14. Nistér, D. Preemptive RANSAC for live structure and motion estimation. Mach. Vis. Appl. 2005, 16, 321–329. [Google Scholar] [CrossRef] [Green Version]
  15. Xu, M.; Lu, J. Distributed RANSAC for the robust estimation of three-dimensional reconstruction. IET Comput. Vis. 2012, 6, 324–333. [Google Scholar] [CrossRef]
  16. Niedfeldt, P.C.; Beard, R.W. Convergence and complexity analysis of recursive-RANSAC: A new multiple target tracking algorithm. IEEE Trans. Autom. Control 2015, 61, 456–461. [Google Scholar] [CrossRef]
  17. Han, J.; Wang, F.; Guo, Y.; Zhang, C.; He, Y. An improved RANSAC registration algorithm based on region covariance descriptor. In Proceedings of the 2015 Chinese Automation Congress (CAC), Wuhan, China, 27–29 November 2015; pp. 746–751. [Google Scholar]
  18. Aiger, D.; Mitra, N.J.; Cohen-Or, D. 4-points congruent sets for robust pairwise surface registration. In ACM Transactions on Graphics (TOG); ACM: New York, NY, USA, 2008; Volume 27, p. 85. [Google Scholar]
  19. Mellado, N.; Aiger, D.; Mitra, N.J. Super 4pcs fast global pointcloud registration via smart indexing. In Computer Graphics Forum; Wiley Online Library: Hoboken, NJ, USA, 2014; Volume 33, pp. 205–215. [Google Scholar]
  20. Shi, X.; Peng, J.; Li, J.; Yan, P.; Gong, H. The Iterative Closest Point Registration Algorithm Based on the Normal Distribution Transformation. Procedia Comput. Sci. 2019, 147, 181–190. [Google Scholar] [CrossRef]
  21. Liu, S.F.; Liang, J.; Gong, C.Y.; Pai, W.Y. Registration method of point clouds using improved digital image correlation coefficient. Opt. Eng. 2018, 57, 113104. [Google Scholar] [CrossRef]
  22. Chang, W.C.; Wu, C.H. Candidate-based matching of 3-D point clouds with axially switching pose estimation. Vis. Comput. 2019, 1–15. [Google Scholar] [CrossRef]
  23. Besl, P.J.; McKay, N.D. A method for registration of 3-D shapes. IEEE Trans. Pattern Anal. Mach. Intell. 1992, 14, 239–256. [Google Scholar] [CrossRef]
  24. Xie, Z.; Xu, S.; Li, X. A high-accuracy method for fine registration of overlapping point clouds. Image Vis. Comput. 2010, 28, 563–570. [Google Scholar] [CrossRef]
  25. Chen, J.; Wu, X.; Wang, M.Y.; Li, X. 3D shape modeling using a self-developed hand-held 3D laser scanner and an efficient HT-ICP point cloud registration algorithm. Opt. Laser Technol. 2013, 45, 414–423. [Google Scholar] [CrossRef]
  26. Zhu, J.; Jin, C.; Jiang, Z.; Xu, S.; Xu, M.; Pang, S. Robust point cloud registration based on both hard and soft assignments. Opt. Laser Technol. 2019, 110, 202–208. [Google Scholar] [CrossRef]
  27. Chetverikov, D.; Stepanov, D.; Krsek, P. Robust Euclidean alignment of 3D point sets: The trimmed iterative closest point algorithm. Image Vis. Comput. 2005, 23, 299–309. [Google Scholar] [CrossRef]
  28. Phillips, J.M.; Liu, R.; Tomasi, C. Outlier robust ICP for minimizing fractional RMSD. In Proceedings of the Sixth International Conference on 3-D Digital Imaging and Modeling (3DIM 2007), Montreal, QC, Canada, 21–23 August 2007; pp. 427–434. [Google Scholar]
  29. Mavridis, P.; Andreadis, A.; Papaioannou, G. Efficient sparse icp. Comput. Aided Geom. Des. 2015, 35, 16–26. [Google Scholar] [CrossRef]
  30. Yang, J.; Li, H.; Jia, Y. Go-icp: Solving 3d registration efficiently and globally optimally. In Proceedings of the IEEE International Conference on Computer Vision, Sydney, Australia, 1–8 December 2013; pp. 1457–1464. [Google Scholar]
  31. Ying, S.; Peng, J.; Du, S.; Qiao, H. A scale stretch method based on ICP for 3D data registration. IEEE Trans. Autom. Sci. Eng. 2009, 6, 559–565. [Google Scholar] [CrossRef]
  32. Wang, C.; Shu, Q.; Yang, Y.; Yuan, F. Point Cloud Registration in Multidirectional Affine Transformation. IEEE Photonics J. 2018, 10, 1–15. [Google Scholar] [CrossRef]
  33. Basdogan, C.; Oztireli, A.C. A new feature-based method for robust and efficient rigid-body registration of overlapping point clouds. Vis. Comput. 2008, 24, 679–688. [Google Scholar] [CrossRef]
  34. Jiang, J.; Cheng, J.; Chen, X. Registration for 3-D point cloud using angular-invariant feature. Neurocomputing 2009, 72, 3839–3844. [Google Scholar] [CrossRef]
  35. Meng, Y.; Zhang, H. Registration of point clouds using sample-sphere and adaptive distance restriction. Vis. Comput. 2011, 27, 543–553. [Google Scholar] [CrossRef]
  36. He, B.; Lin, Z.; Li, Y.F. An automatic registration algorithm for the scattered point clouds based on the curvature feature. Opt. Laser Technol. 2013, 46, 53–60. [Google Scholar] [CrossRef]
  37. Hana, X.F.; Jin, J.S.; Xie, J.; Wang, M.J.; Jiang, W. A comprehensive review of 3D point cloud descriptors. arXiv 2018, arXiv:1802.02297. [Google Scholar]
  38. Elbaz, G.; Avraham, T.; Fischer, A. 3D point cloud registration for localization using a deep neural network auto-encoder. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4631–4640. [Google Scholar]
  39. Aoki, Y.; Goforth, H.; Srivatsan, R.A.; Lucey, S. PointNetLK: Robust & Efficient Point Cloud Registration using PointNet. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–21 June 2019; pp. 7163–7172. [Google Scholar]
  40. Lucas, B.D.; Kanade, T. An iterative image registration technique with an application to stereo vision. In Proceedings of the 7th International Joint Conference on Artificial Intelligence, Vancouver, BC, Canada, 24–28 August 1981; pp. 674–679. [Google Scholar]
  41. Bergen, G.v.d. Efficient collision detection of complex deformable models using AABB trees. J. Graph. Tools 1997, 2, 1–13. [Google Scholar] [CrossRef]
  42. Turk, G.; Levoy, M. Zippered polygon meshes from range images. In Proceedings of the 21st Annual Conference on Computer Graphics and Interactive Techniques, Orlando, FL, USA, 24–29 July 1994; pp. 311–318. [Google Scholar]
  43. Curless, B.; Levoy, M. A Volumetric Method for Building Complex Models From Range Images. 1996. Available online: http://papers.cumincad.org/cgi-bin/works/Show?2ca3 (accessed on 8 August 2019).
  44. Krishnamurthy, V.; Levoy, M. Fitting smooth surfaces to dense polygon meshes. SIGGRAPH 1996, 96, 313–324. [Google Scholar]
  45. NTU-CSIE. Horse Point Cloud Data. 2005. Available online: https://graphics.cmlab.csie.ntu.edu.tw/~robin/courses/gm05/model/ (accessed on 3 May 2019).
  46. Rusu, R.B.; Blodow, N.; Beetz, M. Fast point feature histograms (FPFH) for 3D registration. In Proceedings of the 2009 IEEE International Conference on Robotics and Automation, Kobe, Japan, 12–17 May 2009; pp. 3212–3217. [Google Scholar]
  47. Zhou, Q.Y.; Park, J.; Koltun, V. Fast global registration. In European Conference on Computer Vision; Springer: Berlin, Germany, 2016; pp. 766–782. [Google Scholar]
  48. Open3D. Global Registration. Available online: http://www.open3d.org/docs/tutorial/Advanced/global_registration.html (accessed on 5 May 2019).
  49. Open3D. Fast Global Registration. Available online: http://www.open3d.org/docs/tutorial/Advanced/fast_global_registration.html (accessed on 5 May 2019).
Figure 1. A block diagram of the proposed registration architecture using convolutional neural networks (CNNs).
Figure 1. A block diagram of the proposed registration architecture using convolutional neural networks (CNNs).
Applsci 09 03273 g001
Figure 2. A simple demonstration of the two-stage estimation.
Figure 2. A simple demonstration of the two-stage estimation.
Applsci 09 03273 g002
Figure 3. A comprehensive block diagram of the proposed registration architecture.
Figure 3. A comprehensive block diagram of the proposed registration architecture.
Applsci 09 03273 g003
Figure 4. A demonstration about the establishment of cube and sub-cubes in Simplified Spatial Point Distribution (SSPD) descriptor.
Figure 4. A demonstration about the establishment of cube and sub-cubes in Simplified Spatial Point Distribution (SSPD) descriptor.
Applsci 09 03273 g004
Figure 5. A demonstration of Axis-Aligned Bounding Box (AABB) divided into 8 sub-cubes.
Figure 5. A demonstration of Axis-Aligned Bounding Box (AABB) divided into 8 sub-cubes.
Applsci 09 03273 g005
Figure 6. A demonstration of the corresponding closest points of each corner.
Figure 6. A demonstration of the corresponding closest points of each corner.
Applsci 09 03273 g006
Figure 7. The corresponding closest points according to each corner: (a) Corner 1, (b) Corner 2, (c) Corner 3, (d) Corner 4, (e) Corner 5, (f) Corner 6, (g) Corner 7, (h) Corner 8.
Figure 7. The corresponding closest points according to each corner: (a) Corner 1, (b) Corner 2, (c) Corner 3, (d) Corner 4, (e) Corner 5, (f) Corner 6, (g) Corner 7, (h) Corner 8.
Applsci 09 03273 g007
Figure 8. A 3-D convolutional neural network (CNN) architecture for the rough registration.
Figure 8. A 3-D convolutional neural network (CNN) architecture for the rough registration.
Applsci 09 03273 g008
Figure 9. A 2-D architecture for the precise registration.
Figure 9. A 2-D architecture for the precise registration.
Applsci 09 03273 g009
Figure 10. The comparison about average mean squared error (MSE) of 20 tested samples.
Figure 10. The comparison about average mean squared error (MSE) of 20 tested samples.
Applsci 09 03273 g010
Figure 11. The comparison about average computation time of 20 tested samples.
Figure 11. The comparison about average computation time of 20 tested samples.
Applsci 09 03273 g011
Figure 12. The registration results of Bunny data for test sample 6: (a) The initial state; (b) 3-D CNNs then 2-D CNNs, (c) RANSAC then ICP, (d) Fast Global Registration.
Figure 12. The registration results of Bunny data for test sample 6: (a) The initial state; (b) 3-D CNNs then 2-D CNNs, (c) RANSAC then ICP, (d) Fast Global Registration.
Applsci 09 03273 g012
Figure 13. The registration results of Horse data for test sample 2: (a) The initial state; (b) 3-D CNNs then 2-D CNNs, (c) RANSAC then ICP, (d) Fast Global Registration.
Figure 13. The registration results of Horse data for test sample 2: (a) The initial state; (b) 3-D CNNs then 2-D CNNs, (c) RANSAC then ICP, (d) Fast Global Registration.
Applsci 09 03273 g013
Figure 14. The registration results of Dragon data for test sample 20: (a) The initial state; (b) 3-D CNNs then 2-D CNNs, (c) RANSAC then ICP, (d) Fast Global Registration.
Figure 14. The registration results of Dragon data for test sample 20: (a) The initial state; (b) 3-D CNNs then 2-D CNNs, (c) RANSAC then ICP, (d) Fast Global Registration.
Applsci 09 03273 g014
Figure 15. The registration results of Buddha data for test sample 7: (a) The initial state; (b) 3-D CNNs then 2-D CNNs, (c) RANSAC then ICP, (d) Fast Global Registration.
Figure 15. The registration results of Buddha data for test sample 7: (a) The initial state; (b) 3-D CNNs then 2-D CNNs, (c) RANSAC then ICP, (d) Fast Global Registration.
Applsci 09 03273 g015
Figure 16. The registration results of Armadillo data for test sample 5: (a) The initial state; (b) 3-D CNNs then 2-D CNNs, (c) RANSAC then ICP, (d) Fast Global Registration.
Figure 16. The registration results of Armadillo data for test sample 5: (a) The initial state; (b) 3-D CNNs then 2-D CNNs, (c) RANSAC then ICP, (d) Fast Global Registration.
Applsci 09 03273 g016
Table 1. The comparisons about mean squared error (MSE) and computation time of “Bunny” data for test samples from 1 to 10.
Table 1. The comparisons about mean squared error (MSE) and computation time of “Bunny” data for test samples from 1 to 10.
Test Sample12345678910
Points (n, m)(1116, 1015)(1116, 990)(1116, 1032)(1116, 1012)(1116, 1001)(1116, 979)(1116, 1025)(1116, 982)(1116, 985)(1116, 1015)
BenchmarkMSETimeMSETimeMSETimeMSETimeMSETimeMSETimeMSETimeMSETimeMSETimeMSETime
3-D CNNs then 2-D CNNs3.2190.0253.1720.0273.4180.0263.1920.0252.9480.0262.9570.0243.1500.0252.8510.0233.0960.0253.0400.025
RANdom SAmple Consensus
then Iterative Closest Point
(RANSAC then ICP)3.1130.3273.0360.2143.7500.1403.1000.1732.9790.2912.9330.4592.8790.4372.8450.4652.9170.4512.8120.411
Fast Global Registration3.2660.0453.2970.03213.4500.0323.4200.0333.0420.0433.5180.0482.8800.0483.2860.0552.9170.0483.0000.070
Table 2. The comparisons about MSE and computation time of “Bunny” data for test samples from 11 to 20.
Table 2. The comparisons about MSE and computation time of “Bunny” data for test samples from 11 to 20.
Test Sample11121314151617181920
Points (n, m)(1116, 1005)(1116, 1002)(1116, 973)(1116, 1001)(1116, 1011)(1116, 1004)(1116, 1016)(1116, 997)(1116, 993)(1116, 993)
BenchmarkMSETimeMSETimeMSETimeMSETimeMSETimeMSETimeMSETimeMSETimeMSETimeMSETime
3-D CNNs then 2-D CNNs3.0450.0242.9570.0252.9340.0253.1410.0263.0480.0252.9890.0273.1830.0252.9740.0243.1550.0243.0570.024
RANSAC then ICP2.9040.6453.3570.1282.8750.2883.8750.1183.5220.2212.8670.3072.9230.2772.8890.3922.9760.1452.8270.255
Fast Global Registration3.1350.0763.4590.0312.8830.04614.3220.02812.7040.0303.0500.0372.9410.0402.9140.0493.1060.0323.1240.039
Table 3. The comparisons about MSE and computation time of “Horse” data for test samples from 1 to 10.
Table 3. The comparisons about MSE and computation time of “Horse” data for test samples from 1 to 10.
Test Sample12345678910
Points (n, m)(1009, 971)(1009, 962)(1009, 962)(1009, 948)(1009, 988)(1009, 952)(1009, 957)(1009, 964)(1009, 943)(1009, 960)
BenchmarkMSETimeMSETimeMSETimeMSETimeMSETimeMSETimeMSETimeMSETimeMSETimeMSETime
3-D CNNs then 2-D CNNs2.5590.0212.5730.0222.7650.0222.3300.0222.4190.0222.4670.0222.3330.0222.9260.0222.4610.0222.5280.022
RANSAC then ICP2.3320.7843.3650.5272.6360.4652.3810.3372.4570.4222.4880.1832.3100.7952.3740.7492.4900.4522.5530.271
Fast Global Registration2.4180.13110.1380.0493.5370.0413.7790.0362.6120.0474.9170.0332.4200.1122.4710.1263.5670.0354.3450.031
Table 4. The comparisons about MSE and computation time of “Horse” data for test samples from 11 to 20.
Table 4. The comparisons about MSE and computation time of “Horse” data for test samples from 11 to 20.
Test Sample11121314151617181920
Points (n, m)(1009, 965)(1009, 965)(1009, 962)(1009, 968)(1009, 984)(1009, 964)(1009, 963)(1009, 976)(1009, 976)(1009, 970)
BenchmarkMSETimeMSETimeMSETimeMSETimeMSETimeMSETimeMSETimeMSETimeMSETimeMSETime
3-D CNNs then 2-D CNNs2.6470.0222.6590.0222.4440.0222.5880.0222.3540.0234.5650.0222.5200.0222.6070.0222.4680.0232.5130.022
RANSAC then ICP2.3630.7042.3850.9842.4390.6922.4850.4172.5440.2412.4760.3242.3690.7542.9940.2962.5090.2902.3510.651
Fast Global Registration2.5770.1252.4960.0732.6950.0534.7540.0388.1230.0333.3110.0332.4630.1277.7560.0385.0010.0333.1480.051
Table 5. The comparisons about MSE and computation time of “Dragon” data for test samples from 1 to 10.
Table 5. The comparisons about MSE and computation time of “Dragon” data for test samples from 1 to 10.
Test Sample12345678910
Points (n, m)(1041, 990)(1041, 1020)(1041, 1010)(1041, 1011)(1041, 1014)(1041, 992)(1041, 973)(1041, 992)(1041, 995)(1041, 1003)
BenchmarkMSETimeMSETimeMSETimeMSETimeMSETimeMSETimeMSETimeMSETimeMSETimeMSETime
3-D CNNs then 2-D CNNs3.3500.0213.5650.0233.4620.0223.4250.0223.3200.0243.4400.0223.6620.0223.4590.0223.4300.0213.4370.023
RANSAC then ICP6.6080.1274.2010.1053.4890.2883.8370.1185.0920.1044.2740.1074.3500.1153.2930.3653.2640.8273.6150.110
Fast Global Registration12.7700.04211.0110.0403.6940.0444.6930.04019.8070.04221.2540.04012.6170.0393.3950.0633.2770.1045.3370.039
Table 6. The comparisons about MSE and computation time of “Dragon” data for test samples from 11 to 20.
Table 6. The comparisons about MSE and computation time of “Dragon” data for test samples from 11 to 20.
Test Sample11121314151617181920
Points (n, m)(1041, 999)(1041, 998)(1041, 1010)(1041, 978)(1041, 970)(1041, 984)(1041, 997)(1041, 1012)(1041, 1015)(1041, 1005)
BenchmarkMSETimeMSETimeMSETimeMSETimeMSETimeMSETimeMSETimeMSETimeMSETimeMSETime
3-D CNNs then 2-D CNNs3.6110.0223.4710.0223.4270.0223.3320.0249.6430.0223.3480.0213.4310.0223.3680.0213.5680.0233.5640.022
RANSAC then ICP4.3740.1285.8020.1073.5760.1114.4740.1223.8890.1223.8230.1054.2320.1204.9920.1106.3520.1103.5580.190
Fast Global Registration4.8950.04113.3400.0393.9140.04312.7170.03919.9660.04121.1430.0384.3640.04010.0310.0409.2870.0403.5430.043
Table 7. The comparisons about MSE and computation time of “Buddha” data for test samples from 1 to 10.
Table 7. The comparisons about MSE and computation time of “Buddha” data for test samples from 1 to 10.
Test Sample12345678910
Points (n, m)(1045, 963)(1045, 926)(1045, 975)(1045, 984)(1045, 986)(1045, 965)(1045, 956)(1045, 935)(1045, 949)(1045, 941)
BenchmarkMSETimeMSETimeMSETimeMSETimeMSETimeMSETimeMSETimeMSETimeMSETimeMSETime
3-D CNNs then 2-D CNNs2.8360.0253.0720.0242.7950.0242.8980.0233.1890.0242.7760.0242.8710.0242.8150.0242.7940.0243.1920.023
RANSAC then ICP2.8800.3293.1880.3982.7730.7843.2030.2692.7720.3362.8870.4593.1070.2713.9310.2582.9210.3783.5630.300
Fast Global Registration2.9240.0486.4340.0453.1670.0886.6190.0403.4870.0474.0400.0543.2990.0448.4940.0444.2520.0539.4960.055
Table 8. The comparisons about MSE and computation time of “Buddha” data for test samples from 11 to 20.
Table 8. The comparisons about MSE and computation time of “Buddha” data for test samples from 11 to 20.
Test Sample11121314151617181920
Points (n, m)(1045, 963) (1045, 989)(1045, 934)(1045, 973)(1045, 931)(1045, 976)(1045, 975)(1045, 934)(1045, 1007)
BenchmarkMSETimeMSETimeMSETimeMSETimeMSETimeMSETimeMSETimeMSETimeMSETimeMSETime
3-D CNNs then 2-D CNNs3.1400.0252.9610.0242.8120.0242.8080.0232.9280.0242.8480.0232.8680.0252.8740.0252.8470.0242.9430.025
RANSAC then ICP2.8080.4652.7610.5252.8760.2702.7390.3053.4690.2632.7960.6653.3340.4002.8670.5423.0660.1632.9310.263
Fast Global Registration3.1520.0463.2050.0534.2830.0423.8750.0486.8270.0424.0940.0627.9330.0433.3550.0605.2910.0384.7200.041
Table 9. The comparisons about MSE and computation time of “Armadillo” data for test samples from 1 to 10.
Table 9. The comparisons about MSE and computation time of “Armadillo” data for test samples from 1 to 10.
Test Sample12345678910
Points (n, m)(1046, 984)(1046, 969)(1046, 937)(1046, 966)(1046, 961)(1046, 939)(1046, 958)(1046, 968)(1046, 949)(1046, 952)
BenchmarkMSETimeMSETimeMSETimeMSETimeMSETimeMSETimeMSETimeMSETimeMSETimeMSETime
3-D CNNs then 2-D CNNs2.6320.0252.7380.0242.6890.0242.5070.0242.4120.0232.8990.0252.4790.0242.7330.0262.7840.0242.6780.024
RANSAC then ICP2.3760.6583.0560.7412.4730.3572.6310.2612.4720.5622.3810.3652.3400.9243.2260.7432.3420.5433.3080.704
Fast Global Registration2.4040.13110.0750.0912.8870.04010.4150.0393.3680.0405.4430.0442.3870.1299.9980.0862.9430.0549.8240.106
Table 10. The comparisons about MSE and computation time of “Armadillo” data for test samples from 11 to 20.
Table 10. The comparisons about MSE and computation time of “Armadillo” data for test samples from 11 to 20.
Test Sample11121314151617181920
Points (n, m)(1046, 966)(1046, 977)(1046, 955)(1046, 978)(1046, 963)(1046, 963)(1046, 956)(1046, 949)(1046, 965)(1046, 974)
BenchmarkMSETimeMSETimeMSETimeMSETimeMSETimeMSETimeMSETimeMSETimeMSETimeMSETime
3-D CNNs then 2-D CNNs2.8620.0242.9000.0242.7130.0242.6350.0262.8870.0242.7860.0242.6710.0232.9540.0242.7800.0242.6380.026
RANSAC then ICP2.3920.5142.3670.6313.1620.3282.5510.8082.5540.1983.2810.2602.4440.5032.4710.2302.6990.2062.7880.388
Fast Global Registration3.1560.0422.5220.1309.7960.0399.8750.11512.5660.03810.0990.0373.6370.0406.8060.03712.1010.03410.9380.040

Share and Cite

MDPI and ACS Style

Chang, W.-C.; Pham, V.-T. 3-D Point Cloud Registration Using Convolutional Neural Networks. Appl. Sci. 2019, 9, 3273. https://doi.org/10.3390/app9163273

AMA Style

Chang W-C, Pham V-T. 3-D Point Cloud Registration Using Convolutional Neural Networks. Applied Sciences. 2019; 9(16):3273. https://doi.org/10.3390/app9163273

Chicago/Turabian Style

Chang, Wen-Chung, and Van-Toan Pham. 2019. "3-D Point Cloud Registration Using Convolutional Neural Networks" Applied Sciences 9, no. 16: 3273. https://doi.org/10.3390/app9163273

APA Style

Chang, W. -C., & Pham, V. -T. (2019). 3-D Point Cloud Registration Using Convolutional Neural Networks. Applied Sciences, 9(16), 3273. https://doi.org/10.3390/app9163273

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop