Next Article in Journal
USV-Observed Turbulent Heat Flux Induced by Late Spring Cold Dry Air Incursion over Sub-Mesoscale Warm Regions off Sanriku, Japan
Next Article in Special Issue
Robust Mesh Segmentation Using Feature-Aware Region Fusion
Previous Article in Journal
Terahertz and Millimeter Wave Sensing and Applications
Previous Article in Special Issue
Quality Assessment of a Novel Camera-Based Measurement System for Roughness Determination of Concrete Surfaces—Accuracy Evaluation and Validation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Feature Consistent Point Cloud Registration in Building Information Modeling

1
School of Computer Science and Engineering, Nanjing University of Science and Technology, Nanjing 210000, China
2
Panasonic R&D Center Singapore, Singapore 469332, Singapore
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(24), 9694; https://doi.org/10.3390/s22249694
Submission received: 29 September 2022 / Revised: 28 October 2022 / Accepted: 3 November 2022 / Published: 10 December 2022
(This article belongs to the Special Issue Short-Range Optical 3D Scanning and 3D Data Processing)

Abstract

:
Point Cloud Registration contributes a lot to measuring, monitoring, and simulating in building information modeling (BIM). In BIM applications, the robustness and generalization of point cloud features are particularly important due to the huge differences in sampling environments. We notice two possible factors that may lead to poor generalization, the normal ambiguity of boundaries on hard edges leading to less accuracy in transformation; and the fact that existing methods focus on spatial transformation accuracy, leaving the advantages of feature matching unaddressed. In this work, we propose a boundary-encouraging local frame reference, the P y r a m i d F e a t u r e ( P M D ) , consisting of point-level, line-level, and mesh-level information to extract a more generalizing and continuous point cloud feature to encourage the knowledge of boundaries to overcome the normal ambiguity. Furthermore, instead of registration guided by spatial transformation accuracy alone, we suggest another supervision to extract consistent hybrid features. A large number of experiments have demonstrated the superiority of our PyramidNet (PMDNet), especially when the training (ModelNet40) and testing (BIM) sets are very different, PMDNet still achieves very high scalability.

1. Introduction

Optical 3D scanning shows a growing trend in the construction industry, providing a complete and consistent building engineering database by establishing a virtual building 3D model based on digital technology. Nowadays, Building Information Modeling (BIM), as a key application of optical 3D scanning, dominates project follow-up, building monitoring, and maintenance [1,2,3,4], architecture planning [5,6,7,8], emergency simulation [9,10,11,12], and IoT [13,14,15,16]. Implementing BIM methods helps improve the efficiency and integration of constructions in all stages of their life-cycle, as well as helps provide a platform for engineering information exchange and communication [17,18,19,20].
This work is focused on point cloud registration in building information modeling. In practice, one of the key tasks to solve is how to align pre-built 3D models against the scanned points of real buildings. The aligning results help generate complete and accurate digital models, contributing largely to measuring and applications such as simultaneous localization and mapping (SLAM), 3D reconstruction [21,22,23,24], localization [25,26,27,28], and pose estimation [29,30,31,32].
Existing point cloud registration methods have achieved state-of-the-art performance on common datasets. However, most of them fail in BIM scenarios due to various challenges, noise, and data distribution from varying architectural styles, to name a few. Classic methods [33,34,35,36,37] search for hard correspondence which shows little robustness against noise since Euclidean distance is quite sensitive to offset. Feature-based methods [38,39,40,41] extract local or global reference frames in higher dimensional space to achieve registration. Most of the proposed features take not only point-level information but also line structures, greatly improving the robustness. However, feature-based methods face ambiguity during calculating the normals of boundaries on hard edges, leading to inexplicit representation. Learning-based methods [42,43,44,45,46], however, stand promising because they learn features for establishing correspondence and utilize deep neural networks to reduce the calculations. The only problem is whether they generalize well enough for different clouds.
According to 2D image theory, regions of richer geometric or semantic information can produce more discriminant knowledge. In image classification, more attention is paid to key point detection (SIFT [47], SURF [48], ORB [49]). Likewise, researchers have proposed similar works in 3D point cloud processing (USIP [50], SDK [51]). However, such key points in the cloud tend to face ambiguity during calculating their normals because of the selected neighbors located on different planes.
Motivated by the issues aforementioned, instead of using point-level information alone, this work suggests learning combined geometric knowledge of point, line, and mesh levels to alleviate the impact of ambiguous normals, and hence, improve the generalization of point cloud features. To be specific, we define a cone within the neighbors of a given centroid and calculate the three angles near its apex to form a descriptor representing its local geometry. In this case, a trade-off is made between ambiguous point normals and explicit cone angles.
Existing learning-based works evaluate their losses regarding spatial registration accuracy, while we suggest feature-matching precision of equal importance for better feature extraction in various transformations. We introduce another loss, which calculates the distances between two high-dimensional feature descriptors.
Major contributions of this article are:
1
Introduce a boundary-encouraging point cloud feature, P M D , to represent local geometry with higher generalization for registration, as well as solve the normal ambiguity problem.
2
Introduce feature matching loss to the feature extractor to produce consistent hybrid representation.
3
Our PMDNet shows state-of-the-art performance and higher generalization on samples from different distributions. Moreover, high performance still can be observed even when the clouds become more sparse as the distance increases.

2. Related Work

2.1. Classic Registration Methods

Among classic algorithms (Figure 1a)), ICP [33] and its variants [34,35] try to search the correspondences to minimize the distance (usually Euclidean distance) loss between the projections and the destination points or planes. However, the optimization tends to fall into local optimum because the objective function has multiple local extremums [52], especially in BIM scenarios which usually consist of millions of points. Thus, these methods usually include both coarse and fine registration, where the former aims to provide a better initialization for the latter. To enhance the robustness against noise and outliers, the following researchers view registration as a probabilistic distribution problem, where the input point clouds are treated as two distributions from the same probabilistic model [36,37]. Despite saving efforts to establish the correspondences, they still require a proper initialization because of their non-convex loss function.

2.2. Feature-Based Registration Methods

Instead of establishing Euclidean-based correspondences, feature-based methods extract the local reference frame (LRF) for each point in the input clouds to form feature-based correspondences. These descriptors must be distinct from each other, invariant to transformation, and robust to noise and outliers. Despite such high demands it takes, researchers proposed unique descriptors. PFH [38] calculates the invariant pose of a centroid and its neighbors in high dimensional space. FPFH [39] improves PFH with time efficiency by reducing the dimension of histograms but preserves the local feature. SHOT [40] instead focuses on normals by encoding the normal histogram in different coordinates, generated by concatenating all the local histograms. Furthermore, using Hough voting, PPF [41] calculates the 6-D pose of an LRF in a centroid-off manner. It ignores all the global coordinates and only consists of related information of normals, translation, and angles. However, mesh structures, which outmatch points and lines regarding robustness and continuity, are barely included.

2.3. Learning-Based Registration Methods

Early methods estimate a good initial transformation for the ICP baseline [38,53,54]. Nevertheless, recent works utilize deep neural networks to calculate a global or local reference frame for each point and then iteratively solve the transformation (Figure 1b) [55,56] firstly introduced the CNNs to the point cloud tasks, followed by numerous deep learning methods to achieve registration [42,43,44]. Among those, PointNetLk [45] uses PointNet to calculate the global features of input clouds and minimize the distance between them. DCP [46] tries to search soft correspondence with the transformer and solve the registration by SVD. These works achieved state-of-the-art performance on ModelNet and other common datasets. However, being specific to objects, they fail to generalize. On another, we notice an ambiguity (see Section 3.1) during the calculation of point normals.

2.4. Registration in BIM

In architecture, registration is usually performed via multiple approaches (between images, images and clouds, or clouds). Due to the huge number of points, many methods learn planar geometry instead of local reference frame from regular structures. Turning to planar geometries helps reduce the calculation and improves the overall accuracy due to declined influence of outliers. Taking plane-based registration as an example, plane structures are extracted from both input clouds via RANSAC-based methods [57,58,59], Hough transform [60], or clustering [61]. Afterward, correspondence is estimated using the extracted planes, whose accuracy largely depends on the planar segments and their normals. Ref. [62] homogenizes the as-built and as-planned models by extracting similar cross-sections and thus solves the registration problem, suggesting that the geometric shape of a room, if captured accurately, can be a distinguishing feature. Similarly, ref. [63] proposes a four-degree of freedom (DOF) registration problem and then decomposes it into two steps: (1) horizontal alignment achieved by matching the source and target ortho-projected images using the 2D line features and (2) vertical alignment achieved by making the height of the floor and ceiling in the source and target points equivalent. Ref. [64] segments the input clouds into plane pieces and clusters the parallel ones to eventually determine a transform matrix. However, considering geometries are extremely sensitive to density, their performance falls rapidly as the cloud density decreases because it is hard to extract the planar information on sparse clouds.

3. Feature Consistent Registration

Due to the normal ambiguity, hand-crafted features fail to calculate the explicit normal of boundaries on hard edges, which provides more information for registration than other points. We solve this problem by introducing another local reference frame, P M D , which learns point and mesh structures in the meantime. Also, previous works pay more attention to spatial registration accuracy, leaving behind feature-matching precision, which may help improve the representation in high dimensions.
We propose PMDNet to solve the issues mentioned above, including a feature extractor that generates a hybrid P M D feature, a parameter network to establish soft correspondence, and a solver to estimate transformation.
More specifically, PMDNet learns a soft correspondence between input clouds via hybrid P M D feature distances in an iterative way. The source cloud is transformed during each iteration i by the estimated transformation in the last iteration i 1 . Then, hybrid features are extracted from both the source and reference cloud. In the meantime, PMDNet uses a parameter network to learn the annealing parameters, α , β , to refine the soft correspondence. The transformation is generated by Sinkhorn [65] algorithm. Finally, we calculate two losses in the current iteration, L f e and L t r , and back-propagate them to the feature extractor and the transformation estimator. Figure 2 illustrates the pipeline of the PMDNet. We use RPMNet [66] as the backbone.

3.1. P M D Feature: Local Reference Frame to Encourage Boundaries

Given two sets of points, X = { x j | j = 1 , , J } R J × 3 serving as s r c and Y = { y k | k = 1 , , K } R K × 3 denoted as r e f . The objective is to estimate a rigid transformation G ^ = G ^ ( R ^ , t ^ ) S E ( 3 ) with R ^ S O ( 3 ) and t ^ R 3 , such that:
R ^ , t ^ = arg   min R * S O ( 3 ) , t * R 3 | | X × R * + t * Y | | 2
Normal Ambiguity. Given a centroid point p, the normal of p is calculated within a local reference frame consisting of p and its neighbors p 1 , p 2 , , p n . As Figure 3 shows, mesh Ω 1 , Ω 2 with normals n 1 , n 2 intersect at p. Here comes the ambiguity; p seems to have multiple normals, depending on which mesh its neighbors are selected.
To fix normal ambiguity, we suggest refining the P P F feature with angles of intersected meshes. P P F originally are 4D point pair features describing the surface between the centroid point x c and each neighbor x i in a rotation-invariant manner:
P P F ( x c , x i ) = ( ( n c , Δ x c , i ) , ( n i , Δ x c , i ) , ( n c , n i ) , | | Δ x c , i | | 2 )
where n c , n i are the normals of centroid point x c and its neighbors x i . Hereby we introduce another three components to P P F , that is:
P M D ( x c , x i ) = ( ( n c , Δ x c , i ) , ( n i , Δ x c , i ) , ( n c , n i ) , | | Δ x c , i | | 2 , x c ) ( x c ) = ( ( Ω ( x 1 , x 2 , x c ) , Ω ( x 1 , x 3 , x c ) ) , ( Ω ( x 1 , x 2 , x c ) , Ω ( x 2 , x 3 , x c ) ) , ( Ω ( x 2 , x 3 , x c ) , Ω ( x 1 , x 3 , x c ) ) )
where x 1 , x 2 , x 3 are the closest neighbors of x c , and Ω ( p 1 , p 2 , p 3 ) denotes a mesh consisting of three points, p 1 , p 2 , p 3 . Together, x c , x 1 , x 2 , x 3 form a triangular cone where x c is the apex, and x 1 , x 2 , x 3 locate on the bottom surface. The three newly added components describe the angle of the apex. P M D solves normal ambiguity because it considers not only the angle of points, but also the angle of intersected meshes.
However, the selected points x 1 , x 2 , x 3 do not always belong to an existing mesh. Thus, we suggest that the introduced angles, x c in P M D , should be considered along with ( n i , Δ x c , i ) , ( n c , n i ) in P P F , so as to make a trade-off between them. In this case, P M D appears to be more generalizing than the P P F or other features in that the aforementioned five angles correct each other. As is proposed in RPMNet, P M D is concatenated with x c and Δ x i , forming a 13D descriptor to capture both global and local features.

3.2. L f e : Feature-Aware Loss towards Feature Consistency

Deep learning methods mainly consist of a couple of modules, including a feat extractor and an SVD for computing transformation. Previous works compute the loss regarding SVD, ignoring the extractor, leading to in-explicit connection information between them. RPMNet [66] uses a Euclidean distance loss and a second loss to encourage the inliers. Ref. [67] implements Cross-entropy between estimated and ground-truth correspondence. PointNetLk [45] calculates the loss with estimated transformation G ^ and ground-truth G G T . In these cases, the gradient may remain still when back-propagated to the feature extractor, leading to poor updates.
RPMNet defines the total loss L t r as the weighted sum of the Euclidean distance L e l between the estimated and reference clouds, and a second loss L i n l i e r to encourage inliers.
L e l = 1 N i = 1 N | | P × R ^ + t ^ P × R G T t G T | | 2
L i n l i e r = 1 J j J k K m j , k 1 K k K j J m j , k
L t r = L e l + λ i n l i e r L i n l i e r
However, the L t r is not feature-aware. Thus, we introduce another loss, L f e (Equation (7)), to calculate the difference between two features. L f e is transformation-free and only works in feature extractor:
L f e = 1 J j = 1 J Γ ( F s r c , j , G G T ) Γ ( F s r c , j , G ^ )
where F is the extracted 13D feature, Γ ( F , G ) transforms the input F with a given transformation G. ⊕ calculates the difference between two features.
The overall loss is now the weighted sum of L f e and L t r :
L t o t a l = ( 1 λ f e ) L t r + λ f e L f e

3.3. Annealing Parameter Network

Instead of hard correspondence, we use soft correspondence to predict transforms:
M J × K = { m j , k } = { e β ( | | F x j F y k | | 2 α ) }
subject to (1) k = 1 K M j , k 1 , j ; (2) j = 1 J M j , k 1 , k ; (3) arg   max i * M j 1 , i * arg   max i * M j 2 , i * , j 1 j 2 . Where:
  • F is the hybrid feature in high dimensional space generated by the extractor.
  • α serves as a threshold to preserve inliers and punish outliers.
  • β is an annealing parameter to ensure convergence.
Considering α , β are usually distinct on various datasets, RPMNet uses a parameter network that takes both source and reference point clouds as input to predict α , β in an iterative manner. To be specific, the two clouds are concatenated into P J + K , 3 . After that, another column in which all the elements are either 0 or 1 is added to P . Finally, P is fed to a PointNet baseline with softmax activation in the final layer to ensure the predicted parameters are always positive.

4. Experiments

We use ModelNet40 [68] as a common registration problem to train our PMDNet. It contains 12,311 samples of 40 categories. Figure 4 illustrates part of its samples. The clouds used for PMDNet are furthermore subsampled to 512 points for reducing the calculation. We conducted extensive experiments to evaluate the detailed performance of PMDNet against other methods.
In each experiment, we follow approximately the same process. Firstly, the raw cloud of 2048 points is input. Then, a rigid transformation matrix M S E ( 3 ) is generated with random rotation between [ 0 , 45 ] and random translation between [ 0.5 , 0.5 ] about each axis. Afterward, a copy of the raw serves as the reference cloud (Y), and another copy is used as the source (X), which will be transformed afterward by M. Both X and Y are furthermore shuffled and subsampled randomly to 512 points. Finally, The source (X) and reference (Y) clouds are input to PMDNet to get the estimated transformation G ^ = G ^ ( R ^ , t ^ ) . which would be evaluated against G G T = G G T ( R G T , t G T ) = M 1 using the metrics aforementioned. Figure 5 illustrates the whole process.
The parameters of our PMDNet are listed in Table 1.
All the competing methods are evaluated using their pre-trained models on ModelNet40.

4.1. Metrics

All the metrics are based on the rotation and translation errors:
E r r o r ( R ) = ( R G T 1 R ^ ) , E r r o r ( t ) = | | t G T t ^ | | 2
where { R G T , R ^ } and { t G T , t ^ } denote the ground-truth and estimated rotation and translation, respectively. ( A ) = arg   cos ( t r ( A ) 1 2 ) returns the angle of rotation matrix A. We provide both mean square error (MSE) and mean absolute error (MAE) for consistency with previous works [46]. All the metrics are listed in Table 2. C h a m f e r D i s t a n c e ( C D ) is universally used in registration problems, however, it is extremely sensitive to outliers considering C D i n l i e r 0 and C D o u t l i e r 0 . Hence, we clip this distance with a threshold of d = 0.1 for mitigation. In addition, considering that cloud computing technologies are nowadays widely used in many fields but are often constrained by resources [69], time efficiency is also an essential metric to evaluate the likelihood of a model being deployed to mobile devices.
E R M = 1 J j J | E r r o r ( R ) j |
E T M = 1 J j J | E r r o r ( t ) j |
C D ( X , Y ) = 1 | X | x X min y Y | | x y | | 2 + 1 | Y | y Y min x X | | x y | | 2
C C D ( X , Y ) = x X min ( min y Y ( | | x y | | 2 ) , d ) + y Y min ( min x X ( | | x y | | 2 ) , d )

4.2. Ablation Experiment

In this subsection, we compare the contribution of different components of our P M D feature to determine the best choice in afterward experiments. According to its definition, we set up five controlled groups with various components to determine their influence on the overall performance of clean, noise, and unseen data.
Table 3 illustrates all the detailed results. Most groups show bare differences in noise scenarios, however, their performance on clean and unseen vary largely from one to another. P M D N e t A achieves the best of all, reaching up to over 90% R e c a l l , outperforming P M D N e t E (which is the same definition with P P F ) by over 10% R e c a l l on clean and unseen. Comparing P M D N e t B and P M D N e t C , it is easy to find that, x c alone fails to improve the representation of local geometry. x c has worsened the learning, leading to an obvious decline in accuracy. It goes the same with ( n i , Δ x c , i ) , ( n c , n i ) despite ( n i , Δ x c , i ) , ( n c , n i ) only lead to a slight decline on clean and unseen. Only when ( n i , Δ x c , i ) , ( n c , n i ) and x c are both included at the same time, can they together improve the registration greatly. In previous sections, we have made a reasonable explanation for this when introducing the P M D (Section 3.1). We suggest the added angles in P M D , x c , and the original ones in P P F , ( n i , Δ x c , i ) , ( n c , n i ) be considered together, so they can compensate each other to achieve higher performance and generalization. Because x c only works for those boundaries on multiple hard edges where ( n i , Δ x c , i ) , ( n c , n i ) fails to calculate an explicit normal. This explains the huge gap between P M D N e t A and P M D N e t B , P M D N e t C , P M D N e t s E .
Table 3. Ablation results. P M D N e t A , P M D N e t B , P M D N e t C , P M D N e t D , P M D N e t E is the same as introduced in Table 4. Bold and underline denote best and second best performance.
Table 3. Ablation results. P M D N e t A , P M D N e t B , P M D N e t C , P M D N e t D , P M D N e t E is the same as introduced in Table 4. Bold and underline denote best and second best performance.
IDSceneMAE(R)↓MAE(T)↓CCD( 1 e 3 )↓Recall↑ (1.0, 0.1)Recall↑ (0.1, 0.01)
P M D N e t A Clean0.04670.000390.00322699.83%91.76%
P M D N e t B Clean0.16800.001160.01948798.50%44.67%
P M D N e t C Clean0.09390.000660.00732999.25%76.70%
P M D N e t D Clean0.37730.002160.06277195.59%45.25%
P M D N e t E Clean0.10260.000750.00627099.25%76.45%
P M D N e t A Noise1.12010.009900.84058980.28%1.33%
P M D N e t B Noise1.15710.010000.84006181.19%1.16%
P M D N e t C Noise1.19570.010400.86941379.36%0.74%
P M D N e t D Noise1.41990.011900.99825662.14%0.83%
P M D N e t E Noise1.13650.010100.85664680.61%1.49%
P M D N e t A Unseen0.04230.000370.003137100.00%92.02%
P M D N e t B Unseen0.16420.001170.02260898.65%44.70%
P M D N e t C Unseen0.08580.000630.00671799.68%78.27%
P M D N e t D Unseen0.19470.001480.04571697.70%49.28%
P M D N e t E Unseen0.08030.000580.00505499.52%80.64%
In the following experiments, we keep the components in P M D N e t A since it outperforms other groups. All the following P M D N e t refers to P M D N e t A .

4.3. Registration Comparing on ModelNet40

4.3.1. Generalization Capability

First, we provide the performance on clean data of each method in Table 5, along with the qualitative results of our method in Figure 6. In this case, all the methods are trained, tested, and evaluated on the whole ModelNet40 dataset.
We can see DeepGMR and FGR achieve the best performance on clean data, outmatching PMDNet and RPMNet by approximately 1%, which means the differences between DeepGMR, FGR, RPMNet, and our PMDNet are quite bare. Actually, except for ICP and IDAM, the remaining methods roughly reach the same accuracy and R e c a l l , proving their performance on basic clean data.
Then, we evaluate the performance of unseen data, to test the generalization capability of each competing method. The training set only consists of the first twenty categories of ModelNet40, and all the competing methods are evaluated on the remaining twenty categories. This experiment is quite challenging for the generalization of point cloud features because all the data used in the evaluation is never seen during training. The more generalizing the feature is, the more promising results it outputs.
Table 6 shows the comparison of all the candidate methods. Our method achieves the best performance and greatly outmatches the second place, RPMNet, and other methods, reaching a 100% R e c a l l . DeepGMR, FGR, and IDAM show a large decline (over 70% to less than 10%) here compared with their performance on clean data. Figure 7 illustrates the qualitative results of our method. Be advised that, ICP is not listed here since it is not a learning-based method.
Generally, our method achieves state-of-the-art performance both on unseen categories and clean data.

4.3.2. Gaussian Noise

In this experiment, we evaluate the robustness of each competing method in presence of Gaussian noise. After subsampled, each pair of inputs, the source cloud, and the reference cloud is randomly noised with N ( 0 , 0.01 ) and clipped to [ 0.05 , 0.05 ] to prevent extreme outliers, respectively. In this case, the one-on-one correspondence in dense clouds is corrupted due to the noise. That’s why we have been using sparse clouds from the beginning.
Table 7 illustrates the results. RPMNet and DCP v2 are the top 2 best methods in this experiment, reaching less than 1deg loss and over 90% R e c a l l . Taking ICP (6.5 M A E ( R ) , 0.05 M A E ( t ) , 77% R e c a l l ) as a baseline, we divide the methods into two categories, those worse than ICP (DeepGMR, FGR, and IDAM) and those better than ICP (RPMNet, DCV v2, and PMDNet). Despite being less accurate than DCP v2 or RPMNet, PMDNet still achieves acceptable results, reaching 80% R e c a l l .
We attribute the vulnerability of PMDNet to our P M D feature. It solves the normal ambiguity of boundaries by introducing mesh angles. However, noise causes the coordinates of points to shift—on the one hand, the original boundaries are off the surrounding meshes; and on the other hand, points that are originally not boundaries become the boundaries.

4.4. Registration Comparing on BIM Scenarios

In this section, we test the performance of each competing method on BIM scenarios, more specifically, on clouds of uniform density and clouds with varying density, using their pre-trained models. We select 30 CAD models(.dwg) and uniformly subsample a cloud(.ply) on each of them, consisting of points varying from 0.1 million to 1.0 million. During evaluating, the ground-truth rotation R G T and translation t G T is fixed to [ 45 , 45 , 45 ] and [ 1 , 1 , 1 ] , respectively. Other settings are still the same as introduced above.

4.4.1. Clouds of Uniform Density

In this experiment, to test the basic performance on BIM scenarios, only the dataset is changed from ModelNet40 to BIM clouds, the density of which is left unchanged.
Table 8 illustrates the results of each competing method. PMDNet outperforms any other method. Compared with others, the error of PMDNet is only about 20% of that of others, and the CCD is even one ten thousand of that of the second best. Meanwhile, It is worth noting that PMDNet has once again achieved a 100% R e c a l l , twice as high as that of DCP v2. Figure 8 visually illustrates the input and output of PMDNet.
We perform another robust experiment to examine the robustness of candidate methods in noisy BIM scenarios. In this section, a random noise from N ( 0.01 , 0.01 ) is added and clipped to [–0.05, 0.05] like aforementioned.
Table 9 illustrates the results of each competing method. PMDNet fails on exactly three samples, which leads to high M A E ( R ) and E R M . Despite the three failed samples, the overall performance of PMDNet is still of top class since the R e c a l l is 90%.

4.4.2. Clouds with Varying Density

Our PMDNet achieves great performance on density-uniform scenarios. However, scan in real BIM practice shows a tendency to fade as the distance increases, which leads to sparse points. In this section, a declining density is applied to simulate the real scans, besides which the other settings remain the same as above. Also, we perform experiments on both clean and noisy datasets.
Density Sampling & Noise. For each cloud P R M × 3 , we select a certain axis of X Y Z coordinates, and apply sigmoid and normalize function to calculate a probability ρ 1 , of all M points, which will be applied during sampling.
P ω R M = μ × 1 1 + e σ × P x
ρ 1 R M = P ω P ω
P ϵ R M = N ( 0 , 0.01 ) × P ω
where μ and σ are two introduced weights. ρ 1 , i is the probability of P i to be selected during the afterward sampling. P ϵ is a density-aware Gaussian noise.
Figure 9 visually illustrates the input and output of PMDNet. Table 10 illustrates the metrics of each method on clean data. It is obvious that PMDNet is still outstanding, reaching M A E ( R ) 0.1 , M A E ( t ) 0.001 , 100% R e c a l l . Furthermore, it is interesting that, compared to density-uniform Clean results (Table 8), PMDNet achieves better performance here, while others make little progress. A major factor contributing to this is the higher continuity of the P M D feature. Although the clouds become sparser and noisier with increasing distance, the nearby points remain accurate and dense, producing more information than uniform sampling, which is exactly what was applied in the density-uniform experiment.
Table 11 summaries the results on noise data. Be noticed that, PMDNet fails again on three samples. The major reason is discussed in Section 4.3.2. ( x c ) produces a less accurate representation since the mesh geometry is corrupted by noise. On the other hand, these three samples consist of mostly planes, with little other geometry for learning. Comparing Table 10 and Table 11, we can see a larger decline in accuracy, where M A E ( R ) rises from 0.0622 to 4.6186, and M A E ( t ) from 0.00041 to 0.03232, than that in the density-uniform experiment, where M A E ( R ) rises from 0.1447 to 3.7088, and M A E ( t ) from 0.00089 to 0.01696.

4.5. Time Efficiency

Time efficiency is another important metric regarding registration methods. Thus, we finally test our PMDNet against other learning-based methods in terms of time efficiency. All the methods are evaluated on Windows 11, Intel i5 12400f, NVIDIA RTX 3060TI, 16 GB 3200 MHz RAM.
Table 12 illustrates the results. We perform this experiment on clouds of 512 and 1024 resolution, with each method fixed to 3 iterations. IDAM is definitely the fastest of all methods in this study, followed by DCP v2. Despite being slower than IDAM and DCP v2, PMDNet is over 40% faster compared to RPMNet.

5. Conclusions

In this work, we first introduce a novel local reference frame, the P M D feature, to solve the normal ambiguity of boundaries. Moreover, as we suggest feature matching precision of equal importance as spatial accuracy, we introduce feature loss to registration.
Extensive experiments show our PMDNet achieves state-of-the-art performance. More specifically, PMDNet achieved a 100% R e c a l l in the Unseen scenario with the generic dataset, which is 25% higher than the second-best, RPMNet. Meanwhile, PMDNet also achieved a 100% R e c a l l on density-decreasing BIM scenarios. Last but not least, PMDNet is 40% faster than RPMNet, and 16% slower than DCP v2. However, it is 3 × slower than IDAM.
P M D feature is defined to encourage the key points, however, coordinates show little robustness against noise. That’s why an obvious decline in accuracy is witnessed on noisy clouds. We are interested in further studies focused on robust detection and feature extraction on key points.

Author Contributions

Conceptualization, H.J.; Data curation, G.N. and T.T.; Funding acquisition, P.L.; Investigation, G.N.; Methodology, H.J.; Project administration, Z.W. and P.L.; Resources, T.T.; Software, H.J.; Supervision, P.L.; Writing—original draft, H.J.; Writing—review & editing, P.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research is supported by A*STAR (Agency for Science, Technology and Research of Singapore) under its National Robotics Programme (NRP)—Robotics Domain Specific (RDS) (Grant No. W2122d0155). Disclaimer: Any opinions, findings and conclusions or recommendations expressed in this material are those of the authors and do not reflect the views of the A*STAR.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhang, Y.Y.; Kang, K.; Lin, J.R.; Zhang, J.P.; Zhang, Y. Building information modeling–based cyber-physical platform for building performance monitoring. Int. J. Distrib. Sens. Netw. 2020, 16, 1550147720908170. [Google Scholar] [CrossRef] [Green Version]
  2. Sporr, A.; Zucker, G.; Hofmann, R. Automated HVAC Control Creation Based on Building Information Modeling (BIM): Ventilation System. IEEE Access 2019, 7, 74747–74758. [Google Scholar] [CrossRef]
  3. Alhassan, B.; Omran, J.Y.; Jrad, F.A. Maintenance management for public buildings using building information modeling BIM. Int. J. Inf. Syst. Soc. Change 2019, 10, 42–56. [Google Scholar] [CrossRef]
  4. Xie, Q.; Zhou, X.; Wang, J.; Gao, X.; Chen, X.; Liu, C. Matching Real-World Facilities to Building Information Modeling Data Using Natural Language Processing. IEEE Access 2019, 7, 119465–119475. [Google Scholar] [CrossRef]
  5. Wang, W.C.; Weng, S.W.; Wang, S.H.; Chen, C.Y. Integrating building information models with construction process simulations for project scheduling support. Autom. Constr. 2014, 37, 68–80. [Google Scholar] [CrossRef]
  6. Honcharenko, T.; Tsiutsiura, S.; Kyivska, K.; Balina, O.; Bezklubenko, I. Transform Approach for Formation of Construction Project Management Teams Based on Building Information Modeling. In Proceedings of the ITPM, Slavsko, Ukraine, 16–18 February 2021; pp. 11–21. [Google Scholar]
  7. Hu, C.; Zhang, S. Study on BIM technology application in the whole life cycle of the utility tunnel. In Proceedings of the International Symposium for Intelligent Transportation and Smart City, Shanghai, China, 9–11 May 2019; pp. 277–285. [Google Scholar]
  8. Ryzhakov, D. Innovative Tools for Management the Lifecycle of Strategic Objectives of the Enterprise-Stakeholder in Construction. Int. J. Emerg. Trends Eng. Res. 2020, 8, 4526–4532. [Google Scholar] [CrossRef]
  9. Choi, J.; Choi, J.; Kim, I. Development of BIM-based evacuation regulation checking system for high-rise and complex buildings. Autom. Constr. 2014, 46, 38–49. [Google Scholar] [CrossRef]
  10. Wang, S.H.; Wang, W.C.; Wang, K.C.; Shih, S.Y. Applying building information modeling to support fire safety management. Autom. Constr. 2015, 59, 158–167. [Google Scholar] [CrossRef]
  11. Wang, K.C.; Shih, S.Y.; Chan, W.S.; Wang, W.C.; Wang, S.H.; Gansonre, A.A.; Liu, J.J.; Lee, M.T.; Cheng, Y.Y.; Yeh, M.F. Application of building information modeling in designing fire evacuation—A case study. In Proceedings of the 31st International Symposium on Automation and Robotics in Construction and Mining, ISARC 2014, Sydney, Australia, 9–11 July 2014; pp. 593–601. [Google Scholar]
  12. Lotfi, N.; Behnam, B.; Peyman, F. A BIM-based framework for evacuation assessment of high-rise buildings under post-earthquake fires. J. Build. Eng. 2021, 43, 102559. [Google Scholar] [CrossRef]
  13. Siountri, K.; Skondras, E.; Vergados, D.D. Developing Smart Buildings Using Blockchain, Internet of Things, and Building Information Modeling. Int. J. Interdiscip. Telecommun. Netw. 2020, 12, 1–15. [Google Scholar] [CrossRef]
  14. Marjani, M.; Nasaruddin, F.; Gani, A.; Karim, A.; Hashem, I.A.T.; Siddiqa, A.; Yaqoob, I. Big IoT data analytics: Architecture, opportunities, and open research challenges. IEEE Access 2017, 5, 5247–5261. [Google Scholar]
  15. Lokshina, I.V.; Greguš, M.; Thomas, W.L. Application of integrated building information modeling, IoT and blockchain technologies in system design of a smart building. Procedia Comput. Sci. 2019, 160, 497–502. [Google Scholar] [CrossRef]
  16. Gebken, L.; Drews, P.; Schirmer, I. Enhancing the Building Information Modeling Lifecycle of Complex Structures with IoT: Phases, Capabilities and Use Cases. In Proceedings of the 52nd Hawaii International Conference on System Sciences, Maui, HI, USA, 8–11 January 2019. [Google Scholar]
  17. Trach, R.; Bushuyev, S. Analysis communication network of construction project participants. Przegląd Naukowy Inżynieria i Kształtowanie Środowiska 2020, 29. [Google Scholar] [CrossRef]
  18. Bushuyev, S.; Shkuro, M. Development of proactive method of communications for projects of ensuring the energy efficiency of municipal infrastructure. EUREKA Phys. Eng. 2019, 1, 3–12. [Google Scholar] [CrossRef] [Green Version]
  19. Wei, X.; Bonenberg, W.; Zhou, M.; Wang, J. The Impact of Building Information Modeling Design System on Traditional Urban Design Methods. In Advances in Human Factors in Architecture, Sustainable Urban Planning and Infrastructure, Proceedings of the AHFE 2021 Virtual Conference on Human Factors in Architecture, Sustainable Urban Planning and Infrastructure, San Francisco, CA, USA, 25–29 July 2021; Lecture Notes in Networks and Systems; Charytonowicz, J., Maciejko, A., Falcão, C.S., Eds.; Springer: Berlin, Germany, 2021; Volume 272, pp. 302–309. [Google Scholar]
  20. Sacks, R.; Eastman, C.; Lee, G.; Teicholz, P. BIM Handbook: A Guide to Building Information Modeling for Owners, Designers, Engineers, Contractors, and Facility Managers; John Wiley & Sons: Hoboken, NJ, USA, 2018. [Google Scholar]
  21. Engelmann, F.; Rematas, K.; Leibe, B.; Ferrari, V. From Points to Multi-Object 3D Reconstruction. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 19–25 June 2021; pp. 4588–4597. [Google Scholar]
  22. Xie, J.; Xu, Y.; Zheng, Z.; Zhu, S.C.; Wu, Y.N. Generative PointNet: Deep Energy-Based Learning on Unordered Point Sets for 3D Generation, Reconstruction and Classification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 19–25 June 2021; pp. 14976–14985. [Google Scholar]
  23. Qi, G.; Jinhui, L. A learning based 3D reconstruction method for point cloud. In Proceedings of the 2020 IEEE International Conference on Dependable, Autonomic and Secure Computing, International Conference on Pervasive Intelligence and Computing, International Conference on Cloud and Big Data Computing, International Conference on Cyber Science and Technology Congress (DASC/PiCom/CBDCom/CyberSciTech), Calabria, Italy, 12–15 September 2020; pp. 271–276. [Google Scholar]
  24. Navaneet, K.; Mandikal, P.; Agarwal, M.; Babu, R.V. Capnet: Continuous approximation projection for 3d point cloud reconstruction using 2d supervision. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; Volume 33, pp. 8819–8826. [Google Scholar]
  25. Babu, A.; Yurtdas, K.Y.; Koch, C.E.S.; Yüksel, M. Trajectory Following using Nonlinear Model Predictive Control and 3D Point-Cloud-based Localization for Autonomous Driving. In Proceedings of the 2019 European Conference on Mobile Robots (ECMR), Prague, Czech Republic, 4–6 September 2019; pp. 1–6. [Google Scholar]
  26. Rozenberszki, D.; Majdik, A.L. LOL: Lidar-only Odometry and Localization in 3D point cloud maps. In Proceedings of the 2020 IEEE International Conference on Robotics and Automation (ICRA), Virtual, 31 May–31 August 2020; pp. 4379–4385. [Google Scholar]
  27. O’Sullivan, E.; Zafeiriou, S. 3D Landmark Localization in Point Clouds for the Human Ear. In Proceedings of the 2020 15th IEEE International Conference on Automatic Face and Gesture Recognition (FG 2020), Buenos Aires, Argentina, 16–20 November 2020; pp. 402–406. [Google Scholar]
  28. Xie, Q.; Zhang, Y.; Cao, X.; Xu, Y.; Lu, D.; Chen, H.; Wang, J. Part-in-whole point cloud registration for aircraft partial scan automated localization. Comput.-Aided Des. 2021, 137, 103042. [Google Scholar] [CrossRef]
  29. Zhang, Z.; Hu, L.; Deng, X.; Xia, S. Sequential 3D Human Pose Estimation Using Adaptive Point Cloud Sampling Strategy. In Proceedings of the Thirtieth International Joint Conference on Artificial Intelligence (IJCAI-21), Montreal, QC, USA, 19–26 August 2021. [Google Scholar]
  30. Marcon, M.; Bellon, O.R.P.; Silva, L. Towards real-time object recognition and pose estimation in point clouds. arXiv 2020, arXiv:2011.13669. [Google Scholar]
  31. He, Y.; Sun, W.; Huang, H.; Liu, J.; Fan, H.; Sun, J. Pvn3d: A deep point-wise 3d keypoints voting network for 6dof pose estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 11632–11641. [Google Scholar]
  32. Wei, F.; Sun, X.; Li, H.; Wang, J.; Lin, S. Point-set anchors for object detection, instance segmentation and pose estimation. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; pp. 527–544. [Google Scholar]
  33. Besl, P.J.; McKay, N.D. Method for registration of 3-D shapes. In Proceedings of the Sensor Fusion IV: Control Paradigms and Data Structures, Boston, MA, USA, 2–15 November 1991; Volume 1611, pp. 586–606. [Google Scholar]
  34. Rusinkiewicz, S.; Levoy, M. Efficient variants of the ICP algorithm. In Proceedings of the Third International Conference on 3-D Digital Imaging and Modeling, Quebec City, QC, Canada, 28 May–1 June 2001; pp. 145–152. [Google Scholar]
  35. Zhang, J.; Yao, Y.; Deng, B. Fast and robust iterative closest point. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 3450–3466. [Google Scholar] [CrossRef]
  36. Eckart, B.; Kim, K.; Kautz, J. Hgmr: Hierarchical gaussian mixtures for adaptive 3d registration. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 705–721. [Google Scholar]
  37. Yuan, W.; Eckart, B.; Kim, K.; Jampani, V.; Fox, D.; Kautz, J. Deepgmr: Learning latent gaussian mixture models for registration. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; pp. 733–750. [Google Scholar]
  38. Rusu, R.B.; Blodow, N.; Marton, Z.C.; Beetz, M. Aligning point cloud views using persistent feature histograms. In Proceedings of the 2008 IEEE/RSJ International Conference on Intelligent Robots and Systems, Nice, France, 22–26 September 2008; pp. 3384–3391. [Google Scholar]
  39. Rusu, R.B.; Blodow, N.; Beetz, M. Fast Point Feature Histograms (FPFH) for 3D registration. In Proceedings of the 2009 IEEE International Conference on Robotics and Automation, Kobe, Japan, 12–17 May 2009; pp. 3212–3217. [Google Scholar]
  40. Salti, S.; Tombari, F.; Di Stefano, L. SHOT: Unique signatures of histograms for surface and texture description. Comput. Vis. Image Underst. 2014, 125, 251–264. [Google Scholar] [CrossRef]
  41. Drost, B.; Ulrich, M.; Navab, N.; Ilic, S. Model globally, match locally: Efficient and robust 3D object recognition. In Proceedings of the 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Francisco, CA, USA, 13–18 June 2010; pp. 998–1005. [Google Scholar]
  42. Wang, Y.; Solomon, J.M. Prnet: Self-supervised learning for partial-to-partial registration. arXiv 2019, arXiv:1910.12240. [Google Scholar]
  43. Lu, W.; Wan, G.; Zhou, Y.; Fu, X.; Yuan, P.; Song, S. DeepVCP: An End-to-End Deep Neural Network for Point Cloud Registration. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Korea, 27 October–2 November 2019. [Google Scholar]
  44. Deng, H.; Birdal, T.; Ilic, S. Ppfnet: Global context aware local features for robust 3d point matching. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 195–205. [Google Scholar]
  45. Aoki, Y.; Goforth, H.; Srivatsan, R.A.; Lucey, S. Pointnetlk: Robust & efficient point cloud registration using pointnet. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seoul, Korea, 27 October–2 November 2019; pp. 7163–7172. [Google Scholar]
  46. Wang, Y.; Solomon, J.M. Deep closest point: Learning representations for point cloud registration. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019; pp. 3523–3532. [Google Scholar]
  47. Lowe, D.G. Distinctive image features from scale-invariant keypoints. Int. J. Comput. Vis. 2004, 60, 91–110. [Google Scholar] [CrossRef]
  48. Bay, H. Luc Van Gool, SURF: Speeded-Up Robust Features. In Proceedings of the 9th European Conference on Computer Vision, Graz, Austria, 7–13 May 2006. [Google Scholar]
  49. Rublee, E.; Rabaud, V.; Konolige, K.; Bradski, G. ORB: An efficient alternative to SIFT or SURF. In Proceedings of the 2011 International Conference on Computer Vision, Barcelona, Spain, 6–13 November 2011; pp. 2564–2571. [Google Scholar]
  50. Li, J.; Lee, G.H. Usip: Unsupervised stable interest point detection from 3d point clouds. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019; pp. 361–370. [Google Scholar]
  51. Tinchev, G.; Penate-Sanchez, A.; Fallon, M. Skd: Keypoint detection for point clouds using saliency estimation. IEEE Robot. Autom. Lett. 2021, 6, 3785–3792. [Google Scholar] [CrossRef]
  52. Huang, X.; Mei, G.; Zhang, J.; Abbas, R. A comprehensive survey on point cloud registration. arXiv 2021, arXiv:2103.02690. [Google Scholar]
  53. Johnson, A.E.; Hebert, M. Using spin images for efficient object recognition in cluttered 3D scenes. IEEE Trans. Pattern Anal. Mach. Intell. 1999, 21, 433–449. [Google Scholar] [CrossRef] [Green Version]
  54. Makadia, A.; Patterson, A.; Daniilidis, K. Fully automatic registration of 3D point clouds. In Proceedings of the 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’06), New York, NY, USA, 17–22 June 2006; Volume 1, pp. 1297–1304. [Google Scholar]
  55. Qi, C.R.; Su, H.; Mo, K.; Guibas, L.J. Pointnet: Deep learning on point sets for 3d classification and segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 652–660. [Google Scholar]
  56. Qi, C.R.; Yi, L.; Su, H.; Guibas, L.J. Pointnet++: Deep hierarchical feature learning on point sets in a metric space. arXiv 2017, arXiv:1706.02413. [Google Scholar]
  57. Li, L.; Yang, F.; Zhu, H.; Li, D.; Li, Y.; Tang, L. An improved RANSAC for 3D point cloud plane segmentation based on normal distribution transformation cells. Remote Sens. 2017, 9, 433. [Google Scholar] [CrossRef]
  58. Schnabel, R.; Wahl, R.; Klein, R. Efficient RANSAC for point-cloud shape detection. Comput. Graph. Forum 2007, 26, 214–226. [Google Scholar] [CrossRef]
  59. Nurunnabi, A.; Belton, D.; West, G. Robust segmentation in laser scanning 3D point cloud data. In Proceedings of the 2012 International Conference on Digital Image Computing Techniques and Applications (DICTA), Fremantle, Australia, 3–5 December 2012; pp. 1–8. [Google Scholar]
  60. Grant, W.S.; Voorhies, R.C.; Itti, L. Finding planes in LiDAR point clouds for real-time registration. In Proceedings of the 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems, Tokyo, Japan, 3–7 November 2013; pp. 4347–4354. [Google Scholar]
  61. Li, M.; Gao, X.; Wang, L.; Li, G. Automatic registration of laser-scanned point clouds based on planar features. In Proceedings of the 2nd ISPRS International Conference on Computer Vision in Remote Sensing (CVRS 2015), Xiamen, China, 28–30 April 2015; Volume 9901, pp. 7–13. [Google Scholar]
  62. Mahmood, B.; Han, S.; Lee, D.E. BIM-Based Registration and Localization of 3D Point Clouds of Indoor Scenes Using Geometric Features for Augmented Reality. Remote Sens. 2020, 12, 2302. [Google Scholar] [CrossRef]
  63. Li, Z.; Zhang, X.; Tan, J.; Liu, H. Pairwise Coarse Registration of Indoor Point Clouds Using 2D Line Features. ISPRS Int. J. Geo-Inf. 2021, 10, 26. [Google Scholar] [CrossRef]
  64. Sheik, N.A.; Deruyter, G.; Veelaert, P. Plane-Based Robust Registration of a Building Scan with Its BIM. Remote Sens. 2022, 14, 1979. [Google Scholar] [CrossRef]
  65. Sinkhorn, R. A relationship between arbitrary positive matrices and doubly stochastic matrices. Ann. Math. Stat. 1964, 35, 876–879. [Google Scholar] [CrossRef]
  66. Yew, Z.J.; Lee, G.H. Rpm-net: Robust point matching using learned features. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 11824–11833. [Google Scholar]
  67. Fu, K.; Liu, S.; Luo, X.; Wang, M. Robust point cloud registration framework based on deep graph matching. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 8893–8902. [Google Scholar]
  68. Wu, Z.; Song, S.; Khosla, A.; Yu, F.; Zhang, L.; Tang, X.; Xiao, J. 3d shapenets: A deep representation for volumetric shapes. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1912–1920. [Google Scholar]
  69. Rawat, P.; Bhadoria, R.S.; Gupta, P.; Dimri, P.; Saroha, G. Performance evaluation of an adopted model based on big-bang big-crunch and artificial neural network for cloud applications. Kuwait J. Sci. 2021, 48, 1–13. [Google Scholar] [CrossRef]
Figure 1. (a) Classic optimization-based method. (b) Learning-based method.
Figure 1. (a) Classic optimization-based method. (b) Learning-based method.
Sensors 22 09694 g001
Figure 2. (a) PMDNet overview. Loss propagation is shown in orange. (b) Feature extractor. (c) Annealing parameter prediction network.
Figure 2. (a) PMDNet overview. Loss propagation is shown in orange. (b) Feature extractor. (c) Annealing parameter prediction network.
Sensors 22 09694 g002
Figure 3. (a) Normal Ambiguity. The normal of boundary p is not explicit, instead depended on which mesh the neighbors are located during calculation. (b) an example of normal ambiguity. Normals are differently colored according to their orientations. The boundary between the zoomed purple and green surfaces is not explicit.
Figure 3. (a) Normal Ambiguity. The normal of boundary p is not explicit, instead depended on which mesh the neighbors are located during calculation. (b) an example of normal ambiguity. Normals are differently colored according to their orientations. The boundary between the zoomed purple and green surfaces is not explicit.
Sensors 22 09694 g003
Figure 4. Example of ModelNet40 samples.
Figure 4. Example of ModelNet40 samples.
Sensors 22 09694 g004
Figure 5. Experiment process.
Figure 5. Experiment process.
Sensors 22 09694 g005
Figure 6. Qualitative results of the PMDNet on clean data. (a) source and reference clouds. (b) reference and predicted clouds. (c) correspondence between source and reference. (d) ground-truth correspondence. (e) correspondence between predicted and reference cloud.
Figure 6. Qualitative results of the PMDNet on clean data. (a) source and reference clouds. (b) reference and predicted clouds. (c) correspondence between source and reference. (d) ground-truth correspondence. (e) correspondence between predicted and reference cloud.
Sensors 22 09694 g006
Figure 7. Qualitative results of the PMDNet on unseen categories. (a) source and reference clouds. (b) reference and predicted clouds. (c) correspondence between source and reference. (d) ground-truth correspondence. (e) correspondence between predicted and reference cloud.
Figure 7. Qualitative results of the PMDNet on unseen categories. (a) source and reference clouds. (b) reference and predicted clouds. (c) correspondence between source and reference. (d) ground-truth correspondence. (e) correspondence between predicted and reference cloud.
Sensors 22 09694 g007
Figure 8. Qualitative results of the PMDNet on density-uniform clean BIM data. s r c , r e f , and p r e d clouds are colored green, red, and blue, respectively.
Figure 8. Qualitative results of the PMDNet on density-uniform clean BIM data. s r c , r e f , and p r e d clouds are colored green, red, and blue, respectively.
Sensors 22 09694 g008
Figure 9. Qualitative results of the PMDNet on density-decreasing clean BIM data. s r c , r e f , and p r e d clouds are colored green, red, and blue, respectively.
Figure 9. Qualitative results of the PMDNet on density-decreasing clean BIM data. s r c , r e f , and p r e d clouds are colored green, red, and blue, respectively.
Sensors 22 09694 g009
Table 1. Parameters of PMDNet.
Table 1. Parameters of PMDNet.
ParameterValue
learning rate 1 × 10 4
epochs1024
batch size8
optimizerAdam
Table 2. Evaluate Metrics.
Table 2. Evaluate Metrics.
MetricsRef EquationNotes
err_r_deg_mean (ERM)Equation (11)Mean of isotropic Error of Rotation
err_t_mean (ETM)Equation (12)Mean of isotropic Error of Translation
CCDEquation (14)Clip Chamfer Distance
MAE(R)-Mean Absolute Error of Rotation,
in the unit of degrees
MAE(T)-Mean Absolute Error of Translation,
in the unit of degrees
Recall( ω , δ )-Proportion of samples with,
MAE(R) <  ω ( ) and MAE(t) <  δ (m)
Table 4. Ablation setup. P M D N e t A , P M D N e t B , P M D N e t C , P M D N e t D , P M D N e t E select different components of the P M D feature.
Table 4. Ablation setup. P M D N e t A , P M D N e t B , P M D N e t C , P M D N e t D , P M D N e t E select different components of the P M D feature.
ID x c x c x i ( x c ) ( n r , n i )
P M D N e t A
P M D N e t B
P M D N e t C
P M D N e t D
P M D N e t E
Table 5. Results on clean data. Bold and underline denote best and second best performance.
Table 5. Results on clean data. Bold and underline denote best and second best performance.
MethodMAE(R)↓MAE(t)↓ERM↓ETM↓CCD↓Recall(1.0, 0.1)↑
ICP6.44670.054463.0790.024420.03009074.19%
FGR0.00990.000100.0060.000050.00019099.96%
RPMNet0.24640.000500.1090.000500.00089098.14%
IDAM1.35360.026050.7310.012440.04470075.81%
DeepGMR0.01560.000020.0010.000010.000030100.00%
PMDNet0.04670.000390.0870.000810.00000399.83%
Table 6. Results on unseen data. Bold and underline denote best and second best performance. They are trained on the first twenty categories and tested on the remaining categories.
Table 6. Results on unseen data. Bold and underline denote best and second best performance. They are trained on the first twenty categories and tested on the remaining categories.
MethodMAE(R)↓MAE(t)↓ERM↓ETM↓CCD↓Recall(1.0, 0.1)↑
FGR41.96310.2910623.9500.140670.1237005.13%
RPMNet1.98260.022761.0410.010670.08704075.59%
IDAM19.32490.2072910.1580.100630.1292100.95%
DeepGMR71.06770.4463244.3630.220390.1472800.24%
DCP v22.00720.003703.1500.00503NANA
PMDNet0.04230.000370.0760.000750.000003100.00%
Table 7. Results on Gaussian noise data. Bold and underline denote best and second best performance.
Table 7. Results on Gaussian noise data. Bold and underline denote best and second best performance.
MethodMAE(R)↓MAE(t)↓ERM↓ETM↓CCD↓Recall(1.0, 0.1)↑
ICP6.50300.049443.1270.02250.0538777.59%
FGR10.00790.070805.4050.03380.0691830.75%
RPMNet0.57730.005320.3050.00250.0425796.68%
IDAM3.49160.029151.8180.01410.0543649.59%
DeepGMR2.27360.014981.1780.00710.0502956.52%
DCP v20.73740.001051.0810.0015NANA
PMDNet1.12010.009902.2240.02080.0008480.28%
Table 8. Results on clean density-uniform BIM scenarios. Bold and underline denote best and second best performance, respectively.
Table 8. Results on clean density-uniform BIM scenarios. Bold and underline denote best and second best performance, respectively.
MethodsMAE(R)↓MAE(T)↓ERM↓ETM↓CCD↓Recall(1.0, 0.1)↑
DCP v13.97880.004335.6410.088230.08945313.33%
DCP v21.03280.013191.4150.026140.08806150.00%
IDAM23.70440.0812550.1760.160670.0944560.00%
PMDNet0.14470.000890.5220.001810.000009100.00%
Table 9. Results on noise density-uniform BIM scenarios. Bold and underline denote best and second best performance, respectively. PMDNet fails on three scenarios, contributing to large M A E ( R ) and E R M , but achieves great performance on all the other samples. PMDNet shows the metrics on successfully predicted scenarios.
Table 9. Results on noise density-uniform BIM scenarios. Bold and underline denote best and second best performance, respectively. PMDNet fails on three scenarios, contributing to large M A E ( R ) and E R M , but achieves great performance on all the other samples. PMDNet shows the metrics on successfully predicted scenarios.
MethodsMAE(R)↓MAE(T)↓ERM↓ETM↓CCD↓Recall(1.0, 0.1)↑
DCP v13.89520.044435.5280.089000.08804710.00%
DCP v20.02980.029851.2410.058930.08804156.67%
IDAM21.93740.0830848.0640.160320.0886340.00%
PMDNet3.70880.0169610.0970.042320.00266990.00%
PMDNet 0.30390.002410.4960.004900.000442100.00%
Table 10. Results on clean density-decreasing BIM scenarios. Bold and underline denote best and second best performance, respectively.
Table 10. Results on clean density-decreasing BIM scenarios. Bold and underline denote best and second best performance, respectively.
MethodsMAE(R)↓MAE(T)↓ERM↓ETM↓CCD↓Recall(1.0, 0.1)↑
DCP v13.61610.956635.1301.125230.1984680.00%
DCP v20.99480.969181.4101.118230.1983360.00%
IDAM23.06351.5671026.3031.782470.1999770.00%
PMDNet0.06220.000410.1010.000610.000002100.00%
Table 11. Results on noise density-decreasing BIM scenarios. Bold and underline denote best and second best performance, respectively. PMDNet fails on three scenarios, contributing to large M A E ( R ) and E R M , but achieves great performance on all the other samples. PMDNet shows the metrics on successfully predicted scenarios.
Table 11. Results on noise density-decreasing BIM scenarios. Bold and underline denote best and second best performance, respectively. PMDNet fails on three scenarios, contributing to large M A E ( R ) and E R M , but achieves great performance on all the other samples. PMDNet shows the metrics on successfully predicted scenarios.
MethodsMAE(R)↓MAE(T)↓ERM↓ETM↓CCD↓Recall(1.0, 0.1)↑
DCP v13.19350.933634.2811.095410.1981790.00%
DCP v21.00330.947441.2951.089870.1980620.00%
IDAM23.04951.5238126.6161.716340.2000000.00%
PMDNet4.61860.0323223.9870.132190.00151690.00%
PMDNet 0.43580.003050.7360.006040.00049396.67%
Table 12. Results on time efficiency. Bold and underline denote best and second best performance, respectively. All the results are in the unit of milliseconds.
Table 12. Results on time efficiency. Bold and underline denote best and second best performance, respectively. All the results are in the unit of milliseconds.
PointsDCP v2(3 Iters)RPMNet(3 Iters)IDAM(3 Iters)PMDNet(3 Iters)
51215.0432.475.8417.72
102418.8138.357.1121.58
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Jiang, H.; Lasang, P.; Nader, G.; Wu, Z.; Tanasnitikul, T. Feature Consistent Point Cloud Registration in Building Information Modeling. Sensors 2022, 22, 9694. https://doi.org/10.3390/s22249694

AMA Style

Jiang H, Lasang P, Nader G, Wu Z, Tanasnitikul T. Feature Consistent Point Cloud Registration in Building Information Modeling. Sensors. 2022; 22(24):9694. https://doi.org/10.3390/s22249694

Chicago/Turabian Style

Jiang, Hengyu, Pongsak Lasang, Georges Nader, Zheng Wu, and Takrit Tanasnitikul. 2022. "Feature Consistent Point Cloud Registration in Building Information Modeling" Sensors 22, no. 24: 9694. https://doi.org/10.3390/s22249694

APA Style

Jiang, H., Lasang, P., Nader, G., Wu, Z., & Tanasnitikul, T. (2022). Feature Consistent Point Cloud Registration in Building Information Modeling. Sensors, 22(24), 9694. https://doi.org/10.3390/s22249694

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop