Next Article in Journal
Enhancing the Privacy of Network Services through Trusted Computing
Previous Article in Journal
Application of Soft Computing Techniques for Predicting Thermal Conductivity of Rocks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Line Structure Extraction from LiDAR Point Cloud Based on the Persistence of Tensor Feature

1
Jiangsu Key Laboratory of Spectral Imaging & Intelligence Sense (SIIS), Nanjing University of Science and Technology, Nanjing 210094, China
2
Smart Health Big Data Analysis and Location Services Engineering Research Center of Jiangsu Province, Nanjing University of Posts and Telecommunications, Nanjing 210023, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2022, 12(18), 9190; https://doi.org/10.3390/app12189190
Submission received: 1 August 2022 / Revised: 28 August 2022 / Accepted: 31 August 2022 / Published: 14 September 2022
(This article belongs to the Topic Artificial Intelligence in Sensors)

Abstract

:

Featured Application

line structure extraction, tensor voting, persistent homology.

Abstract

The LiDAR point cloud has been widely used in scenarios of automatic driving, object recognition, structure reconstruction, etc., while it remains a challenging problem in line structure extraction, due to the noise and accuracy, especially in data acquired by consumer electronic devices. To address the issue, a line structure extraction method based on the persistence of tensor feature is proposed, and subsequently applied to the data acquired by an iPhone-based LiDAR sensor. The tensor of each point is encoded, voted, and aggregated by its neighborhood, and further decomposed into different geometric features in each dimension. Then, the line feature in the point cloud is represented and computed using the persistence of the tensor feature. Finally, the line structure is extracted based on the persistent homology according to the discrete Morse theory. With the LiDAR point cloud collected by the iPhone 12 Pro MAX, experiments are conducted, line structures are extracted from two different datasets, and results perform well in comparison with other related results.

1. Introduction

Light detection and ranging (LiDAR) is a method to measure the distance between the object and the receiver based on the reflection of light and obtain massive points for the surface of an area instantly [1,2,3]. With high efficiency and versatile performance, it has been extensively applied in surveying and mapping, automatic driving, scene understanding [4,5], etc. In recent years, LiDAR sensors have been equipped in consumer electronic devices, such as Kinect, iPhone, etc., which has made it more and more convenient to acquire point cloud data for common users, and many applications have been constructed [6,7], such as line structure extraction, shape recognition, object detection and classification, and some high-level applications (scene understanding, simultaneous localization and mapping, etc.).
With these easy-to-use LiDAR devices, plenty of point cloud datasets are provided, and abundant information can be extracted, such as geometric feature, semantic labeling, and scenario relations [8]. However, it also brings many problems in dealing with these redundant point cloud datasets, since not all points are needed, and it’s difficult to extract the structure information, compress the data, and represent these redundant datasets by simple geometric structures [9,10]. To solve these problems and obtain structure information, different studies have been conducted using, e.g., deep learning-based feature extraction methods and geometric model fitting methods [7,8,11,12]. These methods either need predefined geometric models [12,13] or large amounts of training datasets [7]. In addition, some post-processing operations are also needed to maintain connection relations of geometric structures, and results can be affected by the quality of the point cloud datasets [14,15]. (1) How to extract line information from the quality-unstable point cloud dataset collected by consumer electronic devices, and (2) how to construct line structures without predefined geometric models and manually selected training datasets, remain challenges.
To address the issue, a line structure extraction method based on the tensor voting [16,17] and the persistent homology theory [18,19,20] is proposed. The line structure extraction framework is designed for point cloud datasets collected by the iPhone-based LiDAR sensor. The line feature in each point is encoded by the tensor voting, then the line segment is represented as the connection of each point with the highest local line feature value, and the line structure is constructed from line segments along with their structure connections. Contributions are as follows: (1) We compute the line feature from the point cloud based on the tensor voting theory, decompose the tensor feature, and make a combination of different dimensional geometric features, to get the line candidate dataset. (2) We represent line feature by the tensor of the Morse–Smale complex, calculate the line structure from LiDAR point cloud based on the persistent homology theory, and extract the line structure using the connect relations of critical points. (3) We propose the unified framework, design the algorithm to extract line structure from point cloud collected by the iPhone-based LiDAR sensor, and make a comparison between line structure extraction results.
The remainder of this paper is structured as follows. The next section gives an overview of works related to this paper. Section 3 provides a detailed description of the line structure extraction method based on the persistence of tensor feature, including tensor voting, tensor feature decomposition, critical point representation in the discrete Morse theory, and the line structure extraction. Experiments are conducted. The results are discussed in Section 4, followed by conclusions in Section 5.

2. Related Works

Line structure is the concise representation of 3D scenes, and the process relates to description of line features, segmentation and recognition of objects, and scene understanding of point cloud datasets [5,6,7,8], etc.
One of the heated research fields is edge map detection from depth or RBG image, to get the layout information of indoor scenes [2,8]. The edge feature is computed by view relations between the 3D scene to the 2D image, using neural network or some image-based edge detection algorithms. Hence, it’s also applied to the 3D point cloud [11,21,22]. Chen et al. [21] proposed a 3D line segment detection method from the multi-view stereo: project 3D points into planar sets on different camera matrices and computing line relations based on 2D images, then remove false matching relations, cast line features back to the 3D space of point clouds, and get 3D line segments. Another similar line segment detection strategy is the work conducted by Lu et al. [22]. The point cloud dataset is segmented into super planes based on region growing and merging. Then, points belonging to each plane are projected onto it, and 3D points in each plane become 2D images, followed by line extraction using 2D image contour extraction and least-square-fitting based line detection. In these line feature extraction methods, 3D line features can be duplicated or lost during dimensional transformation. However, line structure extraction directly from 3D point cloud can be conducted by an intuitive way. The first kind of intuitive method is the 3D shape detection from 3D point cloud, according to fitting error between the point cloud and the predefined geometric model. Hough transformation [13] is usually applied to shape detection situations, especially line segments, along with the random sample consensus (RANSAC) [12], which can extract geometric shapes from noisy point cloud data; besides, there are also some shape detection methods based on neural network that also works for the similar propose. Nevertheless, these methods depend on the predefined geometric model, which limited the capability of application. The other kind of intuitive method is the line structure extraction based on the 3D point cloud segmentation [1,4,23,24], where the point cloud dataset is directly divided into different object groups, and the rim of geometric objects is extracted to construct edge structures. Gankhuyag et al. [24] proposed an automatic 2D floorplan CAD generation method based on plane detection, and points segmented into floors are preserved to build 2D CAD line segments. For 3D line structure detection, Lin et al. [4] proposed a line extraction method by segmenting point cloud into facets and extracting boundary of each facet to get the 3D line. Although the structure information can be further refined using some semantic information, lines are extracted under the assumption of predefined geometric model or rules, i.e., the straight-line restriction. There are also some neural-network-based point cloud segmentation methods [5,7,11] that can extract boundaries more precisely than the predefined geometric models or rules, but these methods need large amounts of training data. In addition, some post-processing operations are also applied to maintain connection relations of geometric structures.
Another heated research field is line structure extraction from point clouds based on the line feature representation [25,26]. Using the consecutive geometric feature contained in point cloud, line structures can be detected and traced along the consecutive direction of geometric feature [6,9,10,27,28], such as angle differences between the normal vector and its neighboring points, curvature computed from assumed point cloud surface, etc. These methods usually need normal vectors recorded in the point cloud dataset. While there is no normal information stored, some techniques need to be applied to compute normal vectors [6,9,27], such as local plane fitting, principal component analysis (referred to as PCA), and tensor voting. In PCA-based methods, the normal vector of point is computed using the local covariance matrix based on its neighborhood [27], and different dimensional geometric features can be interpreted from eigen decomposition of the matrix; besides, the optimal size of neighborhood is the critical point and is affected by the quality of point cloud dataset [10]. In tensor-voting-based methods, it can deal with the noisy dataset using the tensor encoding and revoting process and has the capability of N-dimensional geometric feature representation [6,15,16,28,29]. Using tensor voting, normal vector of the point is computed from its neighborhood and subsequently revoted to get the refined result. Then, geometric features of different dimensional are provided, and the subset of the point cloud dataset with the label of line feature is extracted [29]. However, the next step of extracting line structure from the selected line subset, still remains to be a challenge. Compared with geometric-model-based shape detection methods, the topological relation-based method can deal with the graph connection of line structures. One of the widely used topological methods is the persistent homology [30,31,32], where topological relations are computed and captured based on N-dimensional holes during growth of the distance to merge all points. Using the persistent homology theory, Carlsson [19] computed topological patterns contained in a point cloud dataset and built pattern signatures by barcodes. Another form of topological features computed based on the persistent homology is studied by Zhou et al. [20] and Wong et al. [33], and topological features are presented by the persistent image, which is further computed using the persistent diagram. Moreover, Beksi et al. [34] presented a 3D point descriptor according to birth and death of 0-cycles and 1-cycles and took it as topological signatures to classify categories for noisy point cloud datasets. These topological methods can deal with quality-unstable point cloud dataset using the topological feature and preserve graph connections with each line segment, while it is difficult to maintain geometric shapes in the dataset.
To deal with the quality-unstable point cloud dataset and obtain connection relation preserved line structures, a line structure extraction method is proposed in this paper, and point cloud datasets are collected using the iPhone-based LiDAR sensor. In addition, structure refinement based on the assumed geometric rules is not going to be conducted in this paper, since line structures are extracted with no predefined geometric models and no manually selected training datasets.

3. Methodology

Suppose a point p is consisted of N-dimensional coordinate and without normal vector information, which can be denoted as p(x1, x2, …, xN). The line feature is the geometric information of each point, and the higher value of the line feature, the more probability of the point to be a part of lines. Hence, the line segment can be constructed by connecting each point with the highest local line feature value. The line structure is built using line segments and their structure connections. To get the line structure from the unstructured point cloud, the geometric feature to express line feature needs to be constructed, which is computed based on the tensor voting, then line datasets to extract line segments is generated. The next step is extracting consecutive line segment from the subset, which is conducted based on the persistence of tensor feature, using topological relations calculated by the persistent homology. Hence, the line structure extraction process consists of 4 steps:
(1)
initialize the tensor for unstructured point and get geometric features in different dimensions;
(2)
revote the tensor feature in different dimensions and compute the refined geometric feature;
(3)
represent the geometric feature and construct the line candidate subset (referred to as LCS);
(4)
construct the Morse–Smale complex and extract line structures based on the discrete Morse theory.
The pipeline of the line structure extraction framework is depicted in Figure 1.

3.1. Initial Tensor Voting and Dimensional Feature Presentation

According to the manifold theory, the N-dimensional space S is composed of d-dimensional normal subspace SN and (Nd)-dimensional tangent subspace SN−d, which are orthogonal and complementary to each other, as denoted in Equation (1). The geometric feature in each dimension can be detected based on the normal subspace Sd, where the normal vector depicts the orthogonal direction of geometric structure. For example, in the 3-dimensional space (N = 3), the geometric structure with 1-dimensional normal space S1 can be detected as “surface,” the geometric structure with 2-dimensional normal space S2 can be detected as “line”, and the geometric structure with 3-dimensional normal space S3 can be detected as “point”.
S = S d + S N d   &   S d · S N d = 0
However, if there is no normal recorded, the normal space can be computed using tangent information from its neighborhood. Suppose a point p with a point pi fromneighborhood Ω , and the vector from pi to p can be denoted as vi, which contains the tangent information of pi. Then vi is normalized as v ^ i and the tangent subspace S N d i can be computed based on the Kronecker delta of v ^ i , as denoted in Equation (2).
S N d i = k r o n ( v ^ i , v ^ i T )
Since there is no normal vector recorded and no preferred geometric features set for space S, it can be denoted as the identity matrix I for space S. Hence, the geometric structure encoded by normal subspace S d i that will be transmitted from point pi to p can be denoted as complementary form of S N d i according to Equation (1), i.e., S d i = I S N d i . As seen in Figure 2, the geometric feature of each point in neighborhood of p is voted to the point, with geometric information transmitted by the normal subspace.
Besides, the intensity of geometric feature decays along distance ri from pi to p, and the decay function is usually taken as the Gaussian function, i.e., w ( r ) = e ( r ρ ) 2 . Here, the ρ for w is set to be 0.1. Finally, tensor Tp at point p is computed and accumulated using that of the point from its neighborhood, as denoted in Equation (3).
T p = p i Ω w ( r i ) ( I S N d i )

3.2. The Refinement of Initial Tensor Using Different Dimensional Geometric Feature

With tensor T computed through initial tensor voting process, the geometric feature in each dimension can be computed based on the tensor decomposition of T. In N-dimensional space, T is presented by a N × N dimensional matrix, and can be further decomposed by eigenvectors ( e 1 , e 2 , , e N ) and related eigenvalues ( λ 1 , λ 2 , , λ N ) , where λ 1 λ 2 λ N . Suppose the d dimensional geometric feature is going to be transmitted from point p(x1, x2, …, xN) to q(x1, x2, …, xN), the normal vector of normal subspace is denoted as v n = ( e 1 , e 2 , , e d ) , the tangent direction vector is denoted as v t = ( e d + 1 , e d + 2 , , e N ) , and vector from p to q is denoetd as v = qp. Then, the normal vector vk, which is transmitted along a minimal circle path defined by points p and q, meets the relations defined in Equation (4).
v k = v n cos ( 2 θ ) v t sin ( 2 θ )
In Equation (4), angle θ is computed based on cosine value and dot multiply operation using geometric relations of vt, and vk, i.e., θ = a cos ( d o t ( n o r m ( v k ) , v t ) . Based on Equations (2)–(4), the d dimensional tensor feature T d transmitted from p to q can be encoded as Equation (5).
T d = w ( r , θ ) k r o n ( v k , v k T )
w ( r , θ ) = e ( r ρ ) 2 ( cos ( θ ) ) 4 is the decay function with distance r and curvature information encoded as θ . It has been proven that the angle θ will become 0 where dimension d ≥ 2, and there is no need to recompute θ in each dimension, if vk is projected to Sd (King [16]). Hence d-dimension tensor feature T d can be represented as Equation (6).
T d = { w ( r , θ ) k r o n ( v k , v k T ) d = 1 w ( r ) ( S d k r o n ( v n , v n T ) ) d 2
On the other hand, the intensity sd of d dimensional geometric feature can be represented using eigenvalues ( λ 1 , λ 2 , , λ N ) , as denoted in Equation (7).
s d = { λ d λ d + 1 d < N λ N d = N
Based on Equations (6) and (7), tensor voted from p to q in each dimension can be computed by Equation (8).
T p = d = 1 d N s d T d
Finally, the tensor Tq of point q is voted by tensor Tp of each point p in p’s neighborhood Ω , as denoted in Equation (9), and this process is called the tensor refinement. The reader can refer to Wang, et al. [6] for a detailed derivation of the tensor refinement process.
T q = p Ω T p

3.3. The Construction of LCS Based on Geometric Saliency

In 3D space, the point is composed of (x, y, z), which is just position information and no normal vector information recorded, and refined voting results is decomposed by eigenvectors ( e 1 , e 2 , e 3 ) and related eigenvalues ( λ 1 , λ 2 , λ 3 ) , where λ 1 λ 2 λ 3 . The sailency of geometric feature that encoded in each dimension can be computed based on Equation (7):
(1)
in 1-dimensional normal space S 1 = k r o n ( e 1 , e 1 T ) , the geometric feature turns out to be “surface”, since there is only the 1D stick-shaped normal vector, and geometric saliency is s s u r f a c e = λ 1 λ 2 ;
(2)
in 2-dimensional normal space S 2 = k r o n ( e 1 , e 1 T ) + k r o n ( e 2 , e 2 T ) , the geometric feature turns out to be “line”, since there is the 2D surface-shaped normal vector, and geometric saliency is s l i n e = λ 2 λ 3 ;
(3)
in 3-dimensional normal space S 3 = k r o n ( e 1 , e 1 T ) + k r o n ( e 2 , e 2 T ) + k r o n ( e 3 , e 3 T ) , the geometric feature turns out to be “point”, since there is the 3D ball-shaped normal vector, and geometric saliency is s p o i n t = λ 3 .
In above descriptions, the higher geometric saliency value of the point, the more credible the geometric feature category it belongs to; for example, if the point with line geometric saliency s l i n e higher than the saliencies of other dimensions, it can categorize into LCS. However, if the point with surface saliency s s u r f a c e lower than some threshold value, it can also be taken into the LCS. Besides, the point with high point saliency s p o i n t turns out to be the corner structure of geometric framework, hence, it can also be taken into the LCS. With these assumptions, the LCS can be constructed based on threshold values { σ s u r f a c e , σ l i n e , σ p o i n t } for the saliency in different dimensions, as dented in Equation (10).
L C S = { p ( x , y , z , s s u r f a c e , s l i n e , s p o i n t ) S | s s u r f a c e σ s u r f a c e s l i n e σ l i n e s p o i n t σ p o i n t }

3.4. Line Structure Extraction Based on the Discrete Morse Theory

Suppose f : S ( p ) R is a smooth function on a manifold, and { p , f ( p ) | p S } can be taken as a smooth surface in 3D space. Besides, the gradient f ( p ) of the function f at point p is the direction f(p) decrease with the largest rate, and the integral line ι on the manifold is defined as the maximal path passing through p whose vectors agree with gradient f ( p ) . The start and end point of the integral line are critical points, which are non-degenerate and with gradient values f ( p ) = 0. In other words, the critical point can be the maxima, minima, or saddle. Hence, the 1-stable manifold ι ( p ) of the critical point p can be expressed as the line structure whose integral line ends at p (i.e., the maxima). Intuitively, the 1-stable manifold can be taken as the ridge of the surface, where each point of 1-stable manifold is the local maxima if there is a line profile which is orthogonal to the direction of gradient direction of the point, as depicted in Figure 3.
For the line structure extraction, line segments can be extracted with the local maximal line saliency value, i.e., the integral line starts from the saddle to the maxima (the black line in Figure 3), as denoted in Equation (11).
ι ( p ) = { p } { q S L C S | d e s t ( q ) = p }
Using the discrete Morse theory, 1-stable manifold of the integral line can be encoded by the Morse–Smale complex, and the persistent geometric feature can be computed. Then 1-stable manifold ι ( p ) starting from the saddle to the maxima can be extracted. To construct the Morse–Smale complex from the quality-unstable point cloud dataset, the LCS is further resampled by regular grid of space S, and the new LCS is computed as follows:
(1)
compute the bounding box of space S, take the minimum edge as the referenced length, and divide it into κ sub-edges of equal length τ ;
(2)
divide the edge of bounding box in other 2 dimensions using the length τ . Then, space S is divided into subspace ς i of equal size. Take the center position c e n t e r ( ς i ) of ς i as the coordinate of the point in LCS;
(3)
compute the relation φ i : { p ( x , y , z , s s u r f a c e , s l i n e , s p o i n t ) | p L C S } ς i between the point pi in LCS and the subspace ς i , and take the point with the max line saliency as the new attribute for ς i , as denoted in Equation (12);
L C S = { p ( x , y , z , s l i n e ) | ( x , y , z ) = c e n t e r ( ς i ) , s l i n e = m a x ( { s l i n e } ) }
(4)
count the number δ i of φ i in each ς i , and label ς i with δ i > 0 as the mask area ς i of mask space S = { ς i | δ i > 0 } , for the computation of persistent homology.
After the resample process, the Morse–Smale complex is constructed based on the discrete Morse theory, and the persistence of the geometric structure is computed. The line segment is extracted using the threshold δ , and line features with higher persistent value than δ are preserved, i.e., { p c r i t i c a l , l c o n n e c t i o n } = { ι ( p ) | ξ ι ( p ) > δ } . Finally, the graph is computed and the line structure { ι } is constructed with connection information stored in the graph.

3.5. The Algorithm of the Line Structure Extraction Framework

The line structure extraction process is consisted of 4 steps. First, compute the initial tensor using the tangent space and decompose geometric features in different dimensions. Second, revote the tensor to get refined results and construct the LCS. Third, calculate the Morse–Smale complex for the LCS and get persistent features. Finally, extract the persistent line structure based on connection relations of critical points. The algorithm is depicted in Algorithm 1.
Algorithm 1 The algorithm of the line structure extraction framework.
Line structure extraction framework LSE(P,r, σ , κ , δ )
INPUT: point cloud P, searching distance for neighborhood r, saliency thresholds σ { σ s u r f a c e , σ l i n e , σ p o i n t } , resampled grid κ , persistence threshold for line segment δ
OUTPUT: line structure with connection relations { ι }
//step 1: compute the initial tensor and decompose geometric features in different dimensions
FOREACHpinP //for each point in P
Ω { p i P | n o r m ( p i p ) r } = Neighborhood(p, r);
Tp = TensorVoting(p, Ω ); //compute initial tensor based on Equations (1)–(3)
{ e , λ } p = GeoFeatureDec(Tp); //eigen decomposition for tensor Tp
END
//step 2: revote tensor and compute refined geometric feature
FOREACHpinP //for each point in P
FOREACHdinN //for each dimension in N, based on Equations (4)–(7)
sd = SaliencyInDimD( { e , λ } p ); //compute dimensional saliency
Td = TensorInDimD( { e , λ } p ); //compute dimensional tensor
END
Tp = AggTensor (sd, Td); //aggregate tensor in each dimension based on Equation (8)
T = RevoteTensorFromNeig( T p Ω ); //refine the voting result based on Equation (9)
END
//step 3: represent geometric feature and construct the LCS
s { s s u r f a c e , s l i n e , s p o i n t } = DimSaliency(T); //compute dimensional saliency for each point
L C S = LineCandidateSubset( s , σ ); //compute the LCS based on Equation (10)
//step 4: extract line structure based on the discrete Morse theory.
{ L C S , S } = ResampleLCS(P, κ , L C S ); //resample the LCS based on Equation (11)
LinePer = MorseSmaleComplex( L C S , S ); //compute line structure based on Equation (12)
LineSeg { p c r i t i c a l , l c o n n e c t i o n } = LineExtract(LinePer, δ ); //extract line segment using threshold δ
LineStr { ι } = BuildLineStructure(LineSeg); //build line structure
RETURNLine Structure { ι }

4. Experiments and Discussions

This section focuses on the performance of line structure extraction method. The point cloud dataset acquired by the iPhone-based LiDAR sensor is collected, and tensor voting is conducted to get the LCS. Then, the line structure is extracted using the persistent line feature based on the discrete Morse theory. Comparisons are conducted and assessed with different methods. In addition, the line structure in a complicated scene is computed and evaluated.

4.1. The Dataset Collected by the iPhone-Based LiDAR Sensor

The point cloud dataset is acquired by the LiDAR sensor assembled in the iPhone 12 Pro MAX, which is a kind of consumer electronic device. It is a solid-state device based on the time of flight (referred to as TOF) technology, with a scanning range of 5 m and absolute accuracy of 0.01 m, as depicted in Figure 4. In addition, the mean square error in an indoor area of 1.49 m2 is 0.0018 m [6]. The point cloud acquired by the LiDAR sensor is a scene of parterre, in an area of 7.76 m2. To deal with the irregularly distributed sampling density, the dataset is randomly resampled with minimum distance of 0.01 m, and the number of points in the dataset is 23,149, with the average number of 2400 points per m2. As seen from Figure 4, the parterre is located on the plane and is consisted of four vertical sub-places and one flat rectangle ring plane, and the center plane inside the rectangle ring plane is a bare soil ground with uneven surfaces. Besides, there are noisy points and different point densities with unstable qualities in the randomly sampled point cloud data.

4.2. Tensor Voting and Geometric Dimension Representation

The point cloud P is supposed to be the 3D point with only coordinate information as (x, y, z). Hence, the initial tensor voting is conducted for each point p with its neighborhood. Then, the normal vector in each dimension is computed for the space S, and the tensor in normal subspace and tangent subspace are represented. In addition, the searching distance for the neighborhood r is set to be 0.3 m, and the minimum number of points involved in the tensor voting process is 5. In the second round of tensor voting, also called tensor refinement, the geometric feature in each dimension is encoded and revoted by its neighborhood. With the refined tensor, the geometric feature is recomputed, and different dimensional geometric feature is labeled with the saliency s s u r f a c e , s l i n e , s p o i n t , which can be taken as the attribute of each point, i.e., p ( x , y , z , s s u r f a c e , s l i n e , s p o i n t ) , depicted in Figure 5.
As seen in Figure 5, saliencies of different dimensional geometric features encoded by the tensor are shown by the colormap ranging from light to dark. In Figure 5a, the surface feature of each point is labeled from black to white, and the lower the surface saliency, the darker the color, and vice versa. However, in Figure 5b, the line saliency of each point is labeled in the opposite way, as well as the point saliency in Figure 5c.

4.3. The LCS Construction and Resampling

It can be obviously observed that the point with high line saliency in Figure 5a is considered to show a line structure, while the point with low surface saliency in Figure 5a also turns out to show a line structure. Besides, the point with high point saliency in Figure 5c seems to be the key point of show line structure, although some noisy points also have high point saliencies. Hence, the LCS can be constructed based on the threshold values { σ s u r f a c e , σ l i n e , σ p o i n t } . According to the saliency of the computation result, σ s u r f a c e is set to be 3.27, and the point with saliency lower than the threshold is selected. However, σ l i n e is set as 3.56 and σ p o i n t is set as 6.33, and the point with saliency higher than the threshold is selected. With the threshold of different geometric features, the LCS in different dimension is computed, and the statistic are depicted in Table 1.
As seen from Table 1, the number of the LCS point computed using s s u r f a c e , s l i n e , and s p o i n t is 89, 3326, and 982, respectively, and the number of ground truth LCS computed using the spatial relation (within the distance of 0.05 m) between the ground truth line structure and the point cloud, is 4006. Only using single geometric feature, the LCS cannot cover the entire line structure. In addition, as depicted in the fourth row of Table 1, the number of the LCS without duplication is 4155, which is a little higher than that of the ground truth LCS. Hence, although there might be redundancy, it is reasonable to select the LCS based on Equation (10). Finally, the LCS is constructed, as shown in Figure 6.
To deal with unstable quality point cloud and unevenly distributed point density, the LCS is further resampled as the regular grid point of the space S. The more complex the scene, the higher the grid size κ to preserve the detailed line structure, and vice versa. However, it is not a good idea to set κ with a very high value since it may cause broken line segments in the result. For example, if grid size κ is set to be too high such that gird interval is smaller than the minimum distance between two neighboring points, there will be vacant grids and the consecutive line segment will become broken. Hence, the suggested ranging domain is [10, 100]. With grid size κ = 45, the point in LCS is resampled and becomes a regularly distributed LCS’ in the space S. Besides, the computation complexity for the line structure extraction is also simplified and refined.

4.4. Comparisons with Other Line Structure Extraction Methods

With the LCS’ computed in Section 4.3, the Morse–Smale complex is constructed based on the discrete Morse theory, and the persistence of complex structures is computed. The important geometric structure will keep a high persistence value during the filtration of the Morse–Smale complex structure. Moreover, the persistence of 1-stable manifold is the curve structure for the LCS’. Hence, the higher the persistence, the more important the line segment. Finally, line segment { p c r i t i c a l , l c o n n e c t i o n } is extracted from the LCS’ based on the persistence threshold δ . Then, a graph is built and line structure { ι } is computed using the line segment and connection relations, as depicted in Figure 7.
As shown in Figure 7, the ground truth line structure is shown in Figure 7a, and it’s manually drawn from the original dataset. Figure 7b shows the result constructed based on the point cloud segmentation method [23] (with parameter {Max angle: 20°, Max relative distance: 1.0 m, Max distance: 0.2 m, Min points per facet: 10, Max edge length 0.03 m}), and the line structure is extracted using the rim of segmented geometric objects, while the result does not seem reasonable when compared with Figure 7a, since there are many mistakes in the extracted line structure caused by the unstable quality of point cloud dataset. Besides, using different persistence threshold δ , the related line structure is extracted from the LCS’ based on the discrete Morse theory. Moreover, the result is compared with the ground truth line structure and the 3D point cloud segmentation-based line extraction method. In Figure 7c–e, persistence thresholds of 0.01, 6, and 15 are applied to each result, respectively. Although there are more details in Figure 7c, the result seems not good as there are many over-connections in the line structure. However, it does not mean that “the lower the threshold δ , the better the result.” In Figure 7e, there can be observed many misconnected line segments with a high threshold δ . Fortunately, the threshold δ can be interactively set through a trivial operation, and it performs best in the result shown in Figure 7d with threshold δ = 6. Among the results, the line structure extracted based on the proposed method performs superior to others.
The statistic of extracted line structure is conducted, as depicted in Table 2. Line extraction results in Figure 7 are quantitatively compared with the ground truth line structure, and only the line segment within the buffer area (0.05 m) of ground truth line structure is considered to be the effective line structure. The total length of the extracted and effective line structure is calculated. The length of line structure extracted by each method, ranging from Figure 7b to Figure 7e, are 186.94 m, 30.47 m, 24.49 m, and 24.46 m, respectively, while effective lengths is 33.50 m, 23.79 m, 21.39 m, and 21.36 m. Although it has a longer effective length in Figure 7b than the others, there are too many mistakes in the result. In addition, compared with the result in Figure 7d, there are some over-connected line segments in Figure 7c and some misconnected line segments in Figure 7e. Hence, the result of proposed method holds the best performance, and 87.33% of line segments have been effectively extracted from the result. What’s more, the connection relation of line structure is preserved, based on the topologic feature of the proposed method.

4.5. Line Structure Extraction in the Complex Area

To make a further assessment of the capability for the proposed method, a complex area of the indoor stair scene is collected by the iPhone-based LiDAR sensor, and experiments are conducted. The stair dataset is in an area of 10.61 m2. To deal with the irregularly distributed sampling density, the dataset is randomly resampled with minimum distance of 0.01 m, and the number of point in the dataset is 141,837, with the average number of 2400 points per m2. As seen from Figure 8a, there are at least 29 facets in different directions. In addition, some structures are incomplete, and the point density is unevenly distributed, which makes it difficult to extract line structures from the quality-unstable dataset.
With neighborhood searching distance r = 0.1 m, the initial tensor voting process is conducted, followed by the tensor refinement. Then, the geometric feature of each point is computed based on the refined tensor. In Figure 8b–d, the saliency of surface, line, and point of each point is labeled with the color ranging from light to dark. In Figure 8b, the darker color with the low saliency of “surface” is considered to be the line structure, while the darker color with high saliency of “line” and “point” in Figure 8c,d shows a high probability of being the line structure. Hence, the LCS is computed using threshold values { σ s u r f a c e = 18 , σ l i n e = 14 , σ p o i n t = 27 } , as depicted in Table 3.
As seen in Table 3, the point with surface saliency s s u r f a c e lower than σ s u r f a c e , or line saliency s l i n e higher than σ l i n e , or point saliency s p o i n t higher than σ p o i n t , is selected and taken into the LCS. Besides, the number of selected points by σ s u r f a c e , σ l i n e , and σ p o i n t and is 1225, 20,137, and 220, respectively, and the total point number without duplication is 21,106. On the other hand, the ground truth LCS is computed using the spatial relation (within the distance of 0.05 m) between the ground truth line structure and the point cloud, and the total number of points is 48,307. To deal with the quality-unstable point cloud, the LCS is resampled, and the related LCS’ is constructed with different grid sizes κ , as depicted in Figure 9.
In Figure 9, the LCS’ is computed with different grid sizes κ , ranging from 50 to 300. In Figure 9a, it’s the constructed LCS based on Equation (10). In Figure 9b, with κ = 64, the details are not properly preserved. While in Figure 9d, the connectivity of line structure can be broken with the high grid size κ = 256, since the space can be over-divided. To make a balance between the details and the connectivity of line structure, the grid size κ is set as 128. Then, the line structure is extracted from the LCS’ based on the discrete Morse theory, using different persistence thresholds δ , and the result is compared with the ground truth line structure, along with the 3D point cloud segmentation-based line extraction method, as depicted in Figure 10.
Figure 10a shows the ground truth line structure, and it is manually drawn from the original dataset. Figure 10b shows the result computed by the point cloud segmentation-based method [23] (with parameter {Max angle: 10°, Max relative distance: 2.0 m, Max distance: 0.1 m, Min points per facet: 10, Max edge length 0.07 m}), and the line structure is extracted using the rim of segmented geometric objects. In Figure 10c–e, line segments { p c r i t i c a l , l c o n n e c t i o n } are extracted from the LCS’ with different persistence threshold values δ , ranging from 0.01 to 10, and related line structures { ι } are computed. A low or high threshold δ will cause the overconnection or misconnection problems in the extracted line structure, as depicted in Figure 10c–e. Here, the optimal threshold δ = 0.1 is interactively set through a trivial operation, as depicted in Figure 10d.
The related statistics of extracted line structure in the complex area are depicted in Table 4. The total length and effective length of extracted line structure are compared with the ground truth line structure, and only the line segment within buffer area (0.05 m) of the ground truth line segment is considered to be the effective line structure. The length of the ground truth line structure in Figure 10a is 53.98 m, and the total length of line extraction results in Figure 10b–e is 289.58 m, 67.92 m, 67.36 m, and 58.28 m, respectively. However, the effective length of each result is 156.26 m, 56.99 m, 57.77 m, and 43.80 m. Although it has a longer length than others in Figure 10b, there are too many misconnections. With a low threshold value δ , there can be over-connections in Figure 10c, while there can be misconnections in Figure 10e using a high threshold value δ . Among conducted experiments, the line structure extracted by the proposed method (in Figure 10d) outperforms the others. Hence, with 85.76% of line segments effectively extracted, the experiment result in Figure 10d holds the best performance, and it turns out that the proposed method can extract the line structure from the dataset with proper geometric and topological features. In addition, the rate of effectively extracted line segments among experimented datasets remains steady (87.33% for the scene of parterre, 85.76% for the scene of indoor stair), and it turns out to be an effective method to extract line structures in different scenarios.

5. Conclusions

With the widespread applications of LiDAR sensors in consumer electronic devices, point cloud datasets in different scenarios are being collected and numerous studies conducted to interpret the 3D scenes and obtain effective representation. Among these approaches is line structure extraction. To deal with the quality-unstable point cloud datasets acquired by consumer electronic devices, the line structure extraction method based on the persistence of tensor feature is applied. The point is encoded and voted to its neighborhood, different geometric features ranging from one dimension to three dimensions are extracted from the voted tensor, and the LCS is constructed based on the combination of various features. Then, the Morse–Smale complex is constructed according to the tensor feature, and line structure is extracted using the persistent homology theory. With the point cloud dataset collected by the iPhone-based LiDAR sensor, experiments are conducted, line structure is successfully extracted from the dataset with connection information preserved, and results are compared with related methods. Moreover, using the proposed method, line structure can be extracted from the quality-unstable point cloud dataset, with no predefined geometric models and no manually selected training datasets.
Future research should aim for the refinement of line structure extraction results, since lines are extracted based on the persistence of tensor feature, and they are not purely straight compared with the ground truth dataset. Another aspect that needs to be further explored is improving the robustness of the line structure extraction framework and transforming it into an end-to-end line structure extraction model.

Author Contributions

X.W. performed the theory analysis, methodology, performed the experiments, and contributed to drafting the manuscript. H.L. collected and analyzed the data, design, and coding. W.H. performed the literature reviews, improved the writing. Q.C. revised the paper, provided the background knowledge and funding. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (NSFC) (No. 61875088, No. 62005128).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Publicly available datasets were analyzed in this study. The data can be found here: https://doi.org/10.6084/m9.figshare.20407797 (accessed on 20 August 2022).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wang, C.; Hou, S.; Wen, C.; Gong, Z.; Li, Q.; Sun, X.; Li, J. Semantic line framework-based indoor building modeling using backpacked laser scanning point cloud. ISPRS J. Photogramm. Remote Sens. 2018, 143, 150–166. [Google Scholar] [CrossRef]
  2. Mallya, A.; Lazebnik, S. Learning informative edge maps for indoor scene layout prediction. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 936–944. [Google Scholar]
  3. Zhang, W.; Zhang, W.; Gu, J. Edge-semantic learning strategy for layout estimation in indoor environment. IEEE Trans. Cybern. 2019, 50, 2730–2739. [Google Scholar] [CrossRef] [PubMed]
  4. Lin, Y.; Wang, C.; Chen, B.; Zai, D.; Li, J. Facet segmentation-based line segment extraction for large-scale point clouds. IEEE Trans. Geosci. Remote Sens. 2017, 55, 4839–4854. [Google Scholar] [CrossRef]
  5. Zhang, J.; Zhao, X.; Chen, Z.; Lu, Z. A Review of Deep Learning-Based Semantic Segmentation for Point Cloud. IEEE Access 2019, 7, 179118–179133. [Google Scholar] [CrossRef]
  6. Wang, X.; Lyu, H.; Mao, T.; He, W.; Chen, Q. Point Cloud Segmentation from iPhone-Based LiDAR Sensors Using the Tensor Feature. Appl. Sci. 2022, 12, 1817. [Google Scholar] [CrossRef]
  7. Guo, Y.; Wang, H.; Hu, Q.; Liu, H.; Liu, L.; Bennamoun, M. Deep Learning for 3d Point Clouds: A Survey. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 43, 4338–4364. [Google Scholar] [CrossRef]
  8. Zheng, B.; Zhao, Y.; Yu, J.C.; Ikeuchi, K.; Zhu, S.C. Beyond Point Clouds: Scene Understanding by Reasoning Geometry and Physics. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Portland, OR, USA, 23–28 June 2013; pp. 3127–3134. [Google Scholar]
  9. Yang, B.; Dong, Z. A shape-based segmentation method for mobile laser scanning point clouds. ISPRS J. Photogramm. Remote Sens. 2013, 81, 19–30. [Google Scholar] [CrossRef]
  10. Dittrich, A.; Weinmann, M.; Hinz, S. Analytical and numerical investigations on the accuracy and robustness of geometric features extracted from 3D point cloud data. ISPRS J. Photogramm. Remote Sens. 2017, 126, 195–208. [Google Scholar] [CrossRef]
  11. Su, H.; Maji, S.; Kalogerakis, E.; Learned-Miller, E. Multi-View Convolutional Neural Networks for 3d Shape Recognition. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 11–18 December 2015; pp. 945–953. [Google Scholar]
  12. Schnabel, R.; Wahl, R.; Klein, R. Efficient RANSAC for Point-Cloud Shape Detection. In Computer Graphics Forum; Blackwell Publishing Ltd.: Oxford, UK, 2007; Volume 26, pp. 214–226. [Google Scholar]
  13. Hulik, R.; Spanel, M.; Smrz, P.; Materna, Z. Continuous plane detection in point-cloud data based on 3D Hough Transform. J. Vis. Commun. Image Represent. 2014, 25, 86–97. [Google Scholar] [CrossRef]
  14. Bazazian, D.; Casas Pla, J.R.; Ruiz Hidalgo, J. Segmentation-based multi-scale edge extraction to measure the persistence of features in unorganized point clouds. In Proceedings of the 12th International Joint Conference on Computer Vision, Imaging and Computer Graphics Theory and Applications, Porto, Portugal, 27 February–1 March 2017; pp. 317–325. [Google Scholar]
  15. Park, M.K.; Lee, S.J.; Lee, K.H. Multi-scale tensor voting for feature extraction from unstructured point clouds. Graph. Model. 2012, 74, 197–208. [Google Scholar] [CrossRef]
  16. King, B.J. Range Data Analysis by Free-Space Modeling and Tensor Voting; Rensselaer Polytechnic Institute: Troy, NY, USA, 2008. [Google Scholar]
  17. Tang, C.-K.; Medioni, G. Curvature-augmented tensor voting for shape inference from noisy 3D data. IEEE Trans. Pattern Anal. Mach. Intell. 2002, 24, 858–864. [Google Scholar] [CrossRef]
  18. Zomorodian, A.; Carlsson, G. Computing persistent homology. Discret. Comput. Geom. 2005, 33, 249–274. [Google Scholar] [CrossRef]
  19. Carlsson, G. Topological pattern recognition for point cloud data. Acta Numer. 2014, 23, 289–368. [Google Scholar] [CrossRef]
  20. Zhou, C.; Dong, Z.; Lin, H. Learning persistent homology of 3D point clouds. Comput. Graph. 2022, 102, 269–279. [Google Scholar] [CrossRef]
  21. Chen, T.; Wang, Q. 3d line segment detection for unorganized point clouds from multi-view stereo. In Proceedings of the Asian Conference on Computer Vision, Queenstown, New Zealand, 8–12 November 2010; Springer: Berlin/Heidelberg, Germany, 2010; pp. 400–411. [Google Scholar]
  22. Lu, X.; Liu, Y.; Li, K. Fast 3D line segment detection from unorganized point cloud. arXiv 2019, arXiv:1901.02532. [Google Scholar]
  23. Dewez, T.J.B.; Girardeau-Montaut, D.; Allanic, C.; Rohmer, J. Facets: A Cloudcompare Plugin to Extract Geological Planes from Unstructed 3D Point Clouds. In Proceedings of the International Archives of the Photogrammetry, Remote Sensing & Spatial Information Sciences, Prague, Czech Republic, 12–19 June 2016; pp. 799–804. [Google Scholar]
  24. Gankhuyag, U.; Han, J.H. Automatic 2d floorplan cad generation from 3d point clouds. Appl. Sci. 2020, 10, 2817. [Google Scholar] [CrossRef]
  25. Guo, J.; Wu, L.; Zhang, M.; Liu, S.; Sun, X. Towards automatic discontinuity trace extraction from rock mass point cloud without triangulation. Int. J. Rock Mech. Min. Sci. 2018, 112, 226–237. [Google Scholar] [CrossRef]
  26. Guo, B.; Li, Q.; Huang, X.; Wang, C. An improved method for power-line reconstruction from point cloud data. Remote Sens. 2016, 8, 36. [Google Scholar] [CrossRef]
  27. Demantké, J.; Mallet, C.; David, N.; Vallet, B. Dimensionality Based Scale Selection in 3D Lidar Point Clouds. In Proceedings of the ISPRS Workshop on Laser Scanning, Calgary, AB, Canada, 29–31 August 2011; pp. 29–31. [Google Scholar]
  28. Wei, M.; Liang, L.; Pang, W.M.; Wang, J.; Li, W.; Wu, H. Tensor voting guided mesh denoising. IEEE Trans. Autom. Sci. Eng. 2016, 14, 931–945. [Google Scholar] [CrossRef]
  29. Schuster, H.F. Segmentation of LiDAR data using the tensor voting framework. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2004, 35, 1073–1078. [Google Scholar]
  30. Sousbie, T. The persistent cosmic web and its filamentary structure–I. Theory and implementation. Mon. Not. R. Astron. Soc. 2011, 414, 350–383. [Google Scholar] [CrossRef]
  31. Sousbie, T.; Pichon, C.; Kawahara, H. The persistent cosmic web and its filamentary structure–II. Illustrations. Mon. Not. R. Astron. Soc. 2011, 414, 384–403. [Google Scholar] [CrossRef] [Green Version]
  32. Tierny, J.; Favelier, G.; Levine, J.A.; Gueunet, C.; Michaux, M. The topology toolkit. IEEE Trans. Vis. Comput. Graph. 2017, 24, 832–842. [Google Scholar] [CrossRef] [PubMed]
  33. Wong, C.C.; Vong, C.M. Persistent Homology Based Graph Convolution Network for Fine-Grained 3D Shape Segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 7098–7107. [Google Scholar]
  34. Beksi, W.J.; Papanikolopoulos, N. A topology-based descriptor for 3D point cloud modeling: Theory and experiments. Image Vis. Comput. 2019, 88, 84–95. [Google Scholar] [CrossRef]
Figure 1. The pipeline of the line structure extraction framework.
Figure 1. The pipeline of the line structure extraction framework.
Applsci 12 09190 g001
Figure 2. The tensor voting process of the point with its neighborhood.
Figure 2. The tensor voting process of the point with its neighborhood.
Applsci 12 09190 g002
Figure 3. 1-stable manifold of the critical point.
Figure 3. 1-stable manifold of the critical point.
Applsci 12 09190 g003
Figure 4. The iPhone-based LiDAR sensor and the collected point cloud dataset.
Figure 4. The iPhone-based LiDAR sensor and the collected point cloud dataset.
Applsci 12 09190 g004
Figure 5. The tensor voting results in different geometric dimensions.
Figure 5. The tensor voting results in different geometric dimensions.
Applsci 12 09190 g005
Figure 6. The LCS and LCS’ with different t grid sizes κ of the point cloud.
Figure 6. The LCS and LCS’ with different t grid sizes κ of the point cloud.
Applsci 12 09190 g006
Figure 7. The comparison of different line structure extraction results.
Figure 7. The comparison of different line structure extraction results.
Applsci 12 09190 g007
Figure 8. The point cloud in the complex area and the geometric saliency in different dimensions.
Figure 8. The point cloud in the complex area and the geometric saliency in different dimensions.
Applsci 12 09190 g008
Figure 9. The LCS and LCS’ with different grid sizes κ in the complex area.
Figure 9. The LCS and LCS’ with different grid sizes κ in the complex area.
Applsci 12 09190 g009
Figure 10. The comparison of different line structure extraction results in the complex area.
Figure 10. The comparison of different line structure extraction results in the complex area.
Applsci 12 09190 g010
Table 1. The statistic of LCS by different geometric features.
Table 1. The statistic of LCS by different geometric features.
Different Geometric FeatureThe Number of Points
LCS by ssurface86
LCS by sline3326
LCS by spoint982
LCS without duplication4155
ground truth LCS4006
Table 2. The statistic of extracted line structure.
Table 2. The statistic of extracted line structure.
Different ResultsTotal Length (m)Effective Length (m)
Figure 7a17.1617.16
Figure 7b186.9433.50
Figure 7c30.4723.79
Figure 7d24.4921.39
Figure 7e24.4621.36
Table 3. The statistic of LCS by different geometric features in the complex area.
Table 3. The statistic of LCS by different geometric features in the complex area.
Different Geometric FeatureThe Number of Points
LCS by ssurface1225
LCS by sline20,137
LCS by spoint220
LCS without duplication21,106
ground truth LCS48,307
Table 4. The statistic of extracted line structure in the complex area.
Table 4. The statistic of extracted line structure in the complex area.
Different ResultsTotal Length (m)Effective Length (m)
Figure 10a53.9853.98
Figure 10b289.58156.26
Figure 10c67.9256.99
Figure 10d67.3657.77
Figure 10e58.2843.80
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Wang, X.; Lyu, H.; He, W.; Chen, Q. Line Structure Extraction from LiDAR Point Cloud Based on the Persistence of Tensor Feature. Appl. Sci. 2022, 12, 9190. https://doi.org/10.3390/app12189190

AMA Style

Wang X, Lyu H, He W, Chen Q. Line Structure Extraction from LiDAR Point Cloud Based on the Persistence of Tensor Feature. Applied Sciences. 2022; 12(18):9190. https://doi.org/10.3390/app12189190

Chicago/Turabian Style

Wang, Xuan, Haiyang Lyu, Weiji He, and Qian Chen. 2022. "Line Structure Extraction from LiDAR Point Cloud Based on the Persistence of Tensor Feature" Applied Sciences 12, no. 18: 9190. https://doi.org/10.3390/app12189190

APA Style

Wang, X., Lyu, H., He, W., & Chen, Q. (2022). Line Structure Extraction from LiDAR Point Cloud Based on the Persistence of Tensor Feature. Applied Sciences, 12(18), 9190. https://doi.org/10.3390/app12189190

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop