Next Article in Journal
Three-Dimensional Surface Motion Displacement Estimation of the Muz Taw Glacier, Sawir Mountains
Previous Article in Journal
Radar Moving Target Detection Based on Small-Sample Transfer Learning and Attention Mechanism
Previous Article in Special Issue
A Base-Map-Guided Global Localization Solution for Heterogeneous Robots Using a Co-View Context Descriptor
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Optimal Feature-Guided Position-Shape Dual Optimization for Building Point Cloud Facade Detail Enhancement

1
College of Surveying and Geo-Informatics, North China University of Water Resources and Electric Power, Zhengzhou 450046, China
2
College of Geomatics and Geoinformation, Guilin University of Technology, Guilin 541004, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2024, 16(22), 4324; https://doi.org/10.3390/rs16224324
Submission received: 7 September 2024 / Revised: 15 November 2024 / Accepted: 17 November 2024 / Published: 20 November 2024

Abstract

:
Dense three-dimensional point clouds are the cornerstone of modern architectural 3D reconstruction, containing a wealth of semantic structural information about building facades. However, current methods struggle to automatically and accurately extract the complex detailed structures of building facades from unstructured point clouds, with detailed facade modeling often relying heavily on manual interaction. This study introduces an efficient method for semantic structural detail enhancement of building facade point clouds, achieved through feature-guided dual-layer optimization of position and shape. The proposed framework addresses three key challenges: (1) robust extraction of facade semantic feature point clouds to effectively perceive the underlying geometric features of facade structures; (2) improved grouping of similarly structured objects using Hausdorff distance discrimination, overcoming the impact of point cloud omissions and granularity differences; (3) position-shape double optimization for facade enhancement, achieving detailed structural optimization. Validated on three typical datasets, the proposed method not only achieved 98.5% accuracy but also effectively supplemented incomplete scan results. It effectively optimizes semantic structures that widely exist and have the characteristic of repeated appearance on building facades, providing robust support for smart city construction and analytical applications.

1. Introduction

In the context of contemporary urban informatization and digitalization, the precise extraction of building facade structures holds significant value in areas such as urban planning, heritage conservation, disaster assessment, and smart city development [1,2,3,4]. With the rapid advancement of 3D laser scanning technology (LiDAR), it is now possible to efficiently obtain high-precision 3D point cloud data of buildings, enabling the automatic extraction of building facade structures [5]. However, due to factors such as occlusion during the scanning process, variations in reflectivity, and limitations in instrument accuracy, the acquired point cloud data often suffer from issues such as incompleteness, noise, and structural discontinuities, posing challenges to the accurate extraction of building facade structures.
In the extraction of building facade structures, researchers have proposed various methods. These methods are primarily based on the geometric properties of point clouds, such as curvature, normals, and distance metrics [6], and use algorithms like region growing, Random Sample Consensus (RANSAC), and graph cuts to identify and reconstruct building facade structures [7]. With the development of deep learning technology, learning-based methods have also been gradually introduced into this field [8,9,10], leveraging trained datasets to identify and extract facade structure features. While these methods have improved the efficiency and accuracy of extraction to some extent, they still face limitations in handling large-scale data, complex scenes, and missing data issues. Current research faces several core bottlenecks: First, existing methods rely heavily on the quality and completeness of point cloud data, with insufficient handling of missing data and noise generated during the scanning process. Second, the geometric primitives derived from complex feature descriptions still struggle to achieve object-level correction. These structures make object-level correction more difficult. Due to limitations such as the uneven sampling of facade point clouds, variations in reflective materials, and occlusions or missing data, it is challenging to extract robust contour structures from individual window targets. Our approach treats each window as an independent object, incorporating consistent discrimination of objectification and mutual optimization among objects, resulting in improved structural extraction outcomes. We emphasize optimizing the spatial structural logic of semantic objects after extracting structural primitives such as points and lines. Third, the increasing complexity of urban architecture has constrained the generalization capabilities of existing algorithms.
In response to these challenges, we propose a novel point cloud feature description and a dual-layer optimization method for building facade semantic structure extraction based on optimal feature guidance. This method addresses low-quality or occluded facades by performing object-level extraction and inference from both positional and shape perspectives. The method achieves the task of identifying and optimizing the extraction of semantic objects on building facades through three coherent steps: feature screening, similar type recognition, and double-layer optimization. The proposed approach is designed for extracting the facade structures of common high-rise buildings in urban environments and automatically generates accurate spatial boundaries of semantic structure objects. The resulting facade layouts can be directly applied to the enhancement and reconstruction of LOD1/2 (the level of detail) models [11,12,13], achieving high accuracy and applicability. This method provides innovative technological support for urban informatization and digitalization efforts.
The main contributions of this work are as follows:
(1) A novel point cloud feature description and feature point extraction method has been proposed to identify the semantic information of the facade, namely the window structure. Our method incorporates a dynamic neighborhood search mechanism that significantly amplifies the distinction between feature points and non-feature points, thereby enabling a more accurate alignment with the underlying geometric edges of building facades and has been verified in three datasets with defects and occlusions.
(2) We have developed an approach based on the improved Hausdorff distance criterion for effectively identifying objects with similar structures from raw data. This algorithm effectively overcomes the issues of missing data and granularity differences in the scanning of objects with similar structures and achieves global optimization of semantic object grouping through matrix analysis.
(3) We have implemented a dual-layer optimization algorithm for facade semantic structure enhancement, which takes into account both positional and shape aspects. By fully leveraging the spatial and morphological similarities of semantic objects on building facades, and effectively addressing the incompleteness caused by point cloud data and noise interference, the accuracy and efficiency of 3D reconstruction can be significantly improved.
In the following sections, Section 2 introduces the existing related work that has been conducted in this field, providing a foundation for our research. Section 3 provides a comprehensive description of a semantic detail enhancement framework for building facade point clouds, delving into the core innovative aspects from three perspectives. Section 4 outlines the entire experimental process and conducts a thorough evaluation, including an assessment of accuracy. Lastly, in Section 5, we present a comprehensive discussion and the conclusions of this paper.

2. Related Works

Unlike most methods that focus on building roofs, the extraction of building facades presents unique challenges due to the presence of multiple heterogeneous structures and occlusions during data collection, particularly issues such as uneven point cloud density caused by TLS (Terrestrial Laser Scanning). To enhance the semantic details of facades, it is crucial to detect window targets from the extracted features of the building. Therefore, this paper briefly reviews the semantic enhancement of building facade point clouds from two perspectives: (1) point cloud feature extraction and (2) facade structure extraction.

2.1. Point Cloud Feature Extraction

As an unstructured form of 3D data, point clouds contain rich spatial information. However, due to their large volume, lack of order, and absence of topological structure, direct processing and analysis of point clouds present numerous challenges. Consequently, effective feature extraction has become a critical research topic in the field of point cloud processing.
Compared to early methods that focused on statistical values of point cloud geometric features (such as point distribution, normals, and curvature) [14,15,16,17], recent research has increasingly turned to geometric feature extraction methods to better utilize the spatial structural information of point clouds. For example, due to the complexity of handling point clouds in three-dimensional space, some researchers have suggested extracting features in two-dimensional space. By projecting point clouds onto a 2D plane, high-precision feature points can be extracted using the high-accuracy feature line intersections available in 2D space [18]. Although this method achieves high precision in feature extraction, it inevitably leads to occlusions and data loss during the projection process, particularly when dealing with point clouds with complex structures.
When a single geometric feature is insufficient for handling complex extraction tasks, methods combining multiple geometric features (such as curvature, normals, and point distances) have been proposed. These methods use the cross-entropy of multiple geometric features to select relevant features [19]. Compared to using only one geometric feature, this approach improves accuracy; however, it results in longer processing times and lower efficiency when dealing with large-scale data.
To robustly address multi-dimensional unstructured point clouds, a new point cloud feature descriptor called Spherical Shell Point Feature (SSPF) was introduced, which combines point coordinates to generate embedded features [20]. This approach considers both point coordinates and spatial alignment, but it faces challenges in handling point clouds in urban scenes.
Deep learning methods, by automatically learning complex patterns within data, are capable of extracting more expressive features from large-scale point cloud data [21,22,23,24]. Methods that combine multi-scale features with spectral information can distinguish the semantic information of point clouds. By filtering multi-scale point cloud features using a random forest approach, and incorporating spectral information, the constructed point cloud features are then fused with local feature aggregation modules in the network to learn deeper semantic information [25]. This approach not only enables feature extraction but also teaches semantic information; however, it is limited in handling object-level tasks, such as window detection.
The novel residual network module introduces smaller residual blocks within the residual units to achieve a finer-grained multi-scale feature representation. It integrates Res2Net with a cascade structure to extract multi-scale features [26]. This method considers the geometric features of point clouds across multiple scales and expands the receptive field size of each network layer. While Res2Net enhances the receptive field and multi-scale feature representation capabilities by constructing hierarchical residual connections within residual units, it simultaneously increases the complexity of the network architecture, making network design and debugging more challenging. Although there are numerous feature extraction methods, most are not well-suited for handling point clouds of building facades.

2.2. Facade Structure Detection

Existing research on building modeling primarily focuses on the extraction and reconstruction of the main structure and roof contours of buildings [27,28,29,30,31,32], while relatively little attention has been paid to the semantic information of building facades, particularly the extraction and analysis of windows. This limitation, to some extent, affects the accuracy of the models and their effectiveness in practical applications. Existing methods apply ground filtering to facades [33] and enhance traditional mature methods by constructing new semantic segmentation networks [34]. These approaches address the problem by improving traditional mature methods and exploring solutions from the perspective of multi-module coupling networks. However, the former requires careful parameter selection, while the latter must address the issue of incomplete point clouds. Facade semantic features such as doors and windows are reflected in fine structures on one hand, and on the other hand, they can also be easily identified in color textures, where multi-objective functions have been used to select optimal facade texture images. Geometric correction, color balancing, and texture restoration are also considered to correct distortions, color inconsistencies, and occlusions by other objects (such as trees), thereby producing realistic 3D models of building facades [35]. Although this method can achieve visually realistic 3D models, the strong texture but weak structural nature of the image data makes it difficult to accurately restore the true geometric structure.
The geometric and topological relationships of semantic entities can be used to reconstruct complex facade models. By formalizing the geometric and topological relationships of window semantic entities on building facades through facade layout models, it is possible to automatically develop semantic building models using the extracted window parameter sets [36]. While this method ensures consistency between geometry, semantics, and topology in the reconstructed building models, it is limited in handling subtle spatial positional differences. Utilizing the rich radiative information in laser point clouds can enhance the capability of point cloud segmentation. A method based on the fusion of point cloud radiance intensity with images for road segmentation can effectively separate dry, damp, and snowy road surfaces and address the issues caused by data sparsity [37]. However, this method is targeted at roads and has limitations when it comes to window segmentation.
Due to the high time and labor costs of fully supervised learning, weakly supervised learning methods have been proposed. For instance, a weakly supervised building facade segmentation method using Spatially Adaptive Fusion with Consistency Contrast Constraint (SAF-C3) has been developed. This method employs a Spatially Adaptive Fusion (SAF) module to extract discriminative features for building facade point clouds [38]. Although weakly supervised learning reduces annotation costs and improves the generalization ability of models, segmentation at the dataset level fails to account for object-level features such as windows.
The challenges of window reconstruction also include issues related to incomplete building facades. To bridge the gap caused by noise, sparsity, and occlusion, and to generate high-fidelity 3D building models, APC2Mesh was proposed. This method integrates the completed point cloud into the 3D reconstruction pipeline, enabling the learning of dense and geometrically accurate representations of buildings [39]. While APC2Mesh bridges the gap between neural point completion and surface reconstruction, and addresses challenges posed by occlusion, noise, and variable point densities when generating building models from ALS (Airborne Laser Scanning) point sets, it still suffers from minimal edge and corner smoothing issues when applied to complex or high-rise buildings. It is evident that while 3D urban reconstruction has become highly advanced, little work has been conducted on object-level semantic enhancement for building facades.

3. Methods

This paper introduces a novel approach for identifying and optimizing the semantic structure of windows from building facade point clouds. Specifically tailored for complex high-rise buildings commonly encountered in urban management, the method starts with local semantic recognition to derive an optimized window layout for the entire facade, thereby enhancing LOD1/2 models that lack detailed facade information. The approach consists of three key steps: (1) the robust extraction of semantic feature point clouds from the facade; (2) the optimization of object clustering based on the Hausdorff distance; and (3) facade enhancement through a dual-layer optimization of shape and position. Figure 1 provides an overview of this comprehensive framework.

3.1. Robust Extraction of Feature Point Clouds

The first step is to extract semantic structure points from the building facade’s point cloud data as accurately as possible, while excluding non-semantic structure points. Effective feature description is the foundation for mining high-precision information from point clouds. This study focuses on the facade point clouds of large urban buildings, which not only contain abundant planar structures but also exhibit significant variations in architectural contours and subtle differences in semantic structures such as windows and doors, presenting new challenges for feature description methods. Before proceeding with feature extraction, it is essential to preprocess the raw point cloud data using established algorithms, including filtering and classification techniques to extract the building point cloud and performing planar segmentation to isolate the building’s facade point cloud [36,37,38,39]. Based on this, the extraction of facade feature point clouds is achieved through the following two steps:

3.1.1. Optimal Neighborhood Computation

For three-dimensional point clouds with discrete distributions, the local structural information of points needs to be extracted from the spatial relationships of adjacent points. In the semantic feature description and recognition of facade point clouds, the challenge we face is to identify minor structural changes from numerous planar structures and overcome noise interference introduced during the point cloud acquisition process. To this end, we should construct a dynamic neighborhood search mechanism that adjusts based on the differences in the complexity of the local neighborhood structure surrounding a point. Shannon entropy, as a measure of the randomness or uncertainty of information, is commonly used to quantify information [40]. In point cloud analysis, it is often used to assess the amount of information contained within the neighborhood of a point cloud; less information indicates more uniform local features, which aids in dynamically adjusting the neighborhood settings of points.
The Shannon entropy facades are calculated as follows:
E i = ( δ 1 · ln δ 1 + δ 2 · ln δ 2 + δ 3 · ln δ 3 )
In the formula, E i represents the magnitude of Shannon entropy. ( δ 1 , δ 2 , δ 3 ) represent the probabilities of a point belonging to a linear point, a planar point, or is a scattered point, respectively. These three values can be determined using Principal Component Analysis (PCA). In brief, for each point p i in the point cloud and its neighborhood set V = p i | p i = x i , y i , z i T , PCA is used to fit the domain plane, resulting in the covariance matrix c o v p . After performing an eigenvalue decomposition, the eigenvalues ( λ 1 λ 2 λ 3 ) are obtained. Then, the saliency features ( δ 1 , δ 2 , δ 3 ) of point p i are calculated as Equation (2).
δ 1 = λ 1 1 / 2 λ 2 1 / 2 / λ 1 1 / 2 δ 2 = λ 2 1 / 2 λ 3 1 / 2 / λ 1 1 / 2 δ 3 = λ 3 1 / 2 / λ 1 1 / 2
It can be observed that changes in the neighborhood radius affect the value of Shannon entropy. A lower Shannon entropy indicates that the neighborhood of the target point contains less information and has more uniform structural features. This phenomenon guides us in the adaptive selection of the radius. Choosing the optimal radius based on the principle of minimizing Shannon entropy enhances the ability to recognize local subtle structures. It is important to note that a smaller radius does not necessarily mean a lower Shannon entropy, as the influence of noise points decreases with an increase in radius. Therefore, ensuring an adequate neighborhood range and the number of points is essential. We should set reasonable upper and lower limits for the search radius and combine the calculation of Shannon entropy to determine the optimal neighborhood search radius.
For any set of point cloud data, we first need to calculate the average point density ρ. By randomly selecting multiple seed points from the point cloud and calculating the average distance between these seed points and their neighboring points, we determine the point density ρ. We set the maximum value of the search radius to 10ρ and the minimum to 3ρ, and then calculate the optimal radius r according to Equation (3).
r = a r g r i [ r m i n , r m a x ] m i n E i r i

3.1.2. Semantic Feature Point Detection via Centroid Displacement

Point cloud data are complex and varied due to their rich geometric features, ranging from coarse structures to fine edges. Precisely identifying feature points directly from these discrete point clouds is a technical challenge [41,42,43,44]. To address this challenge, we propose a feature point detection scheme based on centroid displacement. Combined with the previously determined optimal neighborhood radius r, this ensures that these feature points can accurately match the underlying geometric edges.
Specifically, we first calculate point features using traditional methods, as shown in Figure 2a. For a point pi in the point cloud, we obtain its neighborhood point set p based on the optimal neighborhood radius r and fit its centroid as c i . We define the vector L i = c i p i , and the larger the magnitude of this vector L i , the more complex the local shape of its neighborhood. However, since L i is a uniform Laplacian vector, its weight strategy is insensitive to surface changes, which limits its effectiveness in describing complex and detailed facade semantic structures. As shown in Figure 2b, in flat areas and shallow structural regions, the vectors from points to the centroid are small and indistinguishable, whereas as shown in Figure 2c, under noise interference, the vectors from points to the centroid lengthen, potentially misidentifying pi as a feature point.
To enhance the distinction between feature points and non-feature points, and to mitigate the impact of noise on feature detection, we propose a method of locally weighted centroid displacement. Building on the calculation of vector L i for each neighboring point of point p i , we first create a temporary point m j (as shown by the green point in Figure 3b), which is used to construct the vector p m = p p + p c . Subsequently, we fit a new centroid d i using all these temporary points m j (as depicted in Figure 3c). We then construct a new centroid displacement vector ε i , which is calculated according to Equation (4).
ε i = 1 p j Ω ( p i ) ω ( i , j ) p j Ω ( p i ) ω ( i , j ) p i m j
Here, ω i , j = e x p ( p i p j 2 / r 2 ) , where Ω ( p i ) represents the set of K nearest neighboring points of point p i , and r denotes the average distance between p i and all its neighboring points. The locally weighted centroid displacement scheme enhances the differences between points in shallow features and those in flat areas by considering the relative positions of points to their neighbors.
Subsequently, the magnitude of the centroid displacement vector ε i for each point will be calculated and statistically analyzed, serving as a superior reference metric to replace L i . Based on the statistical outcomes, a reasonable threshold will be established to differentiate feature points from non-feature points. With these steps completed, the extraction of the facade’s semantic feature point cloud set U is finalized.

3.2. Improved Hausdorff Distance Discrimination

Considering that the facade windows of large urban buildings are often arranged in a regular pattern of identical or similar structural arrays, identifying and grouping windows with the same structure will help overcome data gaps and point density differences caused by scanning view limitations, thereby optimizing the extraction results of semantic structures. Therefore, in this phase, we utilize object-level grouping based on the Hausdorff distance for discrimination.

3.2.1. Utilizing Slice-Based Region Growing for Preliminary Extraction of Semantic Individual Objects

At the onset of processing, we initially employ the region growing method to segment the obtained facade semantic feature point cloud U into individual object. Since common attachments on building facades, such as long pipelines and drying clothes, can interfere with the effectiveness of the region growing algorithm, we first slice the point cloud U vertically. The initial thickness of the slices is set based on common building floor heights and is refined through a sliding window algorithm that adjusts the slice thickness for different levels, gradually optimizing until each slice fully covers all areas of the same floor. Subsequently, the region growing algorithm is applied to each floor slice for preliminary segmentation, and the segmentation results of each slice are merged to form an overall preliminary object segmentation. Ultimately, we obtain several semantic object point cloud clusters, denoted as V i , where i represents the number of segmented semantic objects. As shown in Figure 4.

3.2.2. Objects Similarity Measurement

The Hausdorff distance is a metric for measuring the minimum distance between two point sets, defined as the greatest distance between any two points from the compared sets. This measurement effectively captures the similarity of shapes and is particularly useful for comparing and identifying geometric objects, especially when dealing with nonlinear or irregular shapes. However, traditional Euclidean distance may overestimate in the presence of erroneous data, particularly with nonlinear data, due to its characteristic of treating errors in all dimensions equally.
To address this challenge, we introduce the FPFH (Fast Point Feature Histograms) feature as an alternative to traditional Euclidean distance [16]. FPFH captures the local geometric characteristics of point clouds through statistical analysis, generating a 33-dimensional feature vector that effectively describes the local shape and surface changes in the point cloud. It extends the scope of feature description by considering the relationships between the query point, its neighboring points, and the neighbors of those neighbors, while maintaining high computational efficiency. FPFH excels in multi-scale feature extraction, adapting to point cloud datasets of varying sizes, reducing the need for parameter tuning, and enhancing the algorithm’s adaptability to changes in the scale of point clouds. This ensures the accurate capture of key geometric features at various scales, thereby strengthening the comprehensiveness and robustness of feature description.
For two sets of points A and B, the computation of the Hausdorff distance D A , B = m a x a ϵ A m i n b ϵ B d a , b is as follows Figure 5.
The Hausdorff distance quantifies the similarity between the two sets of points and is used for the subsequent processing steps.

3.2.3. Hierarchical Clustering Grouping Based on Similarity Distance Difference Matrix

Considering the focus of this paper on semantic extraction for large urban building facades, the window structures on building facades typically exhibit regular spatial arrangements and often include a multitude of similarly shaped and sized individual objects. Leveraging this characteristic, we group the individual objects obtained from region growing segmentation in this stage based on structural similarity.
Specifically, we calculate the Hausdorff distances between objects within the point cloud clusters V i to assess the similarity among these semantic object point cloud clusters. Assuming the preliminary segmentation results in n objects, we construct an n × n matrix, which is a diagonal matrix, with the element at the i-th row and j-th column representing the Hausdorff distance between the i-th and j-th segmented objects. By summing and statistically analyzing each row of this matrix, we can use hierarchical clustering methods to group segmented objects with similar structures together [45]. Generally, windows with the same structure, even if not completely identical in the point cloud scanning results, will show clear similarities. Moreover, windows on different facades of the same building are often designed to be the same or similar. Therefore, when grouping, we aim to minimize the number of groups to better adapt to noise and data loss in the point cloud.
Additionally, it is important to consider that attachments on building facades, such as pipes, can interfere with the segmentation of semantic objects, causing adjacent windows to be incorrectly grouped together, or a single object to be segmented into multiple parts due to point cloud data loss. Experiments have shown that while such inevitable missegmentations occur, they are not dominant. Through the assessment of similarity between objects in this stage, these unavoidable missegmentations will be identified and excluded from the existing groups, thereby being removed from the set V i to prevent interference with subsequent computational processes.

3.3. Position-Shape Dual Optimization

In the previous phase, we obtained preliminary object-level point cloud segmentation results and grouped segmented objects with similar structures. In this phase, we will fully consider the spatial and shape-related similarities of window objects on building facades. To address the challenges caused by missing point cloud data, noise interference, and segmentation errors resulting from the limitations of segmentation methods, we will enhance and optimize the semantic structure of the facades from both positional and shape perspectives.

3.3.1. Spatial Distribution Correlated Position Optimization

When processing the grouped point cloud set V i , we first calculate its bounding box and determine the central point c i . Although theoretical, the central points of these semantic structure objects should follow spatial patterns of collinearity, coplanarity, or equal spacing, the characteristics of point cloud scanning and the accuracy of segmentation algorithms actually affect the distribution of these central points c i . Therefore, the primary task of this step is to establish spatial relationships among these central points.
Specifically, for the extracted building facade point cloud, we project it onto a plane and define a planar coordinate system O X Z . The origin O of the coordinate system is located at the lower left corner of the plane’s bounding box, with the Z-axis pointing towards the zenith and the X-axis pointing to the right, perpendicular to the Z-axis. Next, we project the extracted central points c i onto both the X-axis and Z-axis. The projection onto the X-axis is used to identify the collinear relationships of c i , while the projection onto the Z-axis is used to identify the coplanar relationships. Figure 6 illustrates this step, with Figure 6a showing the extracted central points c i (indicated by blue dots), and Figure 6b demonstrating that these fitted central points should theoretically satisfy collinear constraints, although there are minor deviations due to the accumulation of errors.
Based on the segmentation results of the windows, the target center points (blue points c i ) are obtained. By projecting these points onto the X and Z axes and fitting them to lines ( N j , M i ), the intersection of these lines represents the optimized window position (yellow points s i ). Furthermore, the previously obscured window structures in the point cloud can also be inferred (red points), as shown in Figure 6b. The corresponding co-row and co-column lines can be expressed as follows.
X = M i Z = N j
Subsequently, we will use these intersection points s i to optimize the positions of c i , ensuring they strictly adhere to collinear relationships. For each segmented object V i , we will find the nearest s i point and replace the current c i coordinates with it, while removing the corresponding nearest point from the set, as shown in Figure 6c. This process does not alter the original positions of the building point cloud but instead provides a more accurate center point for each segmented object as an anchor (indicated by yellow dots) to support further structural optimization.
It should be noted that the intervals of co-row and co-column relationships should be regular, meaning that M i and N j should conform to some arithmetic sequence combination. In the simplest case, if all windows on a facade have the same structure and are evenly distributed, then M i and N j will form arithmetic sequences. For more complex facades, as shown in Figure 7, the distribution of window columns may exhibit a pattern similar to alternating arithmetic sequences, where for a set M i = ( M 1 , M 2 , , M n ), the intervals between adjacent columns can be described as ( Δ 1 , Δ 2 , Δ 1 , Δ 2 , ).

3.3.2. Robust Extraction of Window Contours with In-Group Superimposed Line Optimization

Considering the uniform structural arrangement of windows on the facade of a building, in the shape optimization phase, we will superimpose and fit the windows of the same group to obtain the most complete and accurate window outline, and match it to the center point of each, thereby addressing the differences in scanning accuracy, omissions, and interference from facade attachments. For each segmented object V i , the classic α -shape algorithm is used to extract the contour boundary from the point cloud data, which can handle the concavity and convexity in the point cloud data, generating a polygonal shell to approximate the region occupied by the point cloud.
Subsequently, based on the grouping results of V i , each group is processed individually. For any group, based on the optimized object center point s i as the insertion anchor point, the obtained contour boundary is superimposed, and the line segments are measured for similarity and merged. Specifically, we first define a metric to measure the similarity between two adjacent line segments as follows:
S l i , l j = S θ ( l i , l j ) + S h ( l i , l j ) + S d ( l i , l j )
where l i , l j is two similar line segments
S θ l i , l j = l j s i n θ S h l i , l j = h 1 2 + h 2 2 h 1 + h 2 S d l i , l j = 0 ,   i f   l i   i n t e r s e c t s   w i t h   l j   d p ,                                                         e l s e
We created a class LINE to store and represent the parameters of a line segment, as shown in Figure 8. S θ ( l i , l j ) translates angular differences into a length representation, where θ represents the angle between line segments l i and l j . S h ( l i , l j ) denotes the perpendicular distance, which is a method that comprehensively considers distance and mitigates the impact of extreme values. Here, h 1 and h 2 represent the projection distances from the two endpoints of line segment l j to line segment l i . S d ( l i , l j ) quantifies the distance between two line segments by calculating the distance between their projected endpoints, where d p indicates the minimum distance between the endpoints of the two line segments. Generally, if two line segments l i and l j are similar, the similarity metrics S θ ( l i , l j ) , S h ( l i , l j ) , S d ( l i , l j ) will have smaller values. In other words, a smaller S ( l i , l j ) value implies that line segments l i and l j are more similar, as shown in Step 2 of Figure 9.
Under typical circumstances, the longer the line segment, the more likely it is to be correct. Therefore, the midpoint of the longest line segment is used as the positioning point for a new line segment (the red point in the figure). By utilizing similarity-based line segment clustering, a dominant direction is calculated, and then we can generate a new line using the central point and the directional vector. Finally, based on orthogonal projection, we use the two endpoints with the maximum distance to crop the line, resulting in a merged line segment, as shown in Figure 9.
Through the aforementioned steps, we have been able to construct a stable window outline for the segmented objects V i within the same group. Windows within the same group should exhibit consistent structural features, allowing us to use this outline to uniformly replace the contours of all individual objects within the group. Each group is processed independently, and by integrating the optimized positional results s i , we can generate a semantic layout map for the entire building facade.

3.3.3. Enhancement of Facade Model

The facade layout map obtained from the aforementioned steps will be applied to optimize the facade structure of 3D building models. Currently, it is easy to obtain LOD1/2 block models of buildings, where facades are often represented by planes with images as textures to describe the facade structures. Utilizing the facade layout map resulting from the methods proposed in this paper, we can achieve an accurate distribution of facade semantic structures that closely aligns with the real structures, guiding the refinement operations of building models. Figure 10 demonstrates the enhancement outcomes of the facade models from the dataset used in this study. In comparison to the block models, the models obtained are suitable for further applications in urban management and analysis.

4. Experiments

4.1. Experimental Setup

The experiments were conducted on the Visual Studio 2015 integrated development platform, using open-source libraries such as PCL (Point Cloud Library) and OpenCV for algorithm debugging. The hardware relied on an Intel (R) i7-8700K CPU and 128G of memory to run the experiments. Following the suggestions of the paper [46], the point cloud collected by TLS can model detailed objects. The results were presented using CloudCompare (v2.12.4) and SketchUp (v24.0.484) software. The line segment extraction results were compared with those of F3D (Fast 3D Line Segment Detection). The extraction results of structural feature points were compared and analyzed with ISS (Intrinsic Shape Signatures) and AC (Angle Criterion).
In the F3D method [47], the process begins with segmenting the point cloud into three-dimensional planes through region growing. Following this, the three-dimensional planes are projected onto two-dimensional images. Finally, the contours extracted from the two-dimensional images, which are two-dimensional line segments, are re-projected onto the three-dimensional planes to obtain the corresponding three-dimensional line segments. In the ISS method [48], the shape descriptor of Intrinsic Shape Signatures (ISS) is introduced to describe the local/semi-local areas of point clouds. An intrinsic shape signature uses a view-independent representation of the 3D shape to match shape patches from different views directly, and a view-dependent transform encoding the viewing geometry to facilitate fast pose estimation, in order to accurately distinguish feature points. In the AC method [49], the edge structural feature points of the point cloud are detected by studying the properties of the point set. For each point, we combine several criteria into an integrated boundary probability. A final boundary loop extraction step uses this probability and exploits additional coherence properties of the boundary to derive a robust distance and identify the structural feature points in the point cloud.
The M3C2 (Multi-Scale Model-to-Model Cloud Comparison) algorithm is a method for calculating the robust distance between two sets of point clouds. The uncertainty of the M3C2 distance is used to evaluate the effectiveness of feature extraction by computing the robust distance between two sets of point clouds. The smaller the number, the higher the spatial similarity between the two sets of data, indicating better feature extraction results. This experiment uses the built-in plugin in the CloudCompare software to complete the calculation of the M3C2 distance.

4.2. Data Source

We evaluated our proposed method on three datasets with distinctive characteristics: an open-source dataset, WHU-TLS [50], and two datasets we collected, NCWU-MLS and SJY-TLS. The details are described below.
The WHU dataset is publicly accessible (https://github.com/WHU-USI3DV/WHU-TLS (accessed on 27 October 2023). Due to the significant distance of the terrestrial laser scanner from the buildings, there is considerable variation in point cloud density, which results in less distinct detail features. The NCWU dataset was acquired using a backpack-mounted laser scanner, and the building facades are heavily occluded by attachments such as clothing and pipes. When collecting the SJY dataset with a terrestrial laser scanner, limitations in scanning angles led to large areas of missing data. Each of these datasets presents significant challenges for window extraction. More detailed information about the datasets can be found in Table 1 and Figure 11.
To confirm the precision of the experimental findings, the precise count of windows was established by referencing open-access street view imagery and photographs from the data acquisition phase. Concurrently, the F3D method [29] was employed in a comparative test to substantiate the efficacy of the methodology presented herein.

4.3. Method Evaluation

4.3.1. Qualitative Evaluation

Utilizing the innovative approach presented in this paper, we conducted application validation using three sets of data. We initially emphasized the overlay of facade extraction results with the original data to facilitate a qualitative analysis of the method. Overall, the method has demonstrated good performance on the datasets. Specifically, it can effectively detect the window structures of facades and generate precise semantic contours. For areas with defects in the dataset, the method is capable of supplementing the missing information based on the analysis of intact parts, and the detection accuracy of the supplemented sections has been confirmed by comparison with real images. Furthermore, the entire facade semantic contour is free of redundancy, with all contours ensuring regularization.
Figure 12 illustrates the experimental outcomes from the NCWU dataset, encompassing the original facade point cloud, the extracted facade layout, and a composite view of both for a clearer comparison. Additionally, it showcases the facade enhancement results for the LOD1 model. Within this dataset, two representative areas, A and B, were selected for analysis. Area A, which has defects in laser scanning, still yields a complete layout in the extracted map. Area B demonstrates the overlay effects of the extracted windows with laser scans and photographs, visually confirming the detection accuracy of the proposed algorithm. The results from the other two datasets can be seen in Figure 13.

4.3.2. Quantitative Evaluation

To fully demonstrate the potential of our method in facade semantic enhancement, we compared it with the F3D method using three datasets. Figure 14 visually presents the comparative results, while Table 2 provides a quantitative analysis.
In the proposed method, a window is represented as a rectangle composed of four line segments, and a window with all four segments detected is considered fully detected. If only two or three line segments are detected, it is regarded as incomplete detection, while detecting fewer than two line segments is considered a missed detection. By combining point cloud data with street-view images, we manually calculate the total number of objects (ground truth) and the number of undetected objects to compute the precision and missed detection rate, which is calculated according to Equation (8).
A c c u r a c y = C o m p l e t e   D e t e c t i o n G r o u n d   T r u t h M i s s   R a t e = M i s s e d   D e t e c t i o n G r o u n d   T r u t h
In the WHU-TLS dataset, the proposed method in this paper achieved a correct rate of 98.5%, which is significantly higher than the 62.6% of the F3D method. On the other two datasets, the proposed method achieved correct rates of 85.1% and 88.0%, respectively. Moreover, the dual optimization of position and shape adopted in this paper maximizes the avoidance of missed detections. In addition, all detected window line segments are closed, whereas those detected by the F3D method may have missing segments or be non-closed, as shown in areas C and D of Figure 14.

4.3.3. Performance of Feature Point Cloud Segment

To quantitatively compare the extraction effects, the M3C2 distance is used to calculate the uncertainty of the distance between the original point cloud and the point cloud after feature extraction, which is used to measure the effectiveness of feature extraction. The smaller this metric is, the higher the consistency between the two point clouds. It can be observed that of the three datasets, the method proposed in this paper outperforms the other two methods. Figure 15 visually presents the comparative results, while Table 3 provides a quantitative analysis.
In contrast, the method presented in this paper is not only capable of extracting features from shallow areas but also, guided by the optimal neighborhood, can adapt to different data, resulting in the best feature extraction outcomes.

4.3.4. Evaluation of Grouping Results for Objects with Same Structure

To obtain authentic and reliable window semantic details, it is essential to accurately judge the correctness of windows. The V i obtained from region growing mainly include two types: (1) windows with complete segmentation results, which have a small improved Hausdorff distance between them and exhibit higher similarity, and (2) windows that are either noise points or incorrectly segmented, which show a significant difference in the improved Hausdorff distance from normal windows. As shown in Figure 16 (1), the window matrix is a symmetric matrix where the horizontal or vertical axis represents a point cloud cluster V i , and the color of each cell is a quantified value of the improved Hausdorff distance; the darker the color, the lower the similarity. Those that have a significant difference from every row (column) can be identified as noise and will be excluded in subsequent processing. The box plot in Figure 16 (2) illustrates the distribution of the improved Hausdorff distances for normal V i and incorrect V i . The length of the box can reflect the central tendency and dispersion of the data. The closer the bottom and top of the correctly segmented windows, the more concentrated the improved Hausdorff distances of this group of V i are, indicating a lower degree of dispersion and higher homogeneity, likely originating from similar groups or conditions, thus showing less variation. In contrast, incorrectly segmented windows exhibit a significantly higher degree of dispersion, which under the conditions of this paper suggests that correctly segmented windows is more likely to consist of completely segmented windows, while incorrectly segmented windows consists of noise and incorrectly segmented windows. The points represented by small circles are outliers, indicating extreme values. It can be observed that the outliers for normal windows are smaller, while the data for noise points shows higher outliers. In this paper, a higher outlier value indicates that the point cloud cluster V i is more likely to be a noise or segmentation failure point.
After manual evaluation, a total of 97 point cloud groups were segmented in the NCWU dataset, of which 75 groups were successfully segmented. In the WHU dataset, 314 point cloud groups were identified, with 276 successfully segmented. For the SJY dataset, 361 point cloud groups were segmented, with 225 groups successfully segmented.

4.3.5. Evaluation of Location Optimization

During the process of position optimization, the blue points represent the center points of windows determined through similarity discrimination, with some omissions and misclassifications. However, by analyzing the spatial distribution of adjacent points, it is possible to infer the positions of windows that should exist, indicated by the red points. The combination of these two sets of points completes the reconstruction of the window structure. Due to data limitations, there are some windows that are difficult to detect, such as those on the first floor that are obscured, represented by empty squares in Figure 17. Some windows are obscured (mostly located on the ground floor), resulting in a lack of point cloud data, so the condition of the windows can only be determined through on-site exploration. The NCWU dataset demonstrates the peculiar arrangement of windows on the first floor (with a main door in the middle of the windows), and the WHU dataset shows the positional incompleteness caused by the obstruction of adjacent ground-level buildings (such as billboards). Our method detected 76 windows in the NCWU dataset and inferred the existence of 4 obstructed windows, which were then confirmed for detection accuracy by combining them with ground truth. However, 14 windows remained undetected. In the WHU dataset, 92 windows were detected, and 20 additional windows were inferred, but the correctness of these inferred windows could not be ascertained due to the lack of ground truth.

5. Discussion and Conclusions

5.1. Discussion and Limitation

Based on these analyses, the following discussion highlights both the innovations and limitations of the method presented in this paper.
(1) Innovations. A novel point cloud feature extraction method is proposed, capable of accurately capturing features in shallow regions and adapting to various complex environments. Moreover, the detection accuracy of window reconstruction is infallible. Through effective feature discrimination and regularization, the reconstructed windows achieve a high degree of precision in both position and shape.
(2) Limitation. The existing methods simplify semantic entities into rectangles, overlooking original shapes such as columns and staircases, leading to inadequate recognition of complex windows. Additionally, the effectiveness of judging the similarity of window structures depends on the previous segmentation results; poor segmentation can lead to errors in discerning the similarity of windows. Although the use of a sliding window to segment buildings helps reduce errors, this approach still results in missed detections. Furthermore, the set thresholds may exclude valid entities.

5.2. Conclusion and Future Work

In summary, this study successfully implemented a dual optimization method for building facade point cloud semantic detail enhancement based on optimal feature guidance in terms of both position and shape. The proposed framework effectively addresses the shortcomings of urban building facade reconstruction, such as the lack of semantic information, shape deformation, and positional inaccuracies. The novelty of our framework lies in the introduction of a new feature extraction method based on optimal neighborhood search, coupled with the dual optimization of position and shape to enhance the semantic details of facade point clouds. This framework not only efficiently extracts point cloud features in complex scenes but also integrates similarity discrimination and dual optimization of position and shape into the semantic facade model reconstruction process. The detection accuracy reached 98.5%, surpassing existing methods. According to the experimental analysis, the proposed framework adapts well to different complex environments and can recover point cloud facade semantic information to the greatest extent, even in areas with significant occlusion.
The extension research aims to advance the extraction of semantic details from photogrammetric point clouds or LiDAR point clouds, reducing potential quality degradation in facade detail reconstruction. We have made relevant progress in point cloud feature extraction and regularization. Furthermore, considering the increasing prevalence of large-scale models such as GPT (Generative Pretrained Transformer) and Stable Diffusion, we plan to integrate the point cloud facade semantic information model with these models to explore methods for recognizing and completing the semantic reconstruction of complex building structures. The extraction of cultural heritage details, patterns, and graphically weighted features involves the extraction and fitting of curves on the plane and even line segments in three-dimensional space, which is an independent and complex research topic. The methods for extracting and weighting features of cultural heritage details, patterns, and graphics are part of our further research content.

Author Contributions

Conceptualization, S.L.; Methodology, S.L. and F.Y.; Software, S.L., F.Y. and K.M.; Validation, F.Y. and Q.H.; Formal Analysis, S.L. and F.W.; Investigation, S.L. and F.W.; Data Curation, S.L., F.Y. and Q.H.; Writing—Original Draft Preparation, S.L.; Writing—Review and Editing, S.L. and K.M.; Visualization, S.L. and F.Y.; Project Administration, S.L. and K.M.; Funding Acquisition, S.L. and W.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Key Research Projects of Higher Education Institutions in Henan Province, No. 25B420001, the Guangxi Natural Science Foundation Grants, No. 2021GXNSFBA220001, and the China Henan Province Science and Technology project Grants, No. 232102320068.

Data Availability Statement

The WHU-TLS presented in this study are openly available in reference number [28]. The NCWU-MLS and SJY-TLS datasets are our own collections. Please feel free to reach out to the corresponding author to access them.

Acknowledgments

The authors would like to thank the editor and reviewers for their contributions on the paper.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Wang, P.; Ge, L.; Zhang, B.; Shi, B.; Duan, H.; Ke, C. GB-SAR Geocoding Considering the Radar Attitude Inclination Angles for Building Facade Deformation Extraction. Int. J. Remote Sens. 2024, 45, 8455–8470. [Google Scholar] [CrossRef]
  2. Romero-Jaren, R.; Arranz, J.J. Automatic Segmentation and Classification of BIM Elements from Point Clouds. Autom. Constr. 2021, 124, 103576. [Google Scholar] [CrossRef]
  3. Paiva, P.V.V.; Cogima, C.K.; Dezen-Kempter, E.; Carvalho, M.A.G. Historical Building Point Cloud Segmentation Combining Hierarchical Watershed Transform and Curvature Analysis. Pattern Recognit. Lett. 2020, 135, 114–121. [Google Scholar] [CrossRef]
  4. Lehtola, V.V.; Koeva, M.; Elberink, S.O.; Raposo, P.; Virtanen, J.-P.; Vahdatikhaki, F.; Borsci, S. Digital Twin of a City: Review of Technology Serving City Needs. Int. J. Appl. Earth Obs. Geoinf. 2022, 114, 102915. [Google Scholar] [CrossRef]
  5. Sun, X.; Guo, B.; Li, C.; Sun, N.; Wang, Y.; Yao, Y. Semantic Segmentation and Roof Reconstruction of Urban Buildings Based on LiDAR Point Clouds. ISPRS Int. J. Geo-Inf. 2024, 13, 19. [Google Scholar] [CrossRef]
  6. Yang, B.; Wei, Z.; Li, Q.; Li, J. Automated Extraction of Street-Scene Objects from Mobile Lidar Point Clouds. Int. J. Remote Sens. 2012, 33, 5839–5861. [Google Scholar] [CrossRef]
  7. Hu, Z.; Chen, C.; Yang, B.; Wang, Z.; Ma, R.; Wu, W.; Sun, W. Geometric Feature Enhanced Line Segment Extraction from Large-Scale Point Clouds with Hierarchical Topological Optimization. Int. J. Appl. Earth Obs. Geoinf. 2022, 112, 102858. [Google Scholar] [CrossRef]
  8. Han, X.; Dong, Z.; Yang, B. A Point-Based Deep Learning Network for Semantic Segmentation of MLS Point Clouds. ISPRS J. Photogramm. Remote Sens. 2021, 175, 199–214. [Google Scholar] [CrossRef]
  9. Chen, C.; Jin, A.; Yang, B.; Ma, R.; Sun, S.; Wang, Z.; Zong, Z.; Zhang, F. DCPLD-Net: A Diffusion Coupled Convolution Neural Network for Real-Time Power Transmission Lines Detection from UAV-Borne LiDAR Data. Int. J. Appl. Earth Obs. Geoinf. 2022, 112, 102960. [Google Scholar] [CrossRef]
  10. Pantoja-Rosero, B.G.; Rusnak, A.; Kaplan, F.; Beyer, K. Generation of LOD4 Models for Buildings towards the Automated 3D Modeling of BIMs and Digital Twins. Autom. Constr. 2024, 168, 105822. [Google Scholar] [CrossRef]
  11. Wang, F.; Zhou, G.; Hu, H.; Wang, Y.; Fu, B.; Li, S.; Xie, J. Reconstruction of LoD-2 Building Models Guided by Facade Structures from Oblique Photogrammetric Point Cloud. Remote Sens. 2023, 15, 400. [Google Scholar] [CrossRef]
  12. Ge, Y.; Xiao, X.; Guo, B.; Shao, Z.; Gong, J.; Li, D. A Novel LOD Rendering Method With Multilevel Structure-Keeping Mesh Simplification and Fast Texture Alignment for Realistic 3-D Models. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5640519. [Google Scholar] [CrossRef]
  13. Biljecki, F.; Ledoux, H.; Stoter, J. An Improved LOD Specification for 3D Building Models. Comput. Environ. Urban Syst. 2016, 59, 25–37. [Google Scholar] [CrossRef]
  14. Chen, H.; Cui, W.; Bo, C.; Yang, N. Point Cloud Simplification for the Boundary Preservation Based on Extracted Four Features. Displays 2023, 78, 102414. [Google Scholar] [CrossRef]
  15. Mokhtarian, F.; Suomela, R. Robust Image Corner Detection through Curvature Scale Space. IEEE Trans. Pattern Anal. Mach. Intell. 1998, 20, 1376–1381. [Google Scholar] [CrossRef]
  16. Rusu, R.B.; Blodow, N.; Beetz, M. Fast Point Feature Histograms (FPFH) for 3D Registration. In Proceedings of the 2009 IEEE International Conference on Robotics and Automation, Kobe, Japan, 12–17 May 2009; pp. 3212–3217. [Google Scholar]
  17. He, Y.; Hu, Z.; Wang, R.; Zhu, H.; Fu, G. Features Extraction of Point Clouds Based on Otsu’s Algorithm. Meas. Sci. Technol. 2024, 35, 065205. [Google Scholar] [CrossRef]
  18. Tao, W.; Xiao, Y.; Wang, R.; Lu, T.; Xu, S. A Fast Registration Method for Building Point Clouds Obtained by Terrestrial Laser Scanner via 2-D Feature Points. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2024, 17, 9324–9336. [Google Scholar] [CrossRef]
  19. He, L.; Wang, S.; Hu, Q.; Cai, Q.; Li, M.; Bai, Y.; Wu, K.; Xiang, B. GFOICP: Geometric Feature Optimized Iterative Closest Point for 3-D Point Cloud Registration. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5704217. [Google Scholar] [CrossRef]
  20. Gu, Y.; Xiao, Z.; Li, X. A Spatial Alignment Method for UAV LiDAR Strip Adjustment in Nonurban Scenes. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5702413. [Google Scholar] [CrossRef]
  21. Charles, R.Q.; Su, H.; Kaichun, M.; Guibas, L.J. PointNet: Deep Learning on Point Sets for 3D Classification and Segmentation. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 77–85. [Google Scholar]
  22. Chen, Y.; Liu, J.; Zhang, X.; Qi, X.; Jia, J. LargeKernel3D: Scaling up Kernels in 3D Sparse CNNs. In Proceedings of the 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 18–22 June 2023; pp. 13488–13498. [Google Scholar]
  23. Yang, S.; Hou, M.; Li, S. Point Cloud Semantic Segmentation of Grotto Scenes Using the Knowledge-Guided Deep Learning Method. Int. J. Digit. Earth 2024, 17, 2385081. [Google Scholar] [CrossRef]
  24. Cao, X.; Wang, H.; Zhu, Q.; Wang, Y.; Liu, X.; Li, K.; Su, L. PointStaClu: A Deep Point Cloud Clustering Method Based on Stable Cluster Discrimination. Remote Sens. 2024, 16, 2423. [Google Scholar] [CrossRef]
  25. He, P.; Gao, K.; Liu, W.; Song, W.; Hu, Q.; Cheng, X.; Li, S. OFFS-Net: Optimal Feature Fusion-Based Spectral Information Network for Airborne Point Cloud Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2023, 16, 141–152. [Google Scholar] [CrossRef]
  26. Liu, S.; Sun, H.; Zhang, Z.; Li, Y.; Zhong, R.; Li, J.; Chen, S. A Multiscale Deep Feature for the Instance Segmentation of Water Leakages in Tunnel Using MLS Point Cloud Intensity Images. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5702716. [Google Scholar] [CrossRef]
  27. Kazhdan, M.; Bolitho, M.; Hoppe, H. Poisson Surface Reconstruction. In Proceedings of the Fourth Eurographics Symposium on Geometry Processing; Eurographics Association: Goslar, Germany, 2006; pp. 61–70. [Google Scholar]
  28. Li, L.; Song, N.; Sun, F.; Liu, X.; Wang, R.; Yao, J.; Cao, S. Point2Roof: End-to-End 3D Building Roof Modeling from Airborne LiDAR Point Clouds. ISPRS J. Photogramm. Remote Sens. 2022, 193, 17–28. [Google Scholar] [CrossRef]
  29. Kada, M. 3D RECONSTRUCTION OF SIMPLE BUILDINGS FROM POINT CLOUDS USING NEURAL NETWORKS WITH CONTINUOUS CONVOLUTIONS (CONVPOINT). Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2022, XLVIII-4/W4-2022, 61–66. [Google Scholar] [CrossRef]
  30. Moyano, J.; Musicco, A.; Nieto-Julián, J.E.; Domínguez-Morales, J.P. Geometric Characterization and Segmentation of Historic Buildings Using Classification Algorithms and Convolutional Networks in HBIM. Autom. Constr. 2024, 167, 105728. [Google Scholar] [CrossRef]
  31. Li, L.; Chen, J.; Su, X.; Han, H.; Fan, C. Deep Learning Network for Indoor Point Cloud Semantic Segmentation with Transferability. Autom. Constr. 2024, 168, 105806. [Google Scholar] [CrossRef]
  32. Ma, J.; Jiang, X.; Zheng, D.; Liao, X.; Lu, J.; Zhao, Y. TransWallNet: High-Performance Semantic Segmentation of Large-Scale and Multifeatured Point Clouds of Building Gables. J. Constr. Eng. Manag. 2024, 150, 04024092. [Google Scholar] [CrossRef]
  33. Cai, S.; Zhang, S.; Zhang, W.; Fan, H.; Shao, J.; Yan, G.; Yu, S.; Li, A.; Zhou, G. A General and Effective Method for Wall and Protrusion Separation from Facade Point Clouds. J. Remote Sens. 2023, 3, 0069. [Google Scholar] [CrossRef]
  34. Su, Y.; Liu, W.; Yuan, Z.; Cheng, M.; Zhang, Z.; Shen, X.; Wang, C. DLA-Net: Learning Dual Local Attention Features for Semantic Segmentation of Large-Scale Building Facade Point Clouds. Pattern Recognit. 2022, 123, 108372. [Google Scholar] [CrossRef]
  35. Zhou, G.; Bao, X.; Ye, S.; Wang, H.; Yan, H. Selection of Optimal Building Facade Texture Images From UAV-Based Multiple Oblique Image Flows. IEEE Trans. Geosci. Remote Sens. 2021, 59, 1534–1552. [Google Scholar] [CrossRef]
  36. Wang, Y.; Jiao, W.; Fan, H.; Zhou, G. A Framework for Fully Automated Reconstruction of Semantic Building Model at Urban-Scale Using Textured LoD2 Data. ISPRS J. Photogramm. Remote Sens. 2024, 216, 90–108. [Google Scholar] [CrossRef]
  37. Guo, Y.; Wang, J.; Yu, G.; Zhang, C.; Lin, X.; Xiong, H.; Shi, G. Road Weather Condition Recognition via Fusing Images and Upsampled Point Cloud Reflection Intensities. IEEE Sens. J. 2024, 24, 19286–19296. [Google Scholar] [CrossRef]
  38. Su, Y.; Cheng, M.; Yuan, Z.; Liu, W.; Zeng, W.; Zhang, Z.; Wang, C. Spatial Adaptive Fusion Consistency Contrastive Constraint: Weakly Supervised Building Facade Point Cloud Semantic Segmentation. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5703214. [Google Scholar] [CrossRef]
  39. Hope Akwensi, P.; Bharadwaj, A.; Wang, R. APC2Mesh: Bridging the Gap from Occluded Building Façades to Full 3D Models. ISPRS J. Photogramm. Remote Sens. 2024, 211, 438–451. [Google Scholar] [CrossRef]
  40. Saraiva, P. On Shannon Entropy and Its Applications. Kuwait J. Sci. 2023, 50, 194–199. [Google Scholar] [CrossRef]
  41. Zhang, Y.; Feng, W.; Quan, Y.; Ye, G.; Dauphin, G. Dynamic Spatial–Spectral Feature Optimization-Based Point Cloud Classification. Remote Sens. 2024, 16, 575. [Google Scholar] [CrossRef]
  42. Li, W.; Pan, J.; Hasegawa, K.; Li, L.; Tanaka, S. Missing Region Completion Network for Large-Scale Laser-Scanned Point Clouds: Application to Transparent Visualization of Cultural Heritage. Remote Sens. 2024, 16, 2758. [Google Scholar] [CrossRef]
  43. Yu, J.; Wang, J.; Zang, D.; Xie, X. A Feature Line Extraction Method for Building Roof Point Clouds Considering the Grid Center of Gravity Distribution. Remote Sens. 2024, 16, 2969. [Google Scholar] [CrossRef]
  44. Qu, C.; Zhang, Y.; Sun, Z.; Ma, F.; Zhang, X. Point Clouds Denoising Method Based on SPAD Discrete Probability Model and Completion. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5702115. [Google Scholar] [CrossRef]
  45. Lu, X.; Yao, J.; Tu, J.; Li, K.; Li, L.; Liu, Y. PAIRWISE LINKAGE FOR POINT CLOUD SEGMENTATION. ISPRS Ann. Photogramm. Remote Sens. Spat. Inf. Sci. 2016, III–3, 201–208. [Google Scholar] [CrossRef]
  46. Warchoł, A.; Karaś, T.; Antoń, M. SELECTED QUALITATIVE ASPECTS OF LIDAR POINT CLOUDS: GEOSLAM ZEB-REVO AND FARO FOCUS 3D X130. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2023, XLVIII-1-W3-2023, 205–212. [Google Scholar] [CrossRef]
  47. Lu, X.; Liu, Y.; Li, K. Fast 3D Line Segment Detection From Unorganized Point Cloud. arXiv 2019, arXiv:1901.02532. [Google Scholar]
  48. Zhong, Y. Intrinsic Shape Signatures: A Shape Descriptor for 3D Object Recognition. In Proceedings of the 2009 IEEE 12th International Conference on Computer Vision Workshops, ICCV Workshops, Kyoto, Japan, 27 September–4 October 2009; pp. 689–696. [Google Scholar]
  49. Bendels, G.; Schnabel, R.; Klein, R. Detecting Holes in Point Set Surfaces. J. WSCG 2006, 14, 89–96. [Google Scholar]
  50. Dong, Z.; Liang, F.; Yang, B.; Xu, Y.; Zang, Y.; Li, J.; Wang, Y.; Dai, W.; Fan, H.; Hyyppä, J.; et al. Registration of Large-Scale Terrestrial Laser Scanner Point Clouds: A Review and Benchmark. ISPRS J. Photogramm. Remote Sens. 2020, 163, 327–342. [Google Scholar] [CrossRef]
Figure 1. The workflow of the proposed method.
Figure 1. The workflow of the proposed method.
Remotesensing 16 04324 g001
Figure 2. Calculate the vector from the point to the centroid to describe the local structural features. (The yellow point p i represents a certain point in the point cloud during the traversal process, and a neighborhood will be constructed with it; the blue point p j represents a certain point in the neighborhood of p i ; the purple point c i is the traditional centroid, and the distance from the yellow point to the purple point is the magnitude of the traditional centroid displacement vector; the red point d i is the weighted centroid, and the distance from the yellow point to the red point is the magnitude of the weighted centroid displacement vector).
Figure 2. Calculate the vector from the point to the centroid to describe the local structural features. (The yellow point p i represents a certain point in the point cloud during the traversal process, and a neighborhood will be constructed with it; the blue point p j represents a certain point in the neighborhood of p i ; the purple point c i is the traditional centroid, and the distance from the yellow point to the purple point is the magnitude of the traditional centroid displacement vector; the red point d i is the weighted centroid, and the distance from the yellow point to the red point is the magnitude of the weighted centroid displacement vector).
Remotesensing 16 04324 g002
Figure 3. The locally weighted centroid displacement scheme. (The temporary point m j is obtained through the parallelogram rule constructed by p i , p j , and c i . It should be noted that each p j will produce a m j , and the sum of all m j will result in the red point d i . The specific calculation method is referred to as Formula (4)).
Figure 3. The locally weighted centroid displacement scheme. (The temporary point m j is obtained through the parallelogram rule constructed by p i , p j , and c i . It should be noted that each p j will produce a m j , and the sum of all m j will result in the red point d i . The specific calculation method is referred to as Formula (4)).
Remotesensing 16 04324 g003
Figure 4. Region growing based on the slices. (First, the optimal slice height is determined through a sliding window to ensure that each slice represents a single layer of windows; then, region growing is used to perform preliminary segmentation on each layer of windows).
Figure 4. Region growing based on the slices. (First, the optimal slice height is determined through a sliding window to ensure that each slice represents a single layer of windows; then, region growing is used to perform preliminary segmentation on each layer of windows).
Remotesensing 16 04324 g004
Figure 5. Flowchart of Hausdorff distance based on FPFH.
Figure 5. Flowchart of Hausdorff distance based on FPFH.
Remotesensing 16 04324 g005
Figure 6. Location optimization process of window semantic center point. (The blue point represents the initial center position of the window, which may be displaced due to errors during scanning. The yellow point represents the corrected center of the window, obtained from the intersection of the line fitted based on the blue point. The red point represents the originally obscured position of the window, inferred from the collinearity relationship).
Figure 6. Location optimization process of window semantic center point. (The blue point represents the initial center position of the window, which may be displaced due to errors during scanning. The yellow point represents the corrected center of the window, obtained from the intersection of the line fitted based on the blue point. The red point represents the originally obscured position of the window, inferred from the collinearity relationship).
Remotesensing 16 04324 g006
Figure 7. Center point distribution.
Figure 7. Center point distribution.
Remotesensing 16 04324 g007
Figure 8. Diagram of class LINE.
Figure 8. Diagram of class LINE.
Remotesensing 16 04324 g008
Figure 9. Robust contour structures obtained by group object overlay and line segment optimization. (Firstly, the α-shape is used to extract the contour line segments of all windows and overlay them; secondly, the proposed similarity measure is applied to merge line segments with high similarity to complete shape optimization; finally, the optimized contours are placed in the optimized positions to achieve dual optimization of position and shape).
Figure 9. Robust contour structures obtained by group object overlay and line segment optimization. (Firstly, the α-shape is used to extract the contour line segments of all windows and overlay them; secondly, the proposed similarity measure is applied to merge line segments with high similarity to complete shape optimization; finally, the optimized contours are placed in the optimized positions to achieve dual optimization of position and shape).
Remotesensing 16 04324 g009
Figure 10. Facade enhancement of building model.
Figure 10. Facade enhancement of building model.
Remotesensing 16 04324 g010
Figure 11. The details of the datasets.
Figure 11. The details of the datasets.
Remotesensing 16 04324 g011
Figure 12. Experimental evaluation of dataset NCWU-MLS.
Figure 12. Experimental evaluation of dataset NCWU-MLS.
Remotesensing 16 04324 g012
Figure 13. Experimental evaluation of datasets WHU-TLS and SJY-TLS.
Figure 13. Experimental evaluation of datasets WHU-TLS and SJY-TLS.
Remotesensing 16 04324 g013
Figure 14. An intuitive comparison between the proposed method and the F3D method.
Figure 14. An intuitive comparison between the proposed method and the F3D method.
Remotesensing 16 04324 g014
Figure 15. Comparison of feature point extraction.
Figure 15. Comparison of feature point extraction.
Remotesensing 16 04324 g015
Figure 16. Hausdorff distance matrix-based object structure comparison and grouping.
Figure 16. Hausdorff distance matrix-based object structure comparison and grouping.
Remotesensing 16 04324 g016
Figure 17. Evaluation of center point location optimization.
Figure 17. Evaluation of center point location optimization.
Remotesensing 16 04324 g017
Table 1. Details of datasets.
Table 1. Details of datasets.
DatasetEquipment TypeNumber of PointsArea
Size (m2)
Point Density
(Point/m2)
Window Size (m2)Building Story
WHU-TLSVZ-4001,048,90748 × 65 × 682500216
NCWU-MLSLeica Pegasus8,038,70965 × 33 × 2137003.96
SJY-TLSLeica- P501,243,68632 × 32 × 10535003.6/0.521
Table 2. Comparison of detection results.
Table 2. Comparison of detection results.
DatasetMethodComplete
Detection
Incomplete DetectionMissed DetectionAccuracyMiss Rate
NCWU-MLSF3D58132361.7%24.5%
Ours8001485.1%14.9%
WHU-TLSF3D77457262.6%23.0%
Ours3080598.5%1.5%
SJY-TLSF3D1595611947.6%35.6%
Ours29404088.0%11.9%
Table 3. Feature extraction evaluation. (By calculating the robust distance between the point clouds before and after extraction, the effectiveness of feature extraction can be assessed. The smaller the number, the higher the spatial similarity between the two sets of data, indicating better results of feature extraction).
Table 3. Feature extraction evaluation. (By calculating the robust distance between the point clouds before and after extraction, the effectiveness of feature extraction can be assessed. The smaller the number, the higher the spatial similarity between the two sets of data, indicating better results of feature extraction).
Distance UncertaintyISSACOurs
NCWU-MLS0.190.0320.014
WHU-TLS0.0930.0860.077
SJY-TLS0.0390.0200.019
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, S.; Yan, F.; Ma, K.; Hu, Q.; Wang, F.; Liu, W. Optimal Feature-Guided Position-Shape Dual Optimization for Building Point Cloud Facade Detail Enhancement. Remote Sens. 2024, 16, 4324. https://doi.org/10.3390/rs16224324

AMA Style

Li S, Yan F, Ma K, Hu Q, Wang F, Liu W. Optimal Feature-Guided Position-Shape Dual Optimization for Building Point Cloud Facade Detail Enhancement. Remote Sensing. 2024; 16(22):4324. https://doi.org/10.3390/rs16224324

Chicago/Turabian Style

Li, Shiming, Fengtao Yan, Kaifeng Ma, Qingfeng Hu, Feng Wang, and Wenkai Liu. 2024. "Optimal Feature-Guided Position-Shape Dual Optimization for Building Point Cloud Facade Detail Enhancement" Remote Sensing 16, no. 22: 4324. https://doi.org/10.3390/rs16224324

APA Style

Li, S., Yan, F., Ma, K., Hu, Q., Wang, F., & Liu, W. (2024). Optimal Feature-Guided Position-Shape Dual Optimization for Building Point Cloud Facade Detail Enhancement. Remote Sensing, 16(22), 4324. https://doi.org/10.3390/rs16224324

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop