Next Article in Journal
Validation of ICESat-2 ATLAS Bathymetry and Analysis of ATLAS’s Bathymetric Mapping Performance
Next Article in Special Issue
Semantic Segmentation of Urban Buildings from VHR Remote Sensing Imagery Using a Deep Convolutional Neural Network
Previous Article in Journal
GPU-Based Lossless Compression of Aurora Spectral Data using Online DPCM
Previous Article in Special Issue
The Comparison of Fusion Methods for HSRRSI Considering the Effectiveness of Land Cover (Features) Object Recognition Based on Deep Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Building Extraction Approach Based on the Fusion of LiDAR Point Cloud and Elevation Map Texture Features

1
School of Remote Sensing and Information Engineering, Wuhan University, Wuhan 430079, China
2
Key Laboratory for National Geographic Census and Monitoring, National Administration of Surveying, Mapping and Geoinformation, Wuhan 430079, China
3
Institute of Geological Survey, China University of Geosciences, Wuhan 430074, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2019, 11(14), 1636; https://doi.org/10.3390/rs11141636
Submission received: 15 May 2019 / Revised: 30 June 2019 / Accepted: 3 July 2019 / Published: 10 July 2019
(This article belongs to the Special Issue Remote Sensing based Building Extraction)

Abstract

:
Building extraction is an important way to obtain information in urban planning, land management, and other fields. As remote sensing has various advantages such as large coverage and real-time capability, it becomes an essential approach for building extraction. Among various remote sensing technologies, the capability of providing 3D features makes the LiDAR point cloud become a crucial means for building extraction. However, the LiDAR point cloud has difficulty distinguishing objects with similar heights, in which case texture features are able to extract different objects in a 2D image. In this paper, a building extraction method based on the fusion of point cloud and texture features is proposed, and the texture features are extracted by using an elevation map that expresses the height of each point. The experimental results show that the proposed method obtains better extraction results than that of other texture feature extraction methods and ENVI software in all experimental areas, and the extraction accuracy is always higher than 87%, which is satisfactory for some practical work.

Graphical Abstract

1. Introduction

Remote sensing is the acquisition of information about objects or phenomena without physical contact [1]. A large amount of remote sensing data has been generated and applied, and improvements in the spatial and temporal resolution of remote sensing images have made them become the main data source for object extraction [2], such as tree crown extraction [3], coastal zone detection [4], road recognition [5], etc. Buildings constitute the main component of urban areas, and building extraction using remote sensing images has become a hot research topic as remote sensing technology has the advantage of being fast, large-scale, and economical. Some researchers provided information of the spectral, geometrical, contextual, and rooftop segment patch via the morphological building index (MBI) and saliency cue to extract building information, which had good performance and versatility under different image conditions. However, the image-based building extraction technique is limited by large intra-class differences and small inter-class differences in spectral features [6,7]. 3D information is valid for buildings, especially elevation information, while for the image. it is complicated to realize, and it is mainly reflected by the change of elevation, which is important information of buildings.
As one of the active remote sensing data sources, LiDAR uses laser pulses to measure the distance between the sensor and different objects. It is widely used in geodesy [8], geo-statistics [9], archeology [10], geography [11], the control and navigation of autonomous vehicles [12], etc. Compared with 2D images, which only provide position and shape information, LiDAR can conveniently acquire 3D information on objects in terrain. Therefore, many studies have applied the LiDAR point cloud to conduct building extraction [13]. Wang et al. adopted a building extraction technique based on the point voxel group by using the class-oriented fusion method and “horizontal hollow ratio”, which was effective for large-scale and complex urban environments [14]. Qin et al. demonstrated the use of geometric and radiation features of the waveform and the point cloud with parametric and non-parametric classification methods. The experimental results suggested that it was efficiently used for urban land cover mapping [15]. Zhao et al. utilized connected operators to extract building regions from LiDAR data, neither producing new contours, nor changing positions, which was effective, and the average offset values of simple and complex building boundaries were 0.2–0.4 m [16]. Huang et al. proposed a novel object and region-based top-down strategy to extract buildings, and the experimental result proved that the proposed method achieved good performance and was robust when parameters were within reasonable ranges [17]. Yi et al. detailed a method for reconstructing the volume structure of urban buildings directly from the original LiDAR point cloud. The experimental results demonstrated the advantage of the approach in terms of effectiveness on large-scale and raw LiDAR point data [18].
However, the discreteness of the point cloud may lead to the loss of some features, and it is difficult to distinguish objects with similar heights, while it is able to extract different objects with texture features in 2D images. As elevation map is a kind of 2D image obtained by projecting the point cloud onto 2D planes, and it can provide abundant texture features and has been utilized in the field of building extraction. Fasahat et al. realized building extraction by transforming the point cloud into an elevation map and analyzing gradient information from the elevation map. Experimental results showed the effectiveness in eliminating trees, extracting buildings of all sizes, and extracting buildings with and without a transparent roof [19]. Liu et al. combined remote sensing data of multiple sources to draw height maps of different object types for land cover and land use mapping, which coincided well with the ground survey data with an accuracy of 5.7 m by root mean squared error (RMSE) [20]. Kang et al. achieved the rendering of barren terrain by enhancing the geometric features of elevation maps and increased the number of landscape features, which was most suitable for rendering barren terrain or planet surfaces [21]. He et al. proposed to organize LiDAR point data as three different maps: dense depth map, height map, and surface normal map. It was proven to recover successfully object hierarchies, boundary sharpness, and global integrity regardless of point cloud sparsity, large loss, and 3D to 2D degradation uncertainty [22]. In addition, the texture feature extraction method can be used to obtain features to extract objects on the basis of the elevation map, which can robustly detect buildings from satellite images and outperforms state-of-the-art building detection method [23]. Cao et al. constructed a unified multilevel channel characteristic framework and realized target detection based on histograms of oriented gradient (HoG) features. The experimental results showed that this method could reduce the missed detection rate and improve the detection speed [24]. Du et al. used the gray level co-occurrence matrix (GLCM) features to obtain textures from an elevation map and combined them with point cloud information to achieve area and object-level building extraction, and the results suggested a good potential for large-sized LiDAR data [25]. Niemi et al. inventoried soil damage from forwarding trails and fitted a logistic regression model for predicting the event of soil damage, which showed that DTM-derived local binary patterns (LBP) were useful in terrain trafficability mapping [26].
Point cloud information can reflect the spatial structure of ground objects, but its discrete type may lead to the lack of correlation information of each part. Texture features can reflect the correlation of each part and help to distinguish different objects. Therefore, point cloud and texture features can be fused to achieve complementarity, as well as reflect the features of objects from multiple dimension so as to obtain better results. However, the increased data dimension may lead to an increase in time complexity, and feature selection is always utilized to solve the problem. As the essence of feature selection is a combinatorial optimization problem, which means selecting a satisfactory feature subset to conduct building extraction, it is usually solved by swarm intelligence algorithms [27]. In this paper, by fusing the point cloud and texture features, as well as conducting feature selection, a building extraction technique is realized. Point cloud features are extracted based on the eigenvalue, density, and elevation, and the point cloud is also transformed into an elevation map to extract texture features. After that, the fusion of the point cloud and texture features is used to extract buildings from different experimental areas. Among various swarm intelligence algorithms, particle swarm optimization (PSO) is easy to implement, and stably converges to the optimal solution. Therefore, it is adopted to obtain the superior feature subset for building extraction in the paper.
This paper is structured as follows. In Section 2, the core method and basic principles of this paper are elaborated in detail. The steps of the method are described in Section 3. Section 4 describes the experiments that are carried out according to the method, the experimental data, the final results, and the evaluation of the accuracy. Section 5 summarizes the work of this paper and research prospects.

2. Basic Theory of Gabor Filters

As for 2D images, the Gabor filter is one of the efficient filtering techniques and is based on a sinusoidal plane wave. Its use has been explored in many applications [28,29]. The Gabor filter can not only characterize the spatial frequency structure of an image, but also retain spatial relationship information, and the spatial frequency positioning ability is essential to extract orientation-dependent frequency content from the pattern [30]. Furthermore, as the Gabor filter is invariant to zoom, rotation, and translation, it is suitable for texture representation and recognition [31]. In the spatial domain, a 2D Gabor filter is a Gaussian kernel function modulated by a sinusoidal plane wave, which consists of a real part and an imaginary part representing the orthogonal direction. These two parts can either form a plurality or be used separately [32,33].
The formula for the Gabor filter is expressed as below:
g ( x , y ) = 1 2 π σ x σ y e x p 1 2 x ¯ 2 + y ¯ 2 σ x 2 + σ y 2 + 2 π j W x ¯
x ¯ = x c o s θ + y s i n θ y ¯ = x s i n θ + y c o s θ
where σ x and σ y are parameters that describe the spread of the current pixel in the neighborhood in which weighted summation occurs, W is the central frequency of the complex sinusoid, θ [ 0 , π ) is the orientation of the horizontal to vertical stripes in the equation above, and j represents the imaginary unit.
The extraction of texture features using the Gabor filter includes two main processes: filter design and the effective extraction of texture feature sets from the filter’s output. The process of acquiring texture features from an image using the Gabor filter is as follows. Firstly, the input image is divided into blocks. Secondly, the Gabor filter banks are established, and thirdly, we convolve the Gabor filter templates with each image block in the spatial domain; each image block obtains the filter outputs. These outputs are of the image blocks’ size. Fourthly, each image block is passed through the outputs of the Gabor filter templates and is “condensed” into the texture feature of the image block [34].

3. Building Extraction Based on the Fusion of Point Cloud and Texture Features

3.1. Point Cloud Features

At first, as the LiDAR system generates a number of noise points when acquiring data, which is usually manifested as elevation anomaly points and will affect the accuracy of building extraction, the point cloud is denoised, and elevation anomalies are filtered out. After that, the features of the point cloud, which include various eigenvalues, are obtained. Unlike eigenvectors, eigenvalues have good rotationally-invariant properties [35], and therefore, feature extraction based on the point cloud’s eigenvalues was used for building extraction. Besides, density and elevation are both critical attributes of point cloud. Thus, features based on eigenvalues, density, and elevation were extracted as the reference data of building extraction. The specific meanings and formulas used in the calculations are shown in Table 1.
λ 1 , λ 2 , and λ 3 are eigenvalues of the point cloud, where λ 1 < λ 2 < λ 3 . An analysis of the eigenvalues and eigenvectors can often provide important information for extraction decisions. According to the points in the neighborhood, the covariance matrix of the center point was calculated, and then the eigenvalues of the point were obtained. Based on these eigenvalues, 12 kinds of features can be calculated, including sum of eigenvalues (SU), total variance (TV), eigenvalues (EI), anisotropy (AN), planarity (PL), linearity (LI), surface roughness (SR), and sphericity (SP). AN refers to the uniformity of the point distribution on three arbitrary vertical axes, which helps to separate anisotropic structures, such as power lines and buildings, from vegetation. PL is a measurement of planar characteristics of the point cloud, and planar structures have high PL values. As the surface of a building’s roof reflects laser directly, this feature is remarkable. LI is a measurement of the linear attributes of a point cloud. The power lines and edges of buildings have obvious linear structures, and the linearity of these points is characterized by high values. SR is the average number of points allocated by the point cloud in three directions. The distribution of vegetation points in all directions has no tendency, so the SR values of vegetation are high. The density of the point cloud in the neighborhood of penetrable targets, such as vegetation, reflects the distribution of the point cloud and is usually higher than that of buildings. In the vicinity of the cylinder at the center point, the height differences, including that between the current point and the lowest point (height above (HA)) and that between the highest point and the current point (height below (HB)), were calculated. The standard deviation of the elevation included the elevation of each point in the spherical and cylindrical neighborhoods. Z a v e is the average of the current neighborhood’s interior point elevation; n is the number of points in the current neighborhood’s interior point cloud; and Z i is the i-th point in the neighborhood. The sphere variance (SPV) value is high for objects with few changes in elevation [36]. For high-rise building facades and roofs, the differences between the current and the lowest elevation are usually much larger than that of other points. Therefore, building facades can be distinguished effectively, while the standard deviations of elevations in spherical neighborhoods can be used to identify ground and other horizontal planes. All of the mentioned features above show the properties of point clouds from the point of view of eigenvalues, elevations, and densities. They can provide more effective information for building extraction than single-scale features [37].

3.2. Texture Feature Extraction Based on the Elevation Map

In this study, the point cloud was transformed into an elevation map for texture feature extraction. The process of transformation was based on the elevation distribution of the point cloud, and it was easy to operate. Firstly, the grid size was set as 1 m, and then, the point cloud was rasterized according to its x and y coordinates, while each grid corresponded to one pixel in the elevation map [38]. After that, the height threshold was set, and the elevation variance of all points in the corresponding grid of each pixel was calculated. If the variance was below the threshold, the average elevation of points in the gird was selected as the gray reference value of the corresponding pixel. Otherwise, the height distribution curve was interpolated based on triangulation in the natural neighborhood, and half of the peak value was taken as the gray reference value. For a grid with few or even no points, the median elevation value of the points in the K-nearest-neighbor was taken. After all the above, the gray reference values were normalized to 0–255. In this way, the elevation map corresponding to the point cloud can be obtained as shown in Figure 1.
After the elevation map was obtained, the corresponding texture features were extracted for further building extraction. Compared with other methods, the Gabor filter can capture those features that correspond to different spatial frequencies (scales) and orientations, so it can be used to discriminate features of images. In this study, a 2D Gabor filter was used to extract texture features. The texture features of elevation maps in different orientations and scales were obtained by changing the values of the orientation and frequency parameters. The orientation and frequency values were updated as follows:
θ ( i ) = ( i 1 ) π O , w h e r e i = 1 , 2 , . . . , O
f ( i ) = f m a x ( 2 ) i 1 , w h e r e i = 1 , 2 , . . . , S
where θ ( i ) is the orientation parameter, O is the number of orientation parameters, f ( i ) is the frequency variable, and S is the number of frequency variables. In this study, four frequency values and six orientation values were combined to obtain 24 texture features. The frequency values changed gradually with 0.2, 1.414, 0.1, and 0.0707. The Gabor filter convolution kernel functions were in six different orientations: 0, π / 6 , π / 3 , π / 2 , 2 π / 3 , and 5 π / 6 with the same frequency value.

3.3. Feature Selection for Reducing the Number of Features

In this paper, PSO was used for feature selection to decrease the data dimension, which is a kind of swarm intelligence algorithm using a group of particles. It has been noted that members of a group seem to share information among themselves, which is a fact that leads to increased efficiency of the group. A particle moves toward the optimum based on its present velocity, its previous experience, and the experience of its neighbors. In an n-Dimensional search space, the position and velocity of the i-th particle are represented as vectors X i = x i 1 , . . . , x i n and V i = v i 1 , . . . , v i n . Let P b e s t i and G b e s t be the best position of the i-th particle and the group’s best position so far, respectively. The velocity and position of each particle are updated as follows [39]:
V i k + 1 = ω · V i k + r 1 · c 1 · ( P b e s t i k X i k ) + r 2 · c 2 · ( G b e s t k X i k )
X i k + 1 = X i k + V i k + 1
where V i k is the velocity of the i-th particle at iteration k, ω is the inertia weight factor, c 1 and c 2 are the acceleration coefficients, r 1 and r 2 are random numbers between zero and one, and X i k is the position of the i-th particle at iteration k. In the velocity updating process, the values of the parameters such as ω , c 1 , and c 2 should be determined in advance, which makes it cumbersome to solve large-scale optimization problems.
However, decimal coding may not be suitable for discrete optimization such as feature selection; thus, the position vector of a particle should be coded as a binary form. The velocity of the i-th element in the i-th particle is related to the possibility that the position of the particle takes a value of one or zero. It is implemented by defining an intermediate variable S ( v i j k + 1 ) , called a sigmoid limiting transformation, as follows [40,41]:
S ( v i j k + 1 ) = 1 1 + e x p ( v i j k + 1 )
The value of S ( v i j k + 1 ) can be interpreted as a probability threshold. If a random number selected from a uniform distribution in [0,1] is less than the threshold, the value of the position of the j-th element in the i-th particle at iteration k + 1 (i.e., x i j k + 1 ) is set to one, and otherwise to zero, and the position vector is replaced as follows:
x i j k + 1 = 1 i f r a n d < S ( v i j k + 1 ) 0 o t h e r w i s e
where r a n d denotes random numbers uniformly distributed between zero and one; S ( v i j k + 1 ) is a sigmoid limiting transformation.
In this paper, PSO was used to extract as high an accuracy as possible with few features. To improve the training process, the feature combination was adjusted by PSO, and the optimized results could be obtained by choosing the feature combination with the minimum error as the most suitable one [42]. Finally, a reasonable combination of point cloud and texture features was obtained for building extraction, and the whole process is shown in Figure 2.

3.4. Definition of the Objective Function

To obtain the results of high extraction accuracy and reduce the number of features, an objective function was defined as an auxiliary in this paper. As the Fisher discriminant criterion has been shown to have good performance in building extraction and other extraction problems that include two categories, maximizing the differences between classes and minimizing the differences within classes, and accurately identifying the target category from other classes, it was used to define the objective function for feature selection [43]. The formula of the objective function is expressed as follows:
f i t = ( μ 1 μ 2 ) 2 ( σ 1 2 + σ 2 2 ) · n
where f i t represents the value of the objective function, μ 1 and μ 2 are the eigen mean vectors of two types of objects, σ 1 and σ 2 are the eigen variance vectors of two types of objects, respectively, and n is the number of points. The output of point cloud features was the vectors. The texture features were in the form of a 2D image. They can be converted into vectors, and finally, these two kinds of features can be merged into a vector, while a higher feature vector dimension of each point can be obtained by combining the two vectors. Besides, the larger value of the objective function demonstrated better quality of classification.

3.5. Implementation of the Proposed Method

The proposed method was easy to implement, and the key issues of building extraction were the fusion of point cloud and texture features, as well as feature selection. The process of the proposed method is shown as follows:
  • Step 1: Input the testing images, and compute the feature vectors of the point cloud. Generate elevation maps, and extract texture features via the Gabor filter from them.
  • Step 2: Build the training and testing samples based on the fusion of point cloud and texture features;
  • Step 3: Randomly generate the initial population of PSO in the range of −10–10 via decimal coding, and transform it into binary coding;
  • Step 4: Conduct building extraction, and compute the fitness value of each particle by Equation (9);
  • Step 5: Operation of PSO:
    • Step 5-1: Update the velocity of each particle by using Equation (5);
    • Step 5-2: Switch the population into the form of binary coding by Equation (8);
  • Step 6: Conduct building extraction, and compute the fitness value of each particle by Equation (9);
  • Step 7: If the solution is better, replace the current particle; otherwise, the particle does not change, and then, find the current global best solution;
  • Step 8: Judge whether the maximum number of iterations is reached, and if it is, go to Step 9; otherwise, go to Step 5;
  • Step 9: Output the optimal feature combination, and compare it with other building extraction methods via the extraction accuracy.

4. Experimental Results and Discussion

The experimental environment in this study was a computer with a 2.30-GHz CPU and 8 G of RAM. The data-processing operation was realized using MATLAB 2016a and VS2017 software. The manual extraction process was accomplished using LiDAR software and visual interpretation by researchers with relevant working experience.

4.1. Experimental Platform and Data Information

The data used in this study were point cloud data obtained from a Riegl LMS-Q780 laser scanner in Fuzhou, China. The experimental data included five non-overlapping urban areas, which contained buildings, vegetation, and other types of objects. Since the high density of the experimental point cloud may result in a large amount of calculation, it was necessary to down-sample the data in order to reduce the amount of calculation. According to the density of the point cloud after down-sampling, the data areas were divided into Low-Density Region 1 (LDR 1), LDR 2, the medium-density region (MDR), High-Density Region 1 (HDR 1), and HDR 2. Details on the experimental data are shown in Table 2.
The experimental data were colored according to the elevation rendering, and the results of the manual extraction are shown in Figure 3, Figure 4, Figure 5, Figure 6 and Figure 7.

4.2. Extraction of Texture Features

The process of extracting texture features using the Gabor filter in this study is shown below:
Figure 8 shows the process of the Gabor filter, where it is formed on the basis of different values of the orientation and frequency parameters. Different texture features can be yielded after elevation map convoluting with templates. A group of parameter combination results is shown in Figure 8 with the same frequency value of 0.2, and the orientation varied from 0– 5 π / 6 via steps of π / 6 , while the local display of the common part is also shown on the right side. It can be concluded that variation of the parameter combination caused the change of the convolution module and resulted in differences in texture features, especially on the edge and corner of the buildings.

4.3. Comparative Analysis and Accuracy Evaluation of Building Extraction

In order to prove the effectiveness of the proposed method, the experimental results were compared with those obtained using GLCM, LBP, and HoG for texture feature extraction. Those of building extraction based only on point cloud features (OPCF), building extraction with no feature selection (NFS), and building extraction using ENVI software were also compared in this paper. The extraction accuracy of different methods is shown in Table 3, and the building extraction locations of different experimental areas are shown in Figure 9, Figure 10, Figure 11, Figure 12 and Figure 13.
From Table 3, it can be seen that the extraction accuracy with the proposed method was superior to other texture feature extraction methods and ENVI software. For HDR 2, the extraction accuracy of the proposed method was over 10% higher than that of GLCM HoG and LBP. Especially for LDR 2 and HDR 2, the extraction accuracy of GLCM was only around 60%, while the proposed method could still achieve an extraction accuracy higher than 87%. Although the extraction accuracy by using ENVI software exceeded 80%, and even reached 90% for LDR 2 and HDR 1, the extraction accuracy of the proposed method could still be 1.2874% higher than ENVI software. Comparing with the result of NFS, this suggested that feature selection benefited the building extraction by improving its extraction accuracy and efficiency. In all, when the follow-up operations were the same, the final results obtained by the Gabor filter applied in this paper were more accurate than those of other texture feature extraction methods. After feature selection, not only the extraction accuracy was higher, but also the computational time was shorter, as the data dimension was decreased. Besides, using about 10 features, we were able to achieve such satisfactory results.
As shown in Figure 9, Figure 10, Figure 11, Figure 12 and Figure 13, the experimental results of the proposed method were superior to other texture extraction methods, such as GLCM, HoG, and LBP, as well as NFS, OPCF, and ENVI software in the five experimental areas, as it generated a lower number of errors in the building’s interior area. In addition, the proposed method preserved shape better and the interior integrity of the building. HoG and NFS were incapable of extracting complete buildings in LDR1, and the proposed method was better at preserving the integrity of the large complex building on the top left corner than LBP. For LDR2, GLCM and LBP were unable to be applied for building extraction, and the proposed method produced more correct results obviously, especially in the red circle in Figure 10g, than other methods. For MDR, only ENVI software and the proposed method extracted the complete building in the red circle in Figure 11g. However, less points were extracted as building points by the proposed method in other non-building areas than ENVI software. For HDR1, all of the methods, except NFS, obtained good results in most of the test area. However, more non-building points were obviously extracted as building points in the area of the red circle, which is shown in Figure 12d, and there were also some discrete errors in the non-building areas for other methods, while the proposed method obtained better extraction results, as Figure 12g shows. Furthermore, more buildings were extracted correctly by GLCM, LBP, and OPCF than other methods in HDR2, and for the proposed method, this was less in the red circle areas in Figure 13g.

5. Conclusions

This paper presented a building extraction method based on the fusion of point cloud and texture features, by calculating the feature values, elevation, and density of the point cloud and transforming the point cloud into an elevation map. The Gabor filter was used to extract texture features based on the elevation map, and the features could be assigned to the point cloud again. Then, point cloud and texture features were fused, and feature selection was done to realize more accurate and efficient building extraction. The experiments showed that the fusion of point cloud and texture features was able to obtain higher extraction accuracy than other methods. Besides, because of the large number of features, PSO was used to select a better feature combination to realize building extraction from the point cloud. Compared with the results from other building extraction methods, as well as NFS, OPCF, and ENVI software, the extraction accuracy by using the proposed method could satisfy practical applications preferably. In summary, the proposed method was proven to be efficient and valid for building extraction, with satisfactory extraction accuracy, which always exceeded 87%. It could provide a convenient and effective way to extract buildings in urban areas. On the basis of this work, future work will be performed on the optimization of the texture feature extraction method in the entire data-processing process.

Author Contributions

Conceptualization, X.L. and M.W.; Methodology, J.Y.; Software, J.Y. and M.W.; Validation, X.L., M.W., J.Y. and Y.L.; Formal analysis, X.L. and M.W.; Investigation, Y.L. and J.Y.; Resources, X.L. and M.W.; Data curation, X.L.; Writing—original draft preparation, X.L. and J.Y.; Writing—review and editing, X.L., M.W., J.Y. and Y.L.; Visualization, J.Y.; Supervision, X.L.; Project administration, X.L.; Funding acquisition, X.L.

Funding

This work was funded by the National Key Research & Development Program of China under Grant No. 41771368, the Key Laboratory for National Geographic Census and Monitoring, National Administration of Surveying, Mapping and Geoinformation under Grant No. 2018NGCM06, the Technical Research Service of Airborne LiDAR Data Acquisition and Digital Elevation Model Updating Project in Guangdong Province under Grant No. 0612-1841D0330175, and the Airborne LiDAR Data Acquisition and Digital Elevation Model Updating in Guangdong Province under Grant No. GPCGD173109FG317F.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Li, X.; Ling, F.; Foody, G.M.; Du, Y. A Superresolution Land-Cover Change Detection Method Using Remotely Sensed Images with Different Spatial Resolutions. IEEE Trans. Geosci. Remote Sens. 2016, 54, 3822–3841. [Google Scholar] [CrossRef]
  2. Zhang, L.; Zhang, L.; Du, B. Deep Learning for Remote Sensing Data: A Technical Tutorial on the State of the Art. IEEE Geosci. Remote Sens. Mag. 2016, 4, 22–40. [Google Scholar] [CrossRef]
  3. Panagiotidis, D.; Abdollahnejad, A.; Chiteculo, V. Determining Tree Height and Crown Diameter from High-resolution UAV Imagery. Int. J. Remote Sens. 2017, 38, 2392–2410. [Google Scholar] [CrossRef]
  4. Marullo, S.; Patsaeva, S.; Fiorani, L. Remote sensing of the coastal zone of the European seas. Int. J. Remote Sens. 2018, 39, 9313–9316. [Google Scholar] [CrossRef]
  5. Zhang, J.; Chen, L.; Wang, C.; Zhuo, L.; Tian, Q.; Liang, X. Road Recognition From Remote Sensing Imagery Using Incremental Learning. IEEE Trans. Intell. Transp. Syst. 2017, 99, 1–13. [Google Scholar] [CrossRef]
  6. Huang, X.; Yuan, W.; Li, J.; Zhang, L. A New Building Extraction Postprocessing Framework for High-Spatial-Resolution Remote-Sensing Imagery. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2017, 10, 654–668. [Google Scholar] [CrossRef]
  7. Li, E.; Xu, S.; Meng, W.; Zhang, X. Building Extraction from Remotely Sensed Images by Integrating Saliency Cue. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2017, 10, 906–919. [Google Scholar] [CrossRef]
  8. Magnússon, E.; Belart, J.; Pálsson, F.; Ágústsson, H.; Crochet, P. Geodetic Mass Balance Record with Rigorous Uncertainty Estimates Deducedfrom Aerial Photographs and LiDAR Data–Case Study from Drangajökull Icecap, NW Iceland. Cryosphere 2016, 9, 4733–4785. [Google Scholar] [CrossRef]
  9. Höfler, V.; Wessollek, C.; Karrasch, P. Knowledge-Based Modelling of Historical Surfaces Using LiDAR Data. Earth Resour. Environ. Remote Sens./GIS Appl. VII 2016, 1–11. [Google Scholar] [CrossRef]
  10. Harmon, J.M.; Leone, M.P.; Prince, S.D.; Snyder, M. LiDAR for Archaeological Landscape Analysis: A Case Study of Two Eighteenth-Century Maryland Plantation Sites. Am. Antiq. 2017, 71, 649–670. [Google Scholar] [CrossRef]
  11. Baek, N.; Shin, W.S.; Kim, K.J. Geometric primitive extraction from LiDAR-scanned point clouds. Clust. Comput. 2017, 20, 741–748. [Google Scholar] [CrossRef]
  12. Rozsa, Z.; Sziranyi, T. Obstacle Prediction for Automated Guided Vehicles Based on Point Clouds Measured by a Tilted LiDAR Sensor. IEEE Trans. Intell. Transp. Syst. 2018, 99, 1–13. [Google Scholar] [CrossRef]
  13. Zheng, Y.; Weng, Q.; Zheng, Y. A Hybrid Approach for Three-Dimensional Building Reconstruction in Indianapolis from LiDAR Data. Remote Sens. 2017, 9, 310. [Google Scholar] [CrossRef]
  14. Wang, Y.; Cheng, L.; Chen, Y.; Wu, Y.; Li, M. Building Point Detection from Vehicle-Borne LiDAR Data Based on Voxel Group and Horizontal Hollow Analysis. Remote Sens. 2016, 8, 419. [Google Scholar] [CrossRef]
  15. Qin, Y.; Li, S.; Vu, T.T.; Niu, Z.; Ban, Y. Synergistic Application of Geometric and Radiometric Features of LiDAR Data for Urban Land Cover Mapping. Opt. Express 2015, 23, 13761–13775. [Google Scholar] [CrossRef] [PubMed]
  16. Zhao, Z.; Duan, Y.; Zhang, Y.; Cao, R. Extracting Buildings from and Regularizing Boundaries in Airborne liDAR Data Using Connected Operators. Int. J. Remote Sens. 2016, 37, 889–912. [Google Scholar] [CrossRef]
  17. Huang, R.; Yang, B.; Liang, F.; Dai, W.; Li, J.; Tian, M.; Xu, W. A top-down Strategy for Buildings Extraction from Complex Urban Scenes Using Airborne LiDAR Point Clouds. Infrared Phys. Technol. 2018, 92, 203–218. [Google Scholar] [CrossRef]
  18. Yi, C.; Zhang, Y.; Wu, Q.; Xu, Y.; Remil, O.; Wei, M.; Wang, J. Urban Building Reconstruction from Raw LiDAR Point Data. Comput.-Aided Des. 2017, 93, 1–14. [Google Scholar] [CrossRef]
  19. Siddiqui, F.; Teng, S.; Awrangjeb, M.; Lu, G. A Robust Gradient Based Method for Building Extraction from LiDAR and Photogrammetric Imagery. Sensors 2016, 16, 1110. [Google Scholar] [CrossRef]
  20. Liu, C.; Wang, X.; Huang, H.; Gong, P.; Wu, D.; Jiang, J. The Importance of Data Type, Laser Spot Density and Modelling Method for Vegetation Height Mapping in Continental China. Int. J. Remote Sens. 2016, 37, 6127–6148. [Google Scholar] [CrossRef]
  21. Kang, H.; Sim, Y.; Han, J. Terrain Rendering with Unlimited Detail and Resolution. Graph. Models 2018, 97, 64–79. [Google Scholar] [CrossRef]
  22. He, Y.; Chen, L.; Chen, J.; Li, M. A Novel Way to Organize 3D LiDAR Point Cloud as 2D Depth Map Height Map and Surface Normal Map. In Proceedings of the IEEE International Conference on Robotics and Biomimetics (ROBIO), Zhuhai, China, 6–9 December 2015; pp. 1383–1388. [Google Scholar]
  23. Konstantinidis, D.; Stathaki, T.; Argyriou, V.; Grammalidis, N. Building Detection Using Enhanced HoG–LBP Features and Region Refinement Processes. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2017, 10, 888–905. [Google Scholar] [CrossRef]
  24. Cao, J.; Pang, Y.; Li, X. Learning Multilayer Channel Features for Pedestrian Detection. IEEE Trans. Image Process. 2017, 26, 3210–3220. [Google Scholar] [CrossRef] [PubMed]
  25. Du, S.; Zhang, Y.; Zou, Z.; Xu, S.; He, X.; Chen, S. Automatic Building Extraction from LiDAR Data Fusion of Point and Grid-based Features. ISPRS J. Photogramm. Remote Sens. 2017, 130, 294–307. [Google Scholar] [CrossRef]
  26. Niemi, M.T.; Vastaranta, M.; Vauhkonen, J.; Melkas, T.; Holopainen, M. Airborne LiDAR-derived Eelevation Data in Terrain Trafficability Mapping. Scand. J. For. Res. 2017, 32, 761–773. [Google Scholar] [CrossRef]
  27. Alatas, B. Sports Inspired Computational Intelligence Algorithms for Global Optimization. Artif. Intell. Rev. 2017, 12, 1–49. [Google Scholar] [CrossRef]
  28. Li, C.; Wei, W.; Li, J.; Song, W. A Cloud-based Monitoring System via Face Recognition Using Gabor and CS-LBP Features. J. Supercomput. 2017, 73, 1532–1546. [Google Scholar] [CrossRef]
  29. Kaggwa, F.; Ngubiri, J.; Tushabe, F. Combined Feature Level and Score Level Fusion Gabor Filter-Based Multiple Enrollment Fingerprint Recognition. Int. Conf. Signal Process. 2017, 159–165. [Google Scholar] [CrossRef]
  30. Kim, J.; Um, S.; Min, D. Fast 2D Complex Gabor Filter with Kernel Decomposition. IEEE Trans. Image Process. 2018, 27, 1713–1722. [Google Scholar] [CrossRef]
  31. Luan, S.; Chen, C.; Zhang, B.; Han, J.; Liu, J. Gabor Convolutional Networks. IEEE Trans. Image Process. 2017, 99, 4357–4366. [Google Scholar]
  32. Karanam, S.; Gou, M.; Wu, Z.; Rates-Borras, A.; Camps, O.; Radke, R.J. A Systematic Evaluation and Benchmark for Person Re-Identification: Features, Metrics, and Datasets. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 31, 523–536. [Google Scholar] [CrossRef] [PubMed]
  33. Thanou, D.; Chou, P.; Frossard, P. Graph-Based Compression of Dynamic 3D Point Cloud Sequences. IEEE Trans. Image Process. 2016, 25, 1765–1778. [Google Scholar] [CrossRef] [PubMed]
  34. Song, W.; Lei, Y.; Chen, S.; Pan, Z.; Yang, J.J.; Pan, H.; Du, X.; Cai, W.; Wang, Q. Multiple Facial Image Features-based Recognition for The Automatic Diagnosis of Turner Syndrome. Comput. Ind. 2018, 100, 85–95. [Google Scholar] [CrossRef]
  35. Meng, F.; Wang, X.; Shao, F.; Wang, D.; Hua, X. Energy-Efficient Gabor Kernels in Neural Networks with Genetic Algorithm Training Method. Electronics 2019, 8, 105. [Google Scholar] [CrossRef]
  36. Lei, H.; Jiang, G.; Quan, L. Fast Descriptors and Correspondence Propagation for Robust Global Point Cloud Registration. IEEE Trans. Image Process. 2017, 26, 3614–3623. [Google Scholar] [CrossRef]
  37. Fu, Y.; Chiang, H.D. Toward Optimal Multiperiod Network Reconfiguration for Increasing the Hosting Capacity of Distribution Networks. IEEE Trans. Power Deliv. 2018, 33, 2294–2304. [Google Scholar] [CrossRef]
  38. Yang, J.; Cao, Z.; Qian, Z. A Fast and Robust Local Fescriptor for 3D Point Cloud Registration. Inf. Sci. 2016, 346, 163–179. [Google Scholar] [CrossRef]
  39. Hasanipanah, M.; Amnieh, H.B.; Arab, H.; Zamzam, M.S. Feasibility of PSO–ANFIS model to Estimate Rock Fragmentation Produced by Mine Blasting. Neural Comput. Appl. 2018, 30, 1015–1024. [Google Scholar] [CrossRef]
  40. Lin, J.C.W.; Yang, L.; Fournier-Viger, P.; Hong, T.P.; Voznak, M. A Binary PSO Approach to Mine High-utility Itemsets. Soft Comput. 2017, 21, 5103–5121. [Google Scholar] [CrossRef]
  41. Wang, M.; Wu, C.; Wang, L.; Xiang, D.; Huang, X. A feature selection approach for hyperspectral image based on modified ant lion optimizer. Knowl.-Based Syst. 2019, 168, 39–48. [Google Scholar] [CrossRef]
  42. Phan, A.; Nguyen, M.; Bui, L. Feature weighting and SVM parameters optimization based on genetic algorithms for classification problems. Appl. Intell. 2016, 46, 455–469. [Google Scholar] [CrossRef]
  43. Wan, Y.; Wang, M.; Ye, Z.; Lai, X. A “Tuned” Mask Learnt Approach Based on Gravitational Search Algorithm. Comput. Intell. Neurosci. 2016, 2016, 1–16. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Point cloud rendering image (left) and elevation map (right).
Figure 1. Point cloud rendering image (left) and elevation map (right).
Remotesensing 11 01636 g001
Figure 2. The process to obtain the optimal combination of features.
Figure 2. The process to obtain the optimal combination of features.
Remotesensing 11 01636 g002
Figure 3. LDR 1 elevation coloration and manual extraction results.
Figure 3. LDR 1 elevation coloration and manual extraction results.
Remotesensing 11 01636 g003
Figure 4. LDR 2 elevation coloration and manual extraction results.
Figure 4. LDR 2 elevation coloration and manual extraction results.
Remotesensing 11 01636 g004
Figure 5. MDR elevation coloration and manual extraction results.
Figure 5. MDR elevation coloration and manual extraction results.
Remotesensing 11 01636 g005
Figure 6. HDR 1 elevation coloration and manual extraction results.
Figure 6. HDR 1 elevation coloration and manual extraction results.
Remotesensing 11 01636 g006
Figure 7. HDR 2 elevation coloration and manual extraction results.
Figure 7. HDR 2 elevation coloration and manual extraction results.
Remotesensing 11 01636 g007
Figure 8. Texture features extraction using the Gabor filter.
Figure 8. Texture features extraction using the Gabor filter.
Remotesensing 11 01636 g008
Figure 9. Building extraction results of LDR 1.
Figure 9. Building extraction results of LDR 1.
Remotesensing 11 01636 g009
Figure 10. Building extraction results of LDR 2.
Figure 10. Building extraction results of LDR 2.
Remotesensing 11 01636 g010
Figure 11. Building extraction results of MDR.
Figure 11. Building extraction results of MDR.
Remotesensing 11 01636 g011
Figure 12. Building extraction results of HDR 1.
Figure 12. Building extraction results of HDR 1.
Remotesensing 11 01636 g012
Figure 13. Building extraction results of HDR 2.
Figure 13. Building extraction results of HDR 2.
Remotesensing 11 01636 g013
Table 1. Point cloud features.
Table 1. Point cloud features.
CategoryNameAbbreviationMeaningFormula
Eigenvalue-based featuresSumSUSum of eigenvalues λ 1 + λ 2 + λ 3
Total varianceTVTotal variance ( λ 1 λ 2 λ 3 ) 1 / 3
Eigen entropyEICharacteristic entropy i = 3 3 λ i · I n ( λ i )
AnisotropyANAnisotropy ( λ 1 λ 3 ) / λ 1
PlanarityPLPlanarity ( λ 2 λ 3 ) / λ 1
LinearityLILinearity ( λ 1 λ 2 ) / λ 1
Surface roughnessSRSurface roughness λ 3 / ( λ 1 + λ 2 + λ 3 )
SphericitySPSphericity λ 3 / λ 1
Density-based featurePoint DensityPDPoint Density 0.75 N 3 D π r 3
Elevation-based featuresHeight aboveHAThe height difference between the
current point and the lowest point
Z Z m i n
Height belowHBThe height difference between the
highest point and the current point
Z m a x Z
Sphere VarianceSPVStandard deviation of the height
difference in the spherical neighborhood
i = 1 n ( Z i Z a v e ) 2 n 1
Table 2. Experimental data information. LDR, low-density region; MDR, medium-density region; HDR, high-density region.
Table 2. Experimental data information. LDR, low-density region; MDR, medium-density region; HDR, high-density region.
Experimental DataData AreaNumber of PointsPoint Cloud Density
( m 2 ) Original DataAfter DilutionOriginal DataAfter Dilution
LDR 1174,0804,486,76319,32025.7993390.111040
LDR 2155,5953,989,31021,92625.6836310.140958
MDR186,147585,02423,67526.2615920.183575
HDR 199,4702,283,27529,12723.0621700.294197
HDR 268,0401,897,76020,66327.9361710.303810
Table 3. Comparison of the experimental results with other methods for building extraction (%). OPCF, only point cloud features; NFS, no feature selection.
Table 3. Comparison of the experimental results with other methods for building extraction (%). OPCF, only point cloud features; NFS, no feature selection.
Experimental DataGLCMHoGLBPOPCFNFSENVIProposed
LDR 186.998475.950388.387080.458678.733087.420390.4238
LDR 265.552385.186574.529785.565189.694991.331092.2558
MDR75.890278.935673.334781.702282.352783.518087.1679
HDR 187.506490.826490.047087.496181.604790.266092.1138
HDR 262.391776.697575.279579.436784.276286.275289.1207

Share and Cite

MDPI and ACS Style

Lai, X.; Yang, J.; Li, Y.; Wang, M. A Building Extraction Approach Based on the Fusion of LiDAR Point Cloud and Elevation Map Texture Features. Remote Sens. 2019, 11, 1636. https://doi.org/10.3390/rs11141636

AMA Style

Lai X, Yang J, Li Y, Wang M. A Building Extraction Approach Based on the Fusion of LiDAR Point Cloud and Elevation Map Texture Features. Remote Sensing. 2019; 11(14):1636. https://doi.org/10.3390/rs11141636

Chicago/Turabian Style

Lai, Xudong, Jingru Yang, Yongxu Li, and Mingwei Wang. 2019. "A Building Extraction Approach Based on the Fusion of LiDAR Point Cloud and Elevation Map Texture Features" Remote Sensing 11, no. 14: 1636. https://doi.org/10.3390/rs11141636

APA Style

Lai, X., Yang, J., Li, Y., & Wang, M. (2019). A Building Extraction Approach Based on the Fusion of LiDAR Point Cloud and Elevation Map Texture Features. Remote Sensing, 11(14), 1636. https://doi.org/10.3390/rs11141636

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop