remotesensing-logo

Journal Browser

Journal Browser

Automatic Segmentation, Reconstruction, and Modelling from Laser Scanning Data

A special issue of Remote Sensing (ISSN 2072-4292). This special issue belongs to the section "Remote Sensing Image Processing".

Deadline for manuscript submissions: 15 January 2025 | Viewed by 12963

Special Issue Editors


E-Mail Website
Guest Editor
Institute of Photogrammetry and Remote Sensing, Chinese Academy of Surveying & Mapping, Beijing 100830, China
Interests: LiDAR and hyperspectral remote sensing

E-Mail Website
Guest Editor
Faculty of Applied Science, University of British Columbia, Vancouver, BC, Canada
Interests: transportation engineering; highway design; traffic safety; infrastructure management; LiDAR

Special Issue Information

Dear Colleagues,

At present, with the development of unmanned aerial vehicles (UAVs), autonomous driving systems, and robot technology, laser scanning technology is seen as a critical component to the efficient operation of most of those systems. In the meantime, the miniaturization and highly integration trends of LiDAR components are becoming evident, while the performance of laser scanning systems has also been improved. This has resulted in an influx of massive, very high density and high precision point cloud data at a relatively low cost. The datasets contain an accurate 3D representation of the real world environment and can be collected on local and regional scales, from outdoor to indoor, and underground environments. These data sets have opened a broad range of new applications in a variety of disciplines, e.g., urban development, natural resource management, transportation, electric power, energy, and heritage conservation, for 3D scene modeling, automatic driving, high precision location navigation, facility and infrastructure management, etc.

One challenge when dealing with laser scanning data is that those datasets are unorganized and big data sets. Therefore, the efficient and automatic segmentation, classification, reconstruction and modelling of point clouds collected using laser scanning technology has been the focus of many research papers over the past few years. The identification and recognition of the different elements in a 3D scene is a challenging task due to the various scenarios and different data acquisition systems. The documented approaches, however, mainly focus on a certain kind of object or the detection of learned invariant surface shapes, e.g., street salient/street adjacent objects, modelling of building facades and roofs, detailed modelling of trees, while not necessarily suitable for arbitrary tasks due to the varying requirements of the different fields of research. In addition, many advanced AI and machine learning methods are required to be discussed in automatic segmentation, reconstruction and 3D modelling.

This Special Issue of Remote Sensing aims to attract innovative and well-documented article contributions showcasing recent achievements in the field of LiDAR point cloud segmentation, reconstruction and modelling applications, as well as to identify the obstacles still ahead. Submitted manuscripts may cover, although not limited to, topics related to:

  • LiDAR point cloud segmentation and reconstruction methods and algorithms;
  • Combining LiDAR point cloud and multispectral/hyperspectral image data for segmentation, reconstruction and modelling;
  • Machine/deep learning algorithms for point cloud segmentation and clustering;
  • Application of 3D reconstructed models generated from LiDAR point cloud data;
  • Quality assessment of the segmentation, reconstruction, and modelling process.

Prof. Dr. Zhengjun Liu
Dr. Suliman Gargoum
Guest Editors

Manuscript Submission Information

Manuscripts should be submitted online at www.mdpi.com by registering and logging in to this website. Once you are registered, click here to go to the submission form. Manuscripts can be submitted until the deadline. All submissions that pass pre-check are peer-reviewed. Accepted papers will be published continuously in the journal (as soon as accepted) and will be listed together on the special issue website. Research articles, review articles as well as short communications are invited. For planned papers, a title and short abstract (about 100 words) can be sent to the Editorial Office for announcement on this website.

Submitted manuscripts should not have been published previously, nor be under consideration for publication elsewhere (except conference proceedings papers). All manuscripts are thoroughly refereed through a single-blind peer-review process. A guide for authors and other relevant information for submission of manuscripts is available on the Instructions for Authors page. Remote Sensing is an international peer-reviewed open access semimonthly journal published by MDPI.

Please visit the Instructions for Authors page before submitting a manuscript. The Article Processing Charge (APC) for publication in this open access journal is 2700 CHF (Swiss Francs). Submitted papers should be well formatted and use good English. Authors may use MDPI's English editing service prior to publication or during author revisions.

Benefits of Publishing in a Special Issue

  • Ease of navigation: Grouping papers by topic helps scholars navigate broad scope journals more efficiently.
  • Greater discoverability: Special Issues support the reach and impact of scientific research. Articles in Special Issues are more discoverable and cited more frequently.
  • Expansion of research network: Special Issues facilitate connections among authors, fostering scientific collaborations.
  • External promotion: Articles in Special Issues are often promoted through the journal's social media, increasing their visibility.
  • e-Book format: Special Issues with more than 10 articles can be published as dedicated e-books, ensuring wide and rapid dissemination.

Further information on MDPI's Special Issue polices can be found here.

Published Papers (6 papers)

Order results
Result details
Select all
Export citation of selected articles as:

Research

Jump to: Other

18 pages, 9000 KiB  
Article
Multilevel Geometric Feature Embedding in Transformer Network for ALS Point Cloud Semantic Segmentation
by Zhuanxin Liang and Xudong Lai
Remote Sens. 2024, 16(18), 3386; https://doi.org/10.3390/rs16183386 - 12 Sep 2024
Viewed by 654
Abstract
Effective semantic segmentation of Airborne Laser Scanning (ALS) point clouds is a crucial field of study and influences subsequent point cloud application tasks. Transformer networks have made significant progress in 2D/3D computer vision tasks, exhibiting superior performance. We propose a multilevel geometric feature [...] Read more.
Effective semantic segmentation of Airborne Laser Scanning (ALS) point clouds is a crucial field of study and influences subsequent point cloud application tasks. Transformer networks have made significant progress in 2D/3D computer vision tasks, exhibiting superior performance. We propose a multilevel geometric feature embedding transformer network (MGFE-T), which aims to fully utilize the three-dimensional structural information carried by point clouds and enhance transformer performance in ALS point cloud semantic segmentation. In the encoding stage, compute the geometric features surrounding tee sampling points at each layer and embed them into the transformer workflow. To ensure that the receptive field of the self-attention mechanism and the geometric computation domain can maintain a consistent scale at each layer, we propose a fixed-radius dilated KNN (FR-DKNN) search method to address the limitation of traditional KNN search methods in considering domain radius. In the decoding stage, we aggregate prediction deviations at each level into a unified loss value, enabling multilevel supervision to improve the network’s feature learning ability at different levels. The MGFE-T network can predict the class label of each point in an end-to-end manner. Experiments were conducted on three widely used benchmark datasets. The results indicate that the MGFE-T network achieves superior OA and mF1 scores on the LASDU and DFC2019 datasets and performs well on the ISPRS dataset with imbalanced classes. Full article
Show Figures

Figure 1

27 pages, 5852 KiB  
Article
MVG-Net: LiDAR Point Cloud Semantic Segmentation Network Integrating Multi-View Images
by Yongchang Liu, Yawen Liu and Yansong Duan
Remote Sens. 2024, 16(15), 2821; https://doi.org/10.3390/rs16152821 - 31 Jul 2024
Viewed by 1129
Abstract
Deep learning techniques are increasingly applied to point cloud semantic segmentation, where single-modal point cloud often suffers from accuracy-limiting confusion phenomena. Moreover, some networks with image and LiDAR data lack an efficient fusion mechanism, and the occlusion of images may do harm to [...] Read more.
Deep learning techniques are increasingly applied to point cloud semantic segmentation, where single-modal point cloud often suffers from accuracy-limiting confusion phenomena. Moreover, some networks with image and LiDAR data lack an efficient fusion mechanism, and the occlusion of images may do harm to the segmentation accuracy of a point cloud. To overcome the above issues, we propose the integration of multi-modal data to enhance network performance, addressing the shortcomings of existing feature-fusion strategies that neglect crucial information and struggle with matching modal features effectively. This paper introduces the Multi-View Guided Point Cloud Semantic Segmentation Model (MVG-Net), which extracts multi-scale and multi-level features and contextual data from urban aerial images and LiDAR, and then employs a multi-view image feature-aggregation module to capture highly correlated texture information with the spatial and channel attentions of point-wise image features. Additionally, it incorporates a fusion module that uses image features to instruct point cloud features for stressing key information. We present a new dataset, WK2020, which combines multi-view oblique aerial images with LiDAR point cloud to validate segmentation efficacy. Our method demonstrates superior performance, especially in building segmentation, achieving an F1 score of 94.6% on the Vaihingen Dataset—the highest among the methods evaluated. Furthermore, MVG-Net surpasses other networks tested on the WK2020 Dataset. Compared to backbone network for single point modality, our model achieves overall accuracy improvement of 5.08%, average F1 score advancement of 6.87%, and mean Intersection over Union (mIoU) betterment of 7.9%. Full article
Show Figures

Figure 1

23 pages, 24808 KiB  
Article
Classification of Hyperspectral and LiDAR Data Using Multi-Modal Transformer Cascaded Fusion Net
by Shuo Wang, Chengchao Hou, Yiming Chen, Zhengjun Liu, Zhenbei Zhang and Geng Zhang
Remote Sens. 2023, 15(17), 4142; https://doi.org/10.3390/rs15174142 - 24 Aug 2023
Cited by 4 | Viewed by 1991
Abstract
With the continuous development of surface observation methods and technologies, we can acquire multiple sources of data more effectively in the same geographic area. The quality and availability of these data have also significantly improved. Consequently, how to better utilize multi-source data to [...] Read more.
With the continuous development of surface observation methods and technologies, we can acquire multiple sources of data more effectively in the same geographic area. The quality and availability of these data have also significantly improved. Consequently, how to better utilize multi-source data to represent ground information has become an important research question in the field of geoscience. In this paper, a novel model called multi-modal transformer cascaded fusion net (MMTCFN) is proposed for fusion and classification of multi-modal remote sensing data, Hyperspectral Imagery (HSI) and LiDAR data. Feature fusion and feature extraction are the two stages of the model. First, in the feature extraction stage, a three-branch cascaded Convolutional Neural Network (CNN) framework is employed to fully leverage the advantages of convolutional operators in extracting shallow-level local features. Based on this, we generated multi-modal long-range integrated deep features utilizing the transformer-based vectorized pixel group transformer (VPGT) module during the feature fusion stage. In the VPGT block, we designed a vectorized pixel group embedding that preserves the global features extracted from the three branches in a non-overlapping multi-space manner. Moreover, we introduce the DropKey mechanism into the multi-head self-attention (MHSA) to alleviate overfitting caused by insufficient training samples. Finally, we employ a probabilistic decision fusion strategy to integrate multiple class estimations, assigning a specific category to each pixel. This model was experimented on three HSI-LiDAR datasets with balanced and unbalanced training samples. The proposed model outperforms the other seven SOTA approaches in terms of OA performance, proving the superiority of MMTCFN for the HSI-LiDAR classification task. Full article
Show Figures

Figure 1

20 pages, 20731 KiB  
Article
Sharp Feature-Preserving 3D Mesh Reconstruction from Point Clouds Based on Primitive Detection
by Qi Liu, Shibiao Xu, Jun Xiao and Ying Wang
Remote Sens. 2023, 15(12), 3155; https://doi.org/10.3390/rs15123155 - 16 Jun 2023
Cited by 4 | Viewed by 3683
Abstract
High-fidelity mesh reconstruction from point clouds has long been a fundamental research topic in computer vision and computer graphics. Traditional methods require dense triangle meshes to achieve high fidelity, but excessively dense triangles may lead to unnecessary storage and computational burdens, while also [...] Read more.
High-fidelity mesh reconstruction from point clouds has long been a fundamental research topic in computer vision and computer graphics. Traditional methods require dense triangle meshes to achieve high fidelity, but excessively dense triangles may lead to unnecessary storage and computational burdens, while also struggling to capture clear, sharp, and continuous edges. This paper argues that the key to high-fidelity reconstruction lies in preserving sharp features. Therefore, we introduce a novel sharp-feature-preserving reconstruction framework based on primitive detection. It includes an improved deep-learning-based primitive detection module and two novel mesh splitting and selection modules that we propose. Our framework can accurately and reasonably segment primitive patches, fit meshes in each patch, and split overlapping meshes at the triangle level to ensure true sharpness while obtaining lightweight mesh models. Quantitative and visual experimental results demonstrate that our framework outperforms both the state-of-the-art learning-based primitive detection methods and traditional reconstruction methods. Moreover, our designed modules are plug-and-play, which not only apply to learning-based primitive detectors but also can be combined with other point cloud processing tasks such as edge extraction or random sample consensus (RANSAC) to achieve high-fidelity results. Full article
Show Figures

Figure 1

21 pages, 7827 KiB  
Article
Framework for Geometric Information Extraction and Digital Modeling from LiDAR Data of Road Scenarios
by Yuchen Wang, Weicheng Wang, Jinzhou Liu, Tianheng Chen, Shuyi Wang, Bin Yu and Xiaochun Qin
Remote Sens. 2023, 15(3), 576; https://doi.org/10.3390/rs15030576 - 18 Jan 2023
Cited by 14 | Viewed by 3090
Abstract
Road geometric information and a digital model based on light detection and ranging (LiDAR) can perform accurate geometric inventories and three-dimensional (3D) descriptions for as-built roads and infrastructures. However, unorganized point clouds and complex road scenarios would reduce the accuracy of geometric information [...] Read more.
Road geometric information and a digital model based on light detection and ranging (LiDAR) can perform accurate geometric inventories and three-dimensional (3D) descriptions for as-built roads and infrastructures. However, unorganized point clouds and complex road scenarios would reduce the accuracy of geometric information extraction and digital modeling. There is a standardization need for information extraction and 3D model construction that integrates point cloud processing and digital modeling. This paper develops a framework from semantic segmentation to geometric information extraction and digital modeling based on LiDAR data. A semantic segmentation network is improved for the purpose of dividing the road surface and infrastructure. The road boundary and centerline are extracted by the alpha-shape and Voronoi diagram methods based on the semantic segmentation results. The road geometric information is obtained by a coordinate transformation matrix and the least square method. Subsequently, adaptive road components are constructed using Revit software. Thereafter, the road route, road entity model, and various infrastructure components are generated by the extracted geometric information through Dynamo and Revit software. Finally, a detailed digital model of the road scenario is developed. The Toronto-3D and Semantic3D datasets are utilized for analysis through training and testing. The overall accuracy (OA) of the proposed net for the two datasets is 95.3 and 95.0%, whereas the IoU of segmented road surfaces is 95.7 and 97.9%. This indicates that the proposed net could accomplish superior performance for semantic segmentation of point clouds. The mean absolute errors between the extracted and manually measured geometric information are marginal. This demonstrates the effectiveness and accuracy of the proposed extraction methods. Thus, the proposed framework could provide a reference for accurate extraction and modeling from LiDAR data. Full article
Show Figures

Graphical abstract

Other

Jump to: Research

14 pages, 3371 KiB  
Technical Note
Pose Estimation Based on Bidirectional Visual–Inertial Odometry with 3D LiDAR (BV-LIO)
by Gang Peng, Qiang Gao, Yue Xu, Jianfeng Li, Zhang Deng and Cong Li
Remote Sens. 2024, 16(16), 2970; https://doi.org/10.3390/rs16162970 - 14 Aug 2024
Viewed by 1122
Abstract
Due to the limitation of a single sensor such as only camera or only LiDAR, the Visual SLAM detects few effective features in the case of poor lighting or no texture. The LiDAR SLAM will also degrade in an unstructured environment and open [...] Read more.
Due to the limitation of a single sensor such as only camera or only LiDAR, the Visual SLAM detects few effective features in the case of poor lighting or no texture. The LiDAR SLAM will also degrade in an unstructured environment and open spaces, which reduces the accuracy of pose estimation and the quality of mapping. In order to solve this problem, on account of the high efficiency of Visual odometry and the high accuracy of LiDAR odometry, this paper investigates the multi-sensor fusion of bidirectional visual–inertial odometry with 3D LiDAR for pose estimation. This method can couple the IMU with the bidirectional vision respectively, and the LiDAR odometry is obtained assisted by the bidirectional visual inertial. The factor graph optimization is constructed, which effectively improves the accuracy of pose estimation. The algorithm in this paper is compared with LIO-LOAM, LeGO-LOAM, VINS-Mono, and so on using challenging datasets such as KITTI and M2DGR. The results show that this method effectively improves the accuracy of pose estimation and has high application value for mobile robots. Full article
Show Figures

Figure 1

Back to TopTop