Next Article in Journal
Weekly External Load Correlation in Season Microcycles with Game Running Performance and Training Quantification in Elite Young Soccer Players
Next Article in Special Issue
Mix-VIO: A Visual Inertial Odometry Based on a Hybrid Tracking Strategy
Previous Article in Journal
Multi-Resolution Learning and Semantic Edge Enhancement for Super-Resolution Semantic Segmentation of Urban Scene Images
Previous Article in Special Issue
Place Recognition through Multiple LiDAR Scans Based on the Hidden Markov Model
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

YPL-SLAM: A Simultaneous Localization and Mapping Algorithm for Point–line Fusion in Dynamic Environments

1
College of Agricultural Equipment Engineering, Henan University of Science and Technology, Luoyang 471003, China
2
Longmen Laboratory, Luoyang 471000, China
3
Collaborative Innovation Center of Machinery Equipment Advanced Manufacturing of Henan Province, Luoyang 471003, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(14), 4517; https://doi.org/10.3390/s24144517
Submission received: 29 April 2024 / Revised: 3 June 2024 / Accepted: 3 July 2024 / Published: 12 July 2024

Abstract

:
Simultaneous Localization and Mapping (SLAM) is one of the key technologies with which to address the autonomous navigation of mobile robots, utilizing environmental features to determine a robot’s position and create a map of its surroundings. Currently, visual SLAM algorithms typically yield precise and dependable outcomes in static environments, and many algorithms opt to filter out the feature points in dynamic regions. However, when there is an increase in the number of dynamic objects within the camera’s view, this approach might result in decreased accuracy or tracking failures. Therefore, this study proposes a solution called YPL-SLAM based on ORB-SLAM2. The solution adds a target recognition and region segmentation module to determine the dynamic region, potential dynamic region, and static region; determines the state of the potential dynamic region using the RANSAC method with polar geometric constraints; and removes the dynamic feature points. It then extracts the line features of the non-dynamic region and finally performs the point–line fusion optimization process using a weighted fusion strategy, considering the image dynamic score and the number of successful feature point–line matches, thus ensuring the system’s robustness and accuracy. A large number of experiments have been conducted using the publicly available TUM dataset to compare YPL-SLAM with globally leading SLAM algorithms. The results demonstrate that the new algorithm surpasses ORB-SLAM2 in terms of accuracy (with a maximum improvement of 96.1%) while also exhibiting a significantly enhanced operating speed compared to Dyna-SLAM.

1. Introduction

Visual Simultaneous Localization and Mapping (V-SLAM) plays a critical role in the autonomous navigation of mobile robots. With advancements in robotics, an increasing number of researchers are directing their focus towards SLAM technology [1,2,3]. Classical vision SLAM systems typically include sensor data acquisition, visual odometry, backend optimization, map construction, and loop closure detection within their frameworks [4]; these include PTAM [5], ORB-SLAM2 [6], ORB-SLAM 3 [7], VINS-Mono [8], VINS-Fusion [9], etc. These algorithms are relatively stable in static environments; however, in dynamic environments, the interference of dynamic objects can trigger map drift and complicate data association, which greatly affects the positioning accuracy and map creation of the system. The miscorrelation of the data may even cause the system to crash. Removing dynamic feature points in dynamic environments can reduce the number of feature points, which can lead to object loss [10]. In scenarios with missing texture or motion blur, line features show greater robustness and can represent the structural features of the scene, providing intuitive visual information [11]. Therefore, it is crucial to study the extraction of line features for point–line fusion in dynamic environments. Such research could help to address challenges such as missing map point tracking, degraded localization accuracy, and map drift in dynamic scenes.
To address the problems of dynamic environments, which may lead to system tracking failures and degraded localization accuracy, we propose a SLAM system called YPL-SLAM. It is designed to address system failures and degraded localization accuracy due to insufficient feature point extraction or image blurring in dynamic environments. The YPL-SLAM algorithm combines the YOLOv5s algorithm with the point–line feature fusion technique. This greatly improves the localization accuracy of the system in dynamic scenes, enabling it to extract line features in non-dynamic regions while removing dynamic feature points to ensure that sufficient features are extracted to increase the stability of the system. We improve the ORB-SLAM2 algorithm by detecting objects in the environment through target detection and region segmentation threads, identifying dynamic, potentially dynamic, and static regions in the image, and determining the state of the potentially dynamic situation. Meanwhile, we remove dynamic feature points in the dynamic region and then extract line features in the non-dynamic region. Line features can more comprehensively represent structural features [12]. The main contributions of our work are summarized as follows:
  • We design an improved algorithm, YPL-SLAM, based on ORB-SLAM2, which combines YOLOv5s with point–line feature fusion technology, increases the target detection and region delineation threads, and adopts multi-thread parallelism to ensure the real-time performance and accuracy of the algorithm in a dynamic environment.
  • We construct a method to determine the potential dynamic region. The dynamic region, potentially dynamic regions, and static regions are obtained after target detection and area division. Through processing, the states of the potential dynamic region are determined and the dynamic feature points of the dynamic region are eliminated.
  • We use the weighted fusion strategy considering the image dynamic score and the number of successful feature point and line matches for point–line fusion optimization and processing. The line feature points in the static region are extracted and the point and line feature weights are processed according to the image dynamic score and the number of successful feature point–line matches to obtain the optimized point–line fusion results.

2. Related Works

Currently, the visual SLAM algorithms used for visual odometry are divided into three main categories: the feature point method, the optical flow method, and the direct method. The feature point method extracts feature points between consecutive frames and matches them; finally, it calculates the camera motion through the matching relationship. The optical flow method extracts feature points without matching and solves the camera motion based on the assumption of grayscale invariance. The direct method calculates the camera motion by considering the pixel positions of the spatial points in two consecutive frames based on the optical flow method, as well as building an error model.
The most commonly used algorithms for the extraction of feature points are SIFT [13], SURF [14], and ORB [15]. While these methods boast high accuracy, they are susceptible to instability or even failure when deployed in complex environments. Unlike these feature point methods, the direct methods utilize pixel-level grayscale image information to estimate the motion and scene structure by minimizing photometric errors. These methods are usually more robust in weakly textured scenes. Examples include LSD-SLAM [16], SVO [17], and DSO [18]. However, the direct methods include the assumption of grayscale invariance, and the performance of these methods suffers in scenes with significant changes in illumination. As a direct method, SLAM relies on an image gradient search to optimize the pixel gray values. However, due to the highly non-convex nature of images, particularly when significant camera motion is involved, the algorithm becomes susceptible to local minima. To address these problems, line feature-based SLAM methods have been proposed. Rong et al. [19] introduced PEL-SLAM, a point–line SLAM method leveraging ED-line [20] to enhance environmental adaptation and accuracy. However, this approach simultaneously escalates the computational complexity and heightens the sensitivity to dynamic scenes and illumination variations. Gomez-Ojeda et al. [21] introduced PL-SLAM, a visual SLAM system that integrates both point and line features. This fusion of features enables PL-SLAM to exhibit improved adaptability across diverse environments, particularly in scenarios with limited features or structured scenes. Zhang et al. [22] introduced PL-GM, a SLAM algorithm that merges 3D points and lines generated by RGB-D cameras with 2D point and line features. PL-GM achieves comparable or better performance compared to SLAM methods based on point–line features and point features. However, these methods may exhibit instability or subpar performance when dealing with moving objects or when there are dynamic changes in the scene. Given the important advances in deep learning for target detection, SLAM, combined with deep learning techniques, could be more effective in dynamic environments.
In dynamic environments, the primary objective of the visual SLAM system is to effectively distinguish dynamic targets from outliers. This involves integrating SLAM techniques with deep learning to leverage the capabilities of deep learning models for accurately filtering out dynamic information. This section outlines the related research in this area. On the one hand, SLAM is combined with semantic segmentation techniques. For example, the Dyna-SLAM method, proposed by Bescos et al. [23], combines the Mask R-CNN [24] algorithm with ORB-SLAM2 to effectively detect and segment dynamic objects. However, the method relies on target detection and semantic segmentation for localization tracking, which poses challenges, especially in real-time applications. Wang et al. [25] combined the visual inertial SLAM technique with deep learning to determine the motion probabilities of feature points by using semantic information extracted from images. This improves localization accuracy by eliminating feature points with an overly high motion probability. However, this motion probability grading model is constructed based on life experience and may be limited in multi-target motion scenarios. Yu et al. [26] introduced DS-SLAM, a semantic system that mitigates the influence of dynamic objects through a combination of semantic segmentation and optical flow techniques. However, a decrease in accuracy occurs when there is dense sampling in large-format images. On the other hand, SLAM is combined with target detection techniques. For example, Wu et al. [27] proposed YOLO-SLAM, which combines target detection methods and geometric constraint methods in a tightly coupled manner, aiming to reduce the impact of dynamic targets by adding a lightweight YOLOv3 network into the SLAM system. However, the lightweight YOLOv3 network has limitations regarding dynamic target recognition in complex scenes. Zhang et al. [28] employed a lightweight YOLOv5s network to enhance the system’s operation speed. Additionally, they integrated a pyramid-shaped scene parsing network split header at the YOLOv5s network’s head to achieve semantic extraction within the environment. However, the fast detection speed of the lightweight YOLO target detection algorithm may lead to less accurate results for small targets or heavily occluded targets. Song et al. [29] introduced YF-SLAM, a method leveraging the YOLO-Fastest network to rapidly identify dynamic target regions within dynamic scenes. Subsequently, they employed depth geometric constraints to filter out dynamic feature points, thereby ensuring the real-time operation of the SLAM system. However, when the camera moves in a blurred scene, the depth information is affected, which impacts the localization and map construction quality of the SLAM system. Gong et al. [30] proposed an algorithm for AHY-SLAM using the optical flow method with filtering of the input keyframes in each frame. It extracts the feature points of the keyframes using adaptive thresholding, eliminating the dynamic points using YOLOv5. However, when there are multiple moving objects in the scene, a large number of feature points are eliminated, and its robustness can be limited. Thus, these techniques exhibit poor performance or failure in scenarios involving multiple moving targets or motion blur in the camera. To address these issues, we propose YPL-SLAM, a point–line feature-based SLAM system for dynamic environments.
YPL-SLAM integrates the YOLOv5s object detection technology with point–line feature fusion, removes feature points from dynamic regions, extracts line features from static regions, and employs a weighted fusion strategy based on the image dynamic score and the number of successful matches of feature points and lines. This fusion aims to more efficiently detect and track moving objects in dynamic scenes, thereby reducing the negative impact of multiple dynamic objects on localization. Through this innovative approach, we expect to significantly improve the robustness and accuracy of SLAM systems when dealing with complex scenes exhibiting dynamic characteristics.

3. Methods

In this section, we introduce YPL-SLAM from four perspectives. First, we outline the general framework of YPL-SLAM. The main element of the general framework is the improvement of ORB-SLAM2. We then introduce target detection and region segmentation using the YOLOv5s algorithm, which improves the real-time performance of the system by identifying the region in which the target is located and dividing it into dynamic, potential, and static regions, as compared to commonly used algorithms that incorporate semantic segmentation. Then, the method for determining the state of the potential dynamic region is presented. Finally, we use the LSD [31] algorithm with the LBD [32] descriptor to extract the line features and matches within the static region and optimize the processing of point–line fusion using a weighted fusion strategy that considers the image dynamic score and the number of successful feature point–line matches.

3.1. Overview of the YPL-SLAM System

The ORB-SLAM2 algorithm consists of three threads: tracking, localization building, and loop closing. It utilizes ORB feature extraction for localization building, which combines FAST [33] key point detection and BRIEF [34] descriptors, offering rotational and scale invariance. While ORB-SLAM2 exhibits strong performance in static environments with localization accuracy, we chose to enhance it further. We add target detection and region delineation modules. At the same time, we add a module for the processing of the dynamic and potential dynamic regions in the tracking thread, determine the state of the potential dynamic region, remove dynamic feature points, extract non-dynamic region line features through these processes, and optimize the camera position via point–line fusion, as shown in Figure 1. The processing of the dynamic and potentially dynamic regions contributes to the better handling of moving objects during vision SLAM, and the point–line fusion approach used to optimize the camera position is expected to improve the robustness and localization accuracy of the system. Specific methods for these improvements will be detailed later.

3.2. Target Recognition and Region Segmentation

The YOLOv5s algorithm operates by taking an RGB-D image as input, extracting features through a convolutional neural network, and subsequently parsing these features to determine the target object’s category and bounding box coordinates. Unlike the two-stage detection algorithms of the RCNN series, YOLOv5s streamlines the process by eliminating the step of predicting the positions of candidate boxes. Instead, it treats the problem holistically as a regression task, utilizing convolutional networks for direct end-to-end model training. This approach significantly enhances the speed, surpassing that of other algorithms and ensuring the real-time operation of the entire SLAM system. By leveraging convolutional networks and direct end-to-end training, YOLOv5s accelerates target detection while also providing precise bounding box coordinate information. Moreover, it utilizes regression targets based on pixel information from the entire image, thereby minimizing false alarms in the background and enhancing the separation of the detected targets from background areas.
The YOLOv5s algorithm is used to detect the target in the image and obtain its bounding box location information, which is subsequently classified into three categories according to the region in which the target is located: static regions, potentially dynamic regions, and dynamic regions. The regions where the different targets are located are classified according to the classification criteria in Table 1. Targets with subjective mobility, such as humans and animals, are classified as dynamic regions, while objects that can easily be moved, such as mice, keyboards, and chairs, are classified as potentially dynamic regions, and objects that do not move easily, such as computer screens, cabinets, and windows, are categorized as static regions. Figure 2 illustrates the information about the coordinates of the regions obtained from the classification. Next, the segmented area’s coordinate information is sent to the tracking thread for subsequent target tracking and analysis. This process helps to identify and distinguish the different targets in the image more accurately and provides the basis for further dynamic region analysis. By categorizing the dynamic targets and potential dynamic regions, the system can respond more effectively to different target types, providing targeted information for subsequent processing. This integrated approach to target recognition and region segmentation is expected to improve the system’s ability to perceive and understand dynamic environments.

3.3. Determination of the State of a Potential Dynamic Region

According to the target detection and region delineation module, we can obtain the position information of the potential dynamic regions, dynamic regions, and static regions. The camera position is calculated using the extracted feature points, with the exception of the dynamic region, and RANSAC processing is performed to remove abnormal feature points. Then, the dynamic and potential dynamic regions are processed to obtain the state of the potential dynamic region, and the static feature points in the dynamic region are retained. The status of the region containing the object is determined through target detection. The base matrix F is computed for consecutive frames using ORB feature points extracted from areas outside the dynamic region. Geometric constraints are then applied to calculate the feature points within the potential dynamic regions, allowing for an assessment of the region’s status. The static feature points within the dynamic regions are retained. Figure 3 shows the geometric constraints on the poles.
We can obtain the Epipolar Constraint Formula as follows:
E = t ^ R F = K T E K 1 x 2 T E x 1 = p 2 T F p 1 = 0
In the equation, K represents the camera’s internal reference matrix, and R and t signify the camera’s motion from the previous frame coordinate system to the next frame coordinate system. p 1 and p 2 are characteristic points. x 1 and x 2 are normalized plane coordinates. F represents the fundamental matrix and E represents the essential matrix. We can obtain the following equation:
l 2 = a b c = F p 1
We calculate the distance from the pixel point coordinate p 2 to the polar line l 2 .
d = p 2 T l 2 a 2 + b 2
where d denotes the distance from the point p 2 to the line l 2 , where a , b , and c express the direction vectors of the line l 2 .
Due to the effect of errors, we set a threshold for the distance from p 2 to the straight line l 2 and adjust the threshold according to the specific application scenario and needs. When the calculated d exceeds this threshold, the point is considered an anomaly; otherwise, the point is retained for subsequent calculations.
The polar geometric constraints allow us to determine the anomalous feature points, and the boundary information allows us to determine the number of feature points extracted from different areas. We remove the dynamic feature points for dynamic regions, while, for potentially dynamic regions, the state is then determined according to the following equation:
T = 1 1 + e ( N m + 1 ) / N a
where N m denotes the number of abnormal points in the center of the potential dynamic region, and N a denotes the number of all feature points in the potential dynamic region. When the calculation result exceeds the set threshold ε , it is considered that the region is dynamic; otherwise, the target is considered stationary.

3.4. Extraction of Non-Dynamic Region Line Features

We have extracted the ORB feature points in the image and divided the image into dynamic and potential dynamic regions, which will result in some feature points being filtered out. We improve the stability of the algorithm by extracting the line features in the static region again. In this study, we use the LSD algorithm based on the gradient direction and gradient magnitude of the pixel points to extract the line segments. Then, we use the baseline descriptor to represent the line segment feature information, calculate the LBD vector value of each line feature for the adjacent two frames of the image, and use the nearest neighbor method to solve the line feature matching pairs. When the LBD ratio value of the line feature matching pair is greater than a set threshold value, the line feature matching pair will be retained; otherwise, it is rejected as a mismatch. Finally, the weighted fusion strategy based on the ratio of the image’s dynamic region and the number of successful feature point–line matches assigns respective weights to the point–line errors in the cost function to solve the problem of accuracy degradation due to the direct fusion of different types of data errors. Figure 4 shows the results of our algorithm for different processes performed on the same image frame, visually comparing the different effects of the three steps of extracting the feature points in the image, removing the dynamic feature points, and extracting the line features in the non-dynamic region.
We can see that extracting the feature points in the image leads to the extraction of many dynamic feature points, which seriously affects the accuracy of the algorithm. After performing the removal of dynamic points, if the view has a high percentage of dynamic objects, this will filter out some of the feature points, leading to a reduction in the number of extracted points, which reduces the accuracy of the algorithm or causes a tracking failure. By extracting the line features in the non-dynamic region of the image, the problem of missing feature points can be eliminated to better ensure the stability of the algorithm.

3.5. Point and Line Fusion

A line segment has four degrees of freedom in the three-dimensional coordinate system, so the combination of the Plücker coordinates and orthogonal representation is used to represent the spatial line segment. We define P 1 and P 2 as two points on a straight line in the space, and the spatial straight line Plücker coordinates can be expressed as
L w = P 1 × P 2 w 1 P 1 w 2 P 2 = n v
where w 1 and w 2 are chi-square factors, ν is the direction vector of the line, and n is the normal vector of the plane formed by the coordinate origin and P 1 P 2 .
The straight line is converted from the world coordinate system to the camera coordinate system and then represented using the Plücker coordinate system.
L c = T c w L w = R c w ( t c w ) ^ R c w 0 R   c w n w v w = n c v c
where L c represents a straight line in the camera coordinate system in Plücker coordinates, L w represents a straight line in the world coordinate system in Plücker coordinates, T c w is the transformation matrix from the world coordinate system to the camera coordinate system, R c w is the rotation matrix, t c w is the translation matrix, and t c w ^ represents the antisymmetric matrix of t c w .
By projecting L c onto the image plane coordinate system, we obtain
l c = f x 0 0 0 f y 0 f x c x f x c y f x f y n c = K n c
where l c represents the coordinates of the line in the image plane coordinate system. f x , f y , c x , and c y are the camera’s internal reference, and K is the camera’s internal reference matrix.
The camera position T c w k , 3D spatial point X w , i k , and line L w , i k are taken as vertices, and their relationships are taken as edges to construct the graph optimization model. It can be determined that the error in the camera position–3D point and the camera position–spatial line segment is
e p , i k = x m , i k I K T c w k X w , i k e l , j k = d l m , j k , K T c w k L w , i k
where e p , i k represents the projection error of the feature point, x m , i k represents the corresponding pixel coordinates of the feature point, and I is the projection function, which projects the 3D point onto the 2D image plane. e l , j k is the projection error of the line feature, and d is used to compute the line segment error between the measured value and the computed value, which is based on the distance of the endpoints of the line segment as a measure of the error.
Setting the observation error as a Gaussian distribution, we can obtain the following cost function:
C = λ p k , i ρ p e p , i k   T x k , i 1 e p , i k + λ l k , j ρ l e p , j k   T l k , j 1 e l , j k
where λ p and λ l are the observation error weighting coefficients of the point and line features, respectively; x k , i and x l , i denote the observation covariances of the point and line features; and ρ p , ρ l are the robust kernel functions used to suppress the outliers, which are computed by the following formula:
ρ m = 1 2 e m , i k   2                 e m , i k δ δ e m , i k 1 2 δ o t h e r , m { p , l }
The weight coefficients λ p and λ l are set according to Equations (11) and (12). We synthesize and score each image to obtain M p k . The integrated score M p k of the image consists of two parts: the ratio of the number of successful feature point–line matches r r k and the image dynamic score s r k .
r r k = N   l k / N p k M p k = β r r k + s r k s r k = μ S d k / S 0 β = 1 e N p k + 1
where S d k is the area of the dynamic region of the image in the kth frame, S 0 is the total area of the image, and μ and β are the adaptive scale factors, where β is positively correlated with N p k . The current kth keyframe image successfully matches the number of line features N l k and the number of point features N p k . r r k is the ratio of line features to point features that are successfully matched.
λ p = M p k ¯ r r k ¯ r 1.25 r r k ¯ > r , λ l = 1 λ p
where M p k ¯ is the average of the combined scores of the images, and r r k ¯ is the average of the ratios of line features and point features extracted from the images. r is a threshold value that is used to control the growth of the coefficients in highly dynamic scenarios.

4. Results

In this section, we describe the use of publicly available datasets to validate our algorithm’s effectiveness. Additionally, we compare our SLAM system with state-of-the-art systems from other countries, assessing both their accuracy and real-time performance.

4.1. Dataset and Algorithms for Comparison

YPL-SLAM is a SLAM system designed for dynamic scenes, where RGB-D images are passed into the system and used to achieve simultaneous localization and map construction. Therefore, we used the publicly available TUM dynamic object dataset for the system’s evaluation. This dataset was captured by a Kinect camera using the TOF technique and contains multiple color and depth images at a 640 × 480 resolution. The dataset is divided into two classes based on the scene dynamics: class ‘w’ represents walking scenes, which show high dynamic activity, and class ‘s’ represents sedentary scenes, which show lower dynamic activity. The scenes include two people sitting at a table talking and making gestures. The data were further categorized into four modes of camera motion: (1) the static mode, in which the sensor was manually fixed in one position; (2) the xyz mode, in which the sensor was moved along the x, y, and z axes, with the orientation held constant; (3) the hemispherical mode, in which the sensor was moved in a hemispherical space with a diameter of approximately one meter; and (4) the rpy mode, in which the sensor was in the same position around a major axis’ (roll, pitch, yaw) rotation.
The YPL-SLAM algorithm uses these datasets for comparison with other algorithms. When comparing the algorithms, we consider that YPL-SLAM is derived from the improvements in ORB-SLAM2, which excels in static environments. Hence, we select ORB-SLAM2 for comparison. In dynamic scenes, where Dyna-SLAM and DS-SLAM are recognized as advanced open-source SLAM systems, we include them in the comparison to demonstrate the performance of our proposed algorithm.

4.2. Evaluation Method

We utilize the open-source assessment tool evo to compare the performance of the algorithms. Evo offers a range of utilities for the assessment of SLAM system scripts and measurements, enabling the comparison of different algorithms. It supports evaluations using real-world datasets and provides visualization tools for the analysis of SLAM systems’ outputs. The absolute trajectory error (ATE) serves as a metric to evaluate SLAM systems’ performance. It quantifies the difference between the trajectories output by the SLAM algorithm and the true or reference trajectories, aiding in the assessment of the system’s robustness and accuracy across varying environments and over extended periods. Additionally, we compare the relative pose errors of the different algorithms, which are commonly used to evaluate the drift in displacement estimation.
The following equation is primarily used in calculating the root mean square error of the absolute trajectory error (ATE) between the estimated pose E and the ground truth pose G .
A T E - R M S E = 1 n i = 0 n | | t r a n s ( E ) t r a n s ( G ) | | 2 2
In this equation, n represents the total number of poses to be computed, t r a n s G represents the true pose of the camera, t r a n s E represents the estimated pose of the camera, and | | · | | denotes the Euclidean distance between these two poses.
The formula for RPE-RMSE is as follows:
R P E - R M S E = 1 m i = 1 m | | t r a n s ( Q i 1 Q i + Δ ) 1 ( P i 1 P i + Δ ) | | 2 2 , m = n Δ
where Q i denotes the ground truth pose, P i denotes the estimated pose, and Δ denotes the time interval between two consecutive poses.
We evaluate the performance of our algorithm, YPL-SLAM, compared to ORB-SLAM2, using the following equation:
η = σ τ σ × 100 %
The above equation represents the performance comparison between our algorithm and ORB-SLAM2, where σ denotes the error generated by ORB-SLAM2 and τ denotes the error generated by the YPL-SLAM algorithm.

4.3. Tracking Accuracy Evaluation

Figure 5 illustrates the absolute trajectory error (ATE) and relative pose error (RPE) values of ORB_SLAM2 and YPL-SLAM, computed using the evo tool in low dynamic scenes. Notably, YPL-SLAM demonstrates a lower ATE and RPE, indicating higher accuracy compared to ORB_SLAM2 in such scenarios. Furthermore, Figure 6 provides a comparison of the trajectories estimated by YPL-SLAM with the real trajectories. It shows that the trajectories estimated by YPL-SLAM closely align with the ground truth trajectory. An examination of the X, Y, and Z axes, as well as the pitch, yaw, and roll views, reveals that the estimated results of YPL-SLAM closely match the ground truth data.
Figure 7 and Figure 8 depict images captured in highly dynamic scenes, showcasing the stability and effectiveness of YPL-SLAM in such environments. These images offer a detailed visual representation of YPL-SLAM’s performance in the context of highly dynamic environments. Overall, these results demonstrate that YPL-SLAM achieves satisfactory localization performance in both less and highly dynamic scenes.
Table 2, Table 3 and Table 4 present the quantitative results of the comparison of YPL-SLAM with other algorithms in terms of ATE and RPE. The root mean square error (RMSE) and standard deviation (SD) are reported in the tables to better illustrate the localization accuracy and system stability. The RMSE and SD provide a more intuitive understanding of the accuracy of the results. Table 2 illustrates that the ATE calculated for YPL-SLAM improves across all sequences. Compared to ORB-SLAM2, YPL-SLAM shows notably enhanced performance in highly dynamic scenes, achieving a significant improvement in accuracy by an order of magnitude. Particularly in the fr3_w_xyz sequence, which represents a highly dynamic scenario, the average error and root mean square error show the most substantial improvement, reaching 96.1% and 95.76%, respectively. This indicates that YPL-SLAM effectively enhances the system’s performance in highly dynamic environments, particularly in terms of accuracy and stability. In less dynamic environments, YPL-SLAM also exhibits significant improvements over ORB-SLAM2, yielding very favorable results. However, when compared to Dyna-SLAM and DS-SLAM, the difference in the performance of YPL-SLAM is not significant.
Table 3 and Table 4 present the results of the relative pose error (RPE) comparison between YPL-SLAM and the other algorithms. Table 3 displays the results for the translational part of the RPE, while Table 4 shows the results for the rotational part of the RPE. Compared with ORB-SLAM2, YPL-SLAM shows an improvement in the RMSE and SD in the translational part of the RPE of 74.75% to 88.63% and 77.45% to 91.56%, respectively, and in the rotational part of the RPE by 70.66% to 86.26% and 74.97% to 89.38, respectively, in highly dynamic sequences. The experimental results for YPL-SLAM are superior in some sequences compared to those of Dyna-SLAM and DS-SLAM. Through numerous tests, we have observed that YPL-SLAM achieves stability and reliability in highly dynamic sequences.

4.4. Real-Time Evaluation

To assess the real-time performance of YPL-SLAM, a series of comparative experiments involving ORB-SLAM2, Dyna-SLAM, and YPL-SLAM were conducted, processing each image frame fully. YPL-SLAM and Dyna-SLAM were tested on identical computers equipped with an Intel i7-12700 CPU and an NVIDIA RTX 2060 graphics card to ensure a fair comparison. To evaluate the real-time performance on the same hardware, we compared the processing speeds of Dyna-SLAM and YPL-SLAM. YPL-SLAM utilizes YOLOv5s for target detection and employs parallel processing, making its processing time largely independent of the target detection time. However, in the tracking thread, it is necessary to utilize the results of target detection to extract line features in static regions. Therefore, we add the processing time for target detection to the processing time of each frame. Table 5 presents the experimental results. Despite increasing the overhead for target detection, YPL-SLAM maintains good real-time performance compared to ORB-SLAM2. Furthermore, compared to Dyna-SLAM, YPL-SLAM demonstrates stronger real-time performance. This indicates that YPL-SLAM can operate more efficiently when combining target detection and line feature extraction, providing an advantage in quickly recognizing and removing dynamic objects in real-time applications.

5. Discussion

In this study, we proposed a SLAM method to cope with point–line feature functions in dynamic environments: YPL-SLAM. Compared with ORB-SLAM2, YPL-SLAM adds target identification and region delineation threads, as well as a line feature extraction module and a point–line fusion optimization module. Through its comparison with other SLAM algorithms in the dynamic sequences of the TUM dataset, we obtained the following insights:
  • Compared with the ORB-SLAM2 method, the performance of YPL-SLAM in a highly dynamic environment is more obvious, and the localization accuracy is improved by more than 80%.
  • In comparison with Dyna-SLAM and DS-SLAM, YPL-SLAM still has some shortcomings in highly dynamic scenarios, although the performance difference is not significant in most cases. These shortcomings are mainly due to the fact that Dyna-SLAM utilizes the a priori dynamic region information of each frame for pixel-level semantic segmentation, which results in superior dynamic point filtering capabilities.
  • Comparing the RPE results, it can be seen that the computation results obtained via point–line fusion have fewer errors. The point–line fusion approach plays an optimizing role in the calculation of the relative pose and reduces the error level of the system. It has a significant effect in terms of improving the accuracy and reliability of the attitude estimation process.
  • Based on its ability to guarantee high accuracy, the speed of YPL-SLAM ensures its real-time operation. Moreover, by utilizing YOLOv5s for target recognition and the division of the region where the target is located, and by extracting the line features of the static region, the robustness and accuracy of the system are guaranteed.

6. Conclusions

Addressing the situation in which removing dynamic feature points in dynamic environments with multiple moving objects may lead to a reduction in the number of feature points, which in turn leads to the loss of the object, a new SLAM method, YPL-SLAM, is proposed. It is based on ORB-SLAM2 and utilizes YOLOv5s to perform target recognition. This method divides the target’s area to determine the dynamic region, and the dynamic region’s feature points are removed; then, it extracts the line features of the non-dynamic region and uses a weighted fusion strategy considering the image dynamic scores and the number of successful feature point–line matches for point–line fusion optimization and processing, takes advantage of the high robustness of the line features in the texture-deficient or motion-blurring scenarios, ensuring the stability and accuracy of the system. Compared with the ORB-SLAM2 method, the performance improvement provided by YPL-SLAM in highly dynamic environments is more obvious, and the absolute trajectory error is reduced by up to 96.1% in the highly dynamic sequences of the TUM dataset. In the optimization process of point–line fusion, a weighted fusion strategy that combines the dynamic score of the image and the number of successful feature point–line matches is used for the point–line fusion optimization process. This solves the problems of reducing the number of feature points, resolving the issue of lost objects to a certain extent, and improving accuracy. Meanwhile, the speed of YPL-SLAM can ensure real-time operation.
Although YPL-SLAM has allowed significant progress in processing dynamic scenes, we believe that there is potential for further improvement. In our study, the scene tests in the TUM dataset were conducted under good lighting conditions. Given the critical role of illumination in vision SLAM, we plan to integrate vision and LiDAR in our subsequent work to fully utilize the advantages of LiDAR in poor lighting conditions. The point cloud data provided by LiDAR are insensitive to light, which will help to compensate for the shortcomings of vision SLAM in light-complex scenes. By combining the information from vision and LiDAR, we expect to improve the stability of the SLAM algorithms under various environmental conditions, which will bring greater performance improvements for real-time localization and map building. At the same time, we will consider introducing strategies that enable the system to adapt to different environmental conditions to ensure good performance in various scenarios. Mechanisms for the automatic adjustment of the algorithm’s parameters could be developed so that real-time adjustments can be made at runtime in response to changes in the environment. Such a strategy will enable the system to adapt to changes in illumination, scene complexity, and dynamics, thus improving its robustness. This integrated approach is expected to further advance SLAM’s performance in dynamic scenes and under complex lighting conditions.

Author Contributions

Conceptualization, C.Z.; methodology, C.Z. and K.G.; software, C.Z. and J.L.; validation, X.D., K.G. and X.Y.; formal analysis, C.Z. and S.W.; investigation, S.W. and J.L.; resources, X.D.; data curation, X.Y. and K.G.; writing—original draft preparation, C.Z.; writing review and editing, J.L. and K.G. All authors have read and agreed to the published version of the manuscript.

Funding

Longmen laboratory project (Grant No. LMFKCY2023001 and No. LMFKCY2023002), National Nature Science Foundation of China (Grant No. 52075150).

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The research data used in this paper are from https://vision.in.tum.de/data/datasets/rgbd-dataset/download (accessed on 11 April 2023).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Favorskaya, M.N. Deep Learning for Visual SLAM: The State-of-the-Art and Future Trends. Electronics 2023, 12, 2006. [Google Scholar] [CrossRef]
  2. Jia, Y.; Yan, X.; Xu, Y. A Survey of simultaneous localization and mapping for robot. In Proceedings of the 2019 IEEE 4th Advanced Information Technology, Electronic and Automation Control Conference (IAEAC), Chengdu, China, 20–22 December 2019; pp. 857–861. [Google Scholar]
  3. Al-Tawil, B.; Hempel, T.; Abdelrahman, A.; Al-Hamadi, A. A review of visual SLAM for robotics: Evolution, properties, and future applications. Front. Robot. AI 2024, 11, 1347985. [Google Scholar] [CrossRef] [PubMed]
  4. Cadena, C.; Carlone, L.; Carrillo, H.; Latif, Y.; Scaramuzza, D.; Neira, J.; Reid, I.; Leonard, J.J. Past, Present, and Future of Simultaneous Localization and Mapping: Toward the Robust-Perception Age. IEEE Trans. Robot. 2016, 32, 1309–1332. [Google Scholar] [CrossRef]
  5. Klein, G.; Murray, D. Parallel Tracking and Mapping for Small AR Workspaces. In Proceedings of the 2007 6th IEEE and ACM International Symposium on Mixed and Augmented Reality, Nara, Japan, 13–16 November 2007; pp. 225–234. [Google Scholar]
  6. Mur-Artal, R.; Tardós, J. ORB-SLAM2: An Open-Source SLAM System for Monocular, Stereo, and RGB-D Cameras. IEEE Trans. Robot. 2017, 33, 1255–1262. [Google Scholar] [CrossRef]
  7. Campos, C.; Elvira, R.; Rodríguez, J.M. Montiel, J.D.; Tardós, J. ORB-SLAM3: An Accurate Open-Source Library for Visual, Visual–Inertial, and Multimap SLAM. IEEE Trans. Robot. 2021, 37, 1874–1890. [Google Scholar] [CrossRef]
  8. Qin, T.; Li, P.; Shen, S. VINS-Mono: A Robust and Versatile Monocular Visual-Inertial State Estimator. IEEE Trans. Robot. 2018, 34, 1004–1020. [Google Scholar] [CrossRef]
  9. Qin, T.; Pan, J.; Cao, S.; Shen, S. A General Optimization-based Framework for Local Odometry Estimation with Multiple Sensors. arXiv 2019, arXiv:1901.03638. [Google Scholar]
  10. Yuan, C.; Xu, Y.; Zhou, Q. PLDS-SLAM: Point and Line Features SLAM in Dynamic Environment. Remote Sens. 2023, 15, 1893. [Google Scholar] [CrossRef]
  11. Zhao, Z.; Song, T.; Xing, B.; Lei, Y.; Wang, Z. PLI-VINS: Visual-Inertial SLAM Based on Point-Line Feature Fusion in Indoor Environment. Sensors 2022, 22, 5457. [Google Scholar] [CrossRef]
  12. Li, D.; Liu, S.; Xiang, W.; Tan, Q.; Yuan, K.; Zhang, Z.; Hu, Y. A SLAM System Based on RGBD Image and Point-Line Feature. IEEE Access 2021, 9, 9012–9025. [Google Scholar] [CrossRef]
  13. Lowe, D.G. Distinctive Image Features from Scale-Invariant Keypoints. Int. J. Comput. Vis. 2004, 60, 91–110. [Google Scholar] [CrossRef]
  14. Bay, H.; Ess, A.; Tuytelaars, T.; Van Gool, L. Speeded-Up Robust Features (SURF). Comput. Vis. Image Underst. 2008, 110, 346–359. [Google Scholar] [CrossRef]
  15. Rublee, E.; Rabaud, V.; Konolige, K.; Bradski, G. ORB: An efficient alternative to SIFT or SURF. In Proceedings of the 2011 International Conference on Computer Vision, Barcelona, Spain, 6–13 November 2011; pp. 2564–2571. [Google Scholar]
  16. Engel, J.; Schöps, T.; Cremers, D. LSD-SLAM: Large-Scale Direct Monocular SLAM. In Computer Vision–ECCV 2014, 13th European Conference, Zurich, Switzerland, 6–12 September 2014; Springer: Berlin/Heidelberg, Germany, 2014; pp. 834–849. [Google Scholar]
  17. Forster, C.; Zhang, Z.; Gassner, M.; Werlberger, M.; Scaramuzza, D. SVO: Semidirect Visual Odometry for Monocular and Multicamera Systems. IEEE Trans. Robot. 2017, 33, 249–265. [Google Scholar] [CrossRef]
  18. Engel, J.; Koltun, V.; Cremers, D. Direct Sparse Odometry. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 40, 611–625. [Google Scholar] [CrossRef]
  19. Rong, H.; Gao, Y.; Guan, L.; Alex, R.; Xu, X.; Zhu, Y. Point-Line Visual Stereo SLAM Using EDlines and PL-BoW. Remote Sens. 2021, 13, 3591. [Google Scholar] [CrossRef]
  20. Cuneyt, A.; Cihan, T. EDLines: A real-time line segment detector with a false detection control. Pattern Recognit. Lett. 2011, 32, 1633–1642. [Google Scholar]
  21. Gomez-Ojeda, R.; Moreno, F.-A.; Zuñiga-Noël, D.; Scaramuzza, D.; Gonzalez-Jimenez, J. PL-SLAM: A Stereo SLAM System Through the Combination of Points and Line Segments. IEEE Trans. Robot. 2019, 35, 734–746. [Google Scholar] [CrossRef]
  22. Zhang, C. PL-GM:RGB-D SLAM With a Novel 2D and 3D Geometric Constraint Model of Point and Line Features. IEEE Access 2021, 9, 9958–9971. [Google Scholar] [CrossRef]
  23. Bescos, B.; Fácil, J.M.; Civera, J.; Neira, J. DynaSLAM: Tracking, Mapping, and Inpainting in Dynamic Scenes. IEEE Robot. Autom. Lett. 2018, 3, 4076–4083. [Google Scholar] [CrossRef]
  24. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask R-CNN. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar]
  25. Wang, Y.; Liu, X.; Zhao, M.; Xu, X. VIS-SLAM: A Real-Time Dynamic SLAM Algorithm Based on the Fusion of Visual, Inertial, and Semantic Information. ISPRS Int. J. Geo-Inf. 2024, 13, 163. [Google Scholar] [CrossRef]
  26. Yu, C.; Liu, Z.; Liu, X.; Xie, F.; Yang, Y.; Wei, Q.; Fei, Q. DS-SLAM: A Semantic Visual SLAM towards Dynamic Environments. In Proceedings of the 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Madrid, Spain, 1–5 October 2018; pp. 1168–1174. [Google Scholar]
  27. Wu, W.; Guo, L.; Gao, H.; You, Z.; Liu, Y.; Chen, Z. YOLO-SLAM: A Semantic SLAM System towards Dynamic Environment with Geometric Constraint. Neural Comput. Appl. 2022, 34, 6011–6026. [Google Scholar] [CrossRef]
  28. Zhang, R.; Zhang, X. Geometric Constraint-Based and Improved YOLOv5 Semantic SLAM for Dynamic Scenes. ISPRS Int. J. Geo-Inf. 2023, 12, 211. [Google Scholar] [CrossRef]
  29. Song, Z.; Su, W.; Chen, H.; Feng, M.; Peng, J.; Zhang, A. VSLAM Optimization Method in Dynamic Scenes Based on YOLO-Fastest. Electronics 2023, 12, 3538. [Google Scholar] [CrossRef]
  30. Gong, H.; Gong, L.; Ma, T.; Sun, Z.; Li, L. AHY-SLAM: Toward Faster and More Accurate Visual SLAM in Dynamic Scenes Using Homogenized Feature Extraction and Object Detection Method. Sensors 2023, 23, 4241. [Google Scholar] [CrossRef] [PubMed]
  31. Grompone von Gioi, R.; Jakubowicz, J.; Morel, J.-M.; Randall, G. LSD: A Fast Line Segment Detector with a False Detection Control. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 32, 722–732. [Google Scholar] [CrossRef] [PubMed]
  32. Zhang, L.; Koch, R. An efficient and robust line segment matching approach based on LBD descriptor and pairwise geometric consistency. J. Vis. Commun. Image Represent. 2013, 24, 794–805. [Google Scholar] [CrossRef]
  33. Rosten, E.; Drummond, T. Machine Learning for High-Speed Corner Detection. In Computer Vision—ECCV 2006. ECCV 2006, Proceedings of the 9th European Conference on Computer Vision, Graz, Austria, 7–13 May 2006; Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 2016; Volume 3951. [Google Scholar]
  34. Calonder, M.; Lepetit, V.; Strecha, C.; Fua, P. BRIEF: Binary Robust Independent Elementary Features. In Computer Vision—ECCV 2010. ECCV 2010, Proceedings of the 11th European Conference on Computer Vision, Heraklion, Crete, Greece, 5–11 September 2010; Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 2010; Volume 6314. [Google Scholar]
Figure 1. The SLAM framework of the YPL-SLAM algorithm, with the additional threads for target recognition and region delineation depicted in the blue box. Meanwhile, the yellow box illustrates the line feature extraction and processing module.
Figure 1. The SLAM framework of the YPL-SLAM algorithm, with the additional threads for target recognition and region delineation depicted in the blue box. Meanwhile, the yellow box illustrates the line feature extraction and processing module.
Sensors 24 04517 g001
Figure 2. The YOLOv5s algorithm recognizes the targets, obtains the targets’ semantic information, and classifies these targets into static, dynamic, and potentially dynamic regions. (a,b) Recognizing the targets; (c,d) categorizing the area in which the target is located. Dynamic indicates a dynamic region, uncertain indicates a potential dynamic region, and static indicates a static region.
Figure 2. The YOLOv5s algorithm recognizes the targets, obtains the targets’ semantic information, and classifies these targets into static, dynamic, and potentially dynamic regions. (a,b) Recognizing the targets; (c,d) categorizing the area in which the target is located. Dynamic indicates a dynamic region, uncertain indicates a potential dynamic region, and static indicates a static region.
Sensors 24 04517 g002
Figure 3. Geometric constraints on the poles.
Figure 3. Geometric constraints on the poles.
Sensors 24 04517 g003
Figure 4. The extraction of feature points, the filtering of dynamic feature points, and the extraction of the non-dynamic region’s line features are compared. (a,d,g) are extracted feature points, where the green dots represent the extracted feature points; (b,e,h) are the effects after removing the dynamic feature points; and (c,f,i) are the extracted non-dynamic region’s line features, where the red lines represent the extracted line features.
Figure 4. The extraction of feature points, the filtering of dynamic feature points, and the extraction of the non-dynamic region’s line features are compared. (a,d,g) are extracted feature points, where the green dots represent the extracted feature points; (b,e,h) are the effects after removing the dynamic feature points; and (c,f,i) are the extracted non-dynamic region’s line features, where the red lines represent the extracted line features.
Sensors 24 04517 g004aSensors 24 04517 g004b
Figure 5. In the fr3_s_hs sequence, we compare the absolute trajectory error (ATE) and relative pose error (RPE) between ORB-SLAM2 and YPL-SLAM, particularly focusing on the translation drift. The first and second columns of the table display the ATE results, while the third and fourth columns show the RPE results.
Figure 5. In the fr3_s_hs sequence, we compare the absolute trajectory error (ATE) and relative pose error (RPE) between ORB-SLAM2 and YPL-SLAM, particularly focusing on the translation drift. The first and second columns of the table display the ATE results, while the third and fourth columns show the RPE results.
Sensors 24 04517 g005
Figure 6. In the fr3_s_hs sequence, we compare the ground truth and estimated trajectories. The first column displays the comparison of the 3D trajectories, containing the ground truth and trajectories estimated by both ORB-SLAM2 and YPL-SLAM. The second column presents the fitting results for the X, Y, and Z axes, while the third column shows the fitting results for the roll, pitch, and yaw axes.
Figure 6. In the fr3_s_hs sequence, we compare the ground truth and estimated trajectories. The first column displays the comparison of the 3D trajectories, containing the ground truth and trajectories estimated by both ORB-SLAM2 and YPL-SLAM. The second column presents the fitting results for the X, Y, and Z axes, while the third column shows the fitting results for the roll, pitch, and yaw axes.
Sensors 24 04517 g006
Figure 7. In the fr3_w_xyz sequence, we compare the absolute trajectory error (ATE) and relative pose error (RPE) between ORB-SLAM2 and YPL-SLAM, focusing on the translation drift performance. The first two columns of the table display the ATE results, while the third and fourth columns show the RPE results.
Figure 7. In the fr3_w_xyz sequence, we compare the absolute trajectory error (ATE) and relative pose error (RPE) between ORB-SLAM2 and YPL-SLAM, focusing on the translation drift performance. The first two columns of the table display the ATE results, while the third and fourth columns show the RPE results.
Sensors 24 04517 g007
Figure 8. In the fr3_w_xyz sequence, we compare the ground truth and estimated trajectories. The first column displays the comparison of the 3D trajectories, containing the ground truth and trajectories estimated by both ORB-SLAM2 and YPL-SLAM. The second column presents the fitting results for the X, Y, and Z axes, while the third column shows the fitting results for the roll, pitch, and yaw axes.
Figure 8. In the fr3_w_xyz sequence, we compare the ground truth and estimated trajectories. The first column displays the comparison of the 3D trajectories, containing the ground truth and trajectories estimated by both ORB-SLAM2 and YPL-SLAM. The second column presents the fitting results for the X, Y, and Z axes, while the third column shows the fitting results for the roll, pitch, and yaw axes.
Sensors 24 04517 g008
Table 1. Regional division by target category.
Table 1. Regional division by target category.
Region NameTarget-Semantic Information
Potential dynamic regionkeyboards, chairs, books, mouses, etc.
Dynamic regionpeople, animals, mobile robots, etc.
Static regioncomputer screens, cabinets, windows, etc.
Table 2. Experimental results for absolute trajectory error (ATE [m]).
Table 2. Experimental results for absolute trajectory error (ATE [m]).
SequencesORB-SLAM2Dyna-SLAMDS-SLAMYPL-SLAMImprovement
against ORB-SLAM2
RMSESDRMSESDRMSESDRMSESDRMSE (%)SD (%)
fr3_s_hs0.0660090.0355070.0287540.0142520.0276260.0138210.0286960.01389356.5360.87
fr3_w_hs0.5082350.2341650.0270930.013290.031390.0162290.0276810.01372794.5594.14
fr3_w_rpy0.7604490.373880.0441610.0213850.3482570.2353030.0444710.02528294.1593.24
fr3_w_static0.0596430.0341030.0101980.0047060.0083540.0036630.0098970.00422583.4187.61
fr3_w_xyz0.6818580.3399610.0325470.0175120.0266530.0163540.0265670.01441596.195.76
Table 3. Experimental results for translational relative trajectory error.
Table 3. Experimental results for translational relative trajectory error.
SequencesORB-SLAM2Dyna-SLAMDS-SLAMYPL-SLAMImprovement
against ORB-SLAM2
RMSESDRMSESDRMSESDRMSESDRMSE (%)SD (%)
fr3_s_hs0.0357590.0238720.0243370.0136650.0237790.0132610.0165380.00994153.7558.36
fr3_w_hs0.163260.1301150.0236240.0124290.0295350.0153640.0234740.01224785.6290.59
fr3_w_rpy0.176490.1357350.0357980.0203390.1441310.1062490.0445620.03060274.7577.45
fr3_w_static0.0477030.0411160.0096430.0047070.0102640.0054830.0090580.00489881.0188.09
fr3_w_xyz0.1700890.1153730.0206750.010560.0197060.0119650.0193350.00973288.6391.56
Table 4. Experimental results for rotational relative trajectory error.
Table 4. Experimental results for rotational relative trajectory error.
SequencesORB-SLAM2Dyna-SLAMDS-SLAMYPL-SLAMImprovement
against ORB-SLAM2
RMSESDRMSESDRMSESDRMSESDRMSE (%)SD (%)
fr3_s_hs0.0161180.0082310.0154820.00740.0167140.0089620.0134640.00632316.4723.18
fr3_w_hs0.0842370.0649510.0157270.0082890.0194490.0095180.0156510.00830981.4287.21
fr3_w_rpy0.0867230.0659830.0188610.0106640.0659160.0365910.024660.01651671.5674.97
fr3_w_static0.021280.0174280.0065060.0027360.0066350.0029940.0062440.00291570.6683.27
fr3_w_xyz0.0811390.0539010.0112650.0057140.0103720.0054630.0111490.00572286.2689.38
Table 5. Experimental results for tracking time(s).
Table 5. Experimental results for tracking time(s).
SequencesORB-SLAM2Dyna-SLAMYPL-SLAMImprovement
against Dyna-SLAM (%)
fr3_s_hs0.01911.71730.079195.39%
fr3_w_hs0.02041.82740.073295.99%
fr3_w_rpy0.02011.77850.077295.66%
fr3_w_static0.01831.76470.073595.83%
fr3_w_xyz0.0211.76420.07895.58%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Du, X.; Zhang, C.; Gao, K.; Liu, J.; Yu, X.; Wang, S. YPL-SLAM: A Simultaneous Localization and Mapping Algorithm for Point–line Fusion in Dynamic Environments. Sensors 2024, 24, 4517. https://doi.org/10.3390/s24144517

AMA Style

Du X, Zhang C, Gao K, Liu J, Yu X, Wang S. YPL-SLAM: A Simultaneous Localization and Mapping Algorithm for Point–line Fusion in Dynamic Environments. Sensors. 2024; 24(14):4517. https://doi.org/10.3390/s24144517

Chicago/Turabian Style

Du, Xinwu, Chenglin Zhang, Kaihang Gao, Jin Liu, Xiufang Yu, and Shusong Wang. 2024. "YPL-SLAM: A Simultaneous Localization and Mapping Algorithm for Point–line Fusion in Dynamic Environments" Sensors 24, no. 14: 4517. https://doi.org/10.3390/s24144517

APA Style

Du, X., Zhang, C., Gao, K., Liu, J., Yu, X., & Wang, S. (2024). YPL-SLAM: A Simultaneous Localization and Mapping Algorithm for Point–line Fusion in Dynamic Environments. Sensors, 24(14), 4517. https://doi.org/10.3390/s24144517

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop