Next Article in Journal
Sensitive Competitive Electrochemical Immunosensor for Hg (II) Based on Molybdenum Disulfide/Reduced Graphene Oxide/Gold Nanocomposites
Previous Article in Journal
Beyond Information Distortion: Imaging Variable-Length Time Series Data for Classification
Previous Article in Special Issue
Parallelized SLAM: Enhancing Mapping and Localization Through Concurrent Processing
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Research on a Density-Based Clustering Method for Eliminating Inter-Frame Feature Mismatches in Visual SLAM Under Dynamic Scenes

1
Engineering Research and Design Institute of Agricultural Equipment, Hubei University of Technology, Wuhan 430068, China
2
Hubei Engineering Research Center for Intellectualization of Agricultural Equipment, Wuhan 430068, China
3
School of Mechanical Engineering, Hubei University of Technology, Wuhan 430068, China
4
Hubei Key Laboratory Modern Manufacturing Quality Engineering, School of Mechanical Engineering, Hubei University of Technology, Wuhan 430068, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(3), 622; https://doi.org/10.3390/s25030622
Submission received: 13 December 2024 / Revised: 12 January 2025 / Accepted: 17 January 2025 / Published: 22 January 2025

Abstract

:
Visual SLAM relies on the motion information of static feature points in keyframes for both localization and map construction. Dynamic feature points interfere with inter-frame motion pose estimation, thereby affecting the accuracy of map construction and the overall robustness of the visual SLAM system. To address this issue, this paper proposes a method for eliminating feature mismatches between frames in visual SLAM under dynamic scenes. First, a spatial clustering-based RANSAC method is introduced. This method eliminates mismatches by leveraging the distribution of dynamic and static feature points, clustering the points, and separating dynamic from static clusters, retaining only the static clusters to generate a high-quality dataset. Next, the RANSAC method is introduced to fit the geometric model of feature matches, eliminating local mismatches in the high-quality dataset with fewer iterations. The accuracy of the DSSAC-RANSAC method in eliminating feature mismatches between frames is then tested on both indoor and outdoor dynamic datasets, and the robustness of the proposed algorithm is further verified on self-collected outdoor datasets. Experimental results demonstrate that the proposed algorithm reduces the average reprojection error by 58.5% and 49.2%, respectively, when compared to traditional RANSAC and GMS-RANSAC methods. The reprojection error variance is reduced by 65.2% and 63.0%, while the processing time is reduced by 69.4% and 31.5%, respectively. Finally, the proposed algorithm is integrated into the initialization thread of ORB-SLAM2 and the tracking thread of ORB-SLAM3 to validate its effectiveness in eliminating feature mismatches between frames in visual SLAM.

1. Introduction

Visual SLAM enables self-localization and environmental mapping by capturing and processing feature information from the environment [1]. Dynamic features from moving objects may be incorrectly identified as part of the environment, resulting in localization errors and incorrect map information in visual SLAM. The ability to accurately eliminate dynamic features during the inter-frame matching process is critical to the robot’s localization accuracy and the quality of the environmental map construction. Camera hardware is cost-effective, adaptable, and provides rich image data, making it widely utilized in visual SLAM [2]. A major challenge in visual SLAM is the efficient and accurate removal of dynamic features while processing environmental data in dynamic environments.
Feature matching errors are among the primary causes of performance degradation in visual SLAM [3]. The Random Sample Consensus (RANSAC) method is widely used in visual SLAM to effectively eliminate inter-frame mismatches. However, when a dataset contains numerous outliers, the RANSAC method necessitates more iterations to compute an accurate geometric model. To address this issue, researchers have proposed several improvements. Raguram et al. [4] introduced the USAC method, which improves inlier accuracy and efficiency through modular design and various optimization techniques. The MLESAC method proposed by Torr et al. [5] enhances robustness by optimizing the probability function for model fitting. Chum et al. introduced the PROSAC [6] and LO-RANSAC [7] methods. PROSAC prioritizes sampling based on match-points confidence to improve algorithm speed, while LO-RANSAC incorporates local optimization steps to enhance inlier quality. He et al. [8] proposed an adaptive error threshold RANSAC method determined by inlier matching rates, which improves the algorithm’s accuracy in eliminating feature mismatches in both dynamic and static scenes. Although these methods have improved algorithm performance to some extent, they primarily rely on local neighborhood information of feature points in images, lacking global constraints. This makes them less efficient in high-noise data, dynamic scenes, or under complex geometric constraints [9], and they still face challenges in effectively eliminating mismatched points in dynamic environments. For image pairs with structurally similar or repetitively textured scenes in dynamic environments, the initial feature matching set often contains a large number of mismatches, significantly impacting the accuracy of subsequent camera pose estimation and 3D reconstruction in visual SLAM systems.
To address these issues, the main contributions of this study are as follows:
  • This paper proposes a density-based RANSAC method (DSSAC), which pre-processes the dataset through clustering and dividing it into dynamic and static clusters to accurately eliminate dynamic feature points. Since the data are pre-clustered, the required number of iterations is significantly reduced compared to traditional RANSAC. This method combines geometric analysis and clustering techniques, offering an efficient and accurate strategy for elimination mismatches.
  • RANSAC, as a highly robust fitting method, excels in fitting high-quality data. By combining the DSSAC method with RANSAC, the former effectively removes noise and clustered dynamic points through clustering, while the latter is employed for secondary screening of mismatched points, achieving higher accuracy and computational efficiency.
  • This study evaluates the performance of the DSSAC-RANSAC method and verifies its effectiveness in eliminating mismatches in indoor and outdoor dynamic scenes. Using reprojection error mean, reprojection error variance, and processing time as evaluation metrics, experimental results demonstrate that DSSAC-RANSAC significantly outperforms traditional methods in eliminating dynamic feature points. The algorithm is applied to the initialization thread of ORB-SLAM2 and the tracking thread of ORB-SLAM3 to validate its feasibility within visual SLAM systems.
Section 2 reviews related studies on feature mismatch elimination methods, analyzing their applications and limitations in feature matching and mismatch removal. Section 3 provides a detailed explanation of the proposed improved algorithm, including the core concept, algorithm design, and implementation details. Section 4 presents the experimental section, where the algorithm’s performance is tested on public datasets, and the applicability and robustness of the method are verified using self-collected datasets. The results are systematically organized and analyzed for clarity and insight. Additionally, the proposed algorithm is integrated into the initialization thread of ORB-SLAM2 and the tracking thread of ORB-SLAM3 to further evaluate its practical application. Section 5 concludes the paper and discusses future research directions, as well as potential improvements.

2. Related Works

Random Sample Consensus (RANSAC) is a widely used method for data fitting. It seeks a mathematical model that best fits the data characteristics by randomly selecting points from the sample and evaluates the fitting quality using reprojection error, iterating to identify the optimal model [10]. Fischler M.A. et al. [11] used RANSAC to eliminate mismatched feature points in image feature matching by randomly selecting four pairs of matching points from images A and B, then solving the homography matrix H using the Direct Linear Transform (DLT) method:
c u v 1 = H u v 1
In Equation (1), where c is a constant, p = ( u       v       1 ) t and p is a point in image A, p = ( u       v       1 ) t and p is a point in image B, p and p form a matching pair, and H = h 1       h 2       h 3 h 4       h 5       h 6 h 7       h 8       h 9 . After solving for the homography matrix, the reprojection error between the matching feature points of the two images is calculated and compared to a predefined reprojection error threshold. The suitability of the obtained homography matrix is then determined based on the inlier ratio. Subsequently, points are iteratively selected until the optimal homography matrix H and the inlier set are obtained within a predefined number of iterations.
Density-Based Spatial Clustering of Applications with Noise (DBSCAN) is a density-based spatial clustering method [12]. Unlike other clustering methods, DBSCAN eliminates the need to predetermine the number of clusters and can detect clusters of arbitrary shapes, leading to its extensive application in various domains [13]. In an improvement regarding adaptive parameter estimation, Jiang et al. [14] observed that correct matches during feature matching typically exhibit similar motion patterns, which can be grouped into several motion consistency groups, whereas incorrect matches tend to be randomly distributed in the image domain. Therefore, they modeled the feature matching process in visual SLAM as a spatial clustering problem.
During the clustering process, the neighborhood threshold r and the minimum number of points MinPts are critical parameters affecting clustering quality.
First, the weighted distance is used to compute the distance D between each pair of feature points. If feature points p i and q i exist in image A, and their corresponding feature points in image B are p j and q j , the distance d is calculated as:
d p q = d ( p i , q i ) + d ( p j , q j ) + ω i j × d ( p i j , q i j )
In Equation (2), the weight coefficient wij is defined as:
w p q = 1 + γ × e m i n { d ( p i , q i ) + d ( p j , q j ) }
In Equation (3), where d p q represents the weighted distance between two pairs of feature points, d ( p i , q i ) denotes the Euclidean distance between feature points p i and q i , and γ is a positive parameter enhancing motion consistency.
If there are N pairs of matching points, an N × N distance matrix D can be constructed after calculating all d p q values. The neighborhood distance r is calculated based on the maximum value MA and minimum value MI of D using the following formula:
r = λ ( M A M I ) + M I
In Equation (4), λ is a scaling parameter used to adjust the size of the neighborhood radius. The minimum number of points within a neighborhood, minpts, is defined as:
M i n P t s = N × p c t
In Equation (5), N represents the total number of matching point pairs, and pct is the predefined percentage, typically ranging from 3% to 5%.
In the point set U = { p 1 , p 2 , , p n } , for any point p U , if p satisfies the following conditions:
N r ( p ) = q U | d i s t ( p , q ) r N r ( p ) M i n P t s
The point p is defined as a core point if it meets the criteria. In Equation (6), d i s t ( p , q ) represents the distance between point p and point q, r is the defined neighborhood distance, and MinPts is the minimum number of points required within the neighborhood to qualify p as a core point.
If the number of points in the neighborhood of point p is less than MinPts, but it lies within the neighborhood of a core point, p is defined as a border point. Border points do not participate in the expansion of new clusters but are assigned to the cluster of their neighboring core points. Points that neither belong to the neighborhood of any core point nor meet the minimum point requirement are defined as noise points and are removed after clustering.
The clustering process of DBSCAN is performed by traversing the dataset. Starting with the first point that meets the core point criteria, its cluster is expanded by checking whether its neighboring points also satisfy the core point condition. If so, these points are added to the current cluster, and the process continues until no new core points can be identified. After traversing all points, those that do not belong to any cluster and do not meet the core point condition are labeled as noise points.
Liu et al. [15] adaptively calculated parameters using a k-distance graph before clustering, improving the robustness of the clustering method in handling datasets with varying densities. Bryant et al. [12] integrated RNN methods to precompute adaptive parameters based on the distribution of the data, enhancing the robustness of the method. Liu et al. [16] used KD-trees to optimize the traversal process for cluster expansion, thereby improving the execution speed of the clustering algorithm. Hu et al. [17] leveraged global information from feature matching to construct a feature matching set, using clustering methods to eliminate noise points within matching clusters, effectively removing mismatched pairs.
Although the aforementioned improvements significantly enhance the mismatch elimination capability of the DBSCAN method in static environments, certain limitations persist in dynamic environments. Specifically, multiple adjacent matching point pairs may appear on the surface of dynamic objects. Due to spatial proximity, these point pairs might be incorrectly classified into the same cluster, thus interfering with the correct distinction between dynamic and static features. Such misclassification significantly affects the subsequent pose estimation and map construction accuracy in the visual SLAM system, limiting its robustness and adaptability in dynamic scenarios. Inspired by this, this study combines DBSCAN and RANSAC methods to propose a density-based RANSAC method suited for dynamic environments. In dynamic scenarios with a large number of outliers, this method can accurately eliminate outliers, identify and remove dynamic clusters, and efficiently obtain a set of static matching pairs.

3. Methodology

3.1. DSSAC-RANSAC Framework

This paper proposes a density-based RANSAC method, which applies RANSAC three times to eliminate mismatches. The method introduces global clustering, enabling more accurate image feature matching in dynamic scenes by processing the initial set with density clustering, thereby significantly improving the algorithm’s speed. Since ORB is simpler and more efficient than SIFT, SURF, and other feature detectors [18], with advantages such as rotational invariance, it is widely used in real-time image processing and computer vision. In the following discussion, ORB is used to illustrate the algorithm.
The flowchart of DSSAC-RANSAC is shown in Figure 1, and the algorithm flow is as follows:
1.
Capture and output the corresponding 2D images P A and P B through the camera.
2.
Detect key points in images P A and P B , calculate the descriptors of the key points, and output the ORB feature points.
3.
Perform brute-force matching of ORB features based on Hamming distance [19], obtaining the initial matching set C B F .
4.
Perform down-sampling on the sample set to obtain the matching set C D .
5.
Apply the DSSAC method proposed in this paper for local clustering, obtaining the clustered set C D J , and then remove dynamic features to obtain the static point set C M .
6.
Use the RANSAC method to perform local optimization on the point set C M .

3.2. Density-Based Segmentation RANSAC

Traditional RANSAC is a method that randomly selects points from a dataset to establish a mathematical model, obtaining the best fit by minimizing the reprojection error. The quality of the selected points significantly impacts the fitting performance of the mathematical model. Due to the randomness of point selection, traditional RANSAC requires a large number of iterations to achieve satisfactory matching results, particularly in scenarios with a high number of dynamic features. For different transformations in various scenarios, correct feature matches tend to exhibit coherent motion, with neighboring points sharing similar motion [14]. Based on this characteristic, this paper proposes a density-based segmentation RANSAC (DSSAC) method to preprocess the sample set, establish a reprojection error threshold and inlier ratio, and eliminate dynamic feature points.
DBSCAN is used to perform clustering on all motion vectors, generating multiple clusters. Dense clusters typically contain feature matches with similar motion vectors, while sparse clusters or outliers correspond to mismatches. The algorithm framework is illustrated in Figure 2. Similar to the DBSCAN method, it constructs a feature matching set as follows:
D = { x i T , y i T , m i T } T , ( i = 1 , 2 , , N )
Here, D represents the dataset containing the matched point coordinates and motion vectors of two images. x i , y i are 2D vectors representing the spatial positions of the matching points p i , j and q i , j in images P A and P B , and m i represents the motion vector of the two matched feature points.
The sample set is divided into distinct clusters based on spatial density, and the resulting feature point clusters are defined as:
C = { C 1 , C 1 , C 2 , , C K }
Each cluster C i contains several feature point pairs with similar motion vectors, and i = 1 represents the mismatch cluster.
In the cluster C i , the number of matched feature point pairs N > 4 , which allows at least 4 pairs of feature points ( p i , j , q i , j ) to be randomly selected as samples for estimating the homography matrix H of the cluster. Using these sample point pairs, the optimal homography matrix H is obtained by minimizing the geometric error between the matching pairs. Project the position p i , j of each feature point pair in C from image P A to image P B to obtain the projected position q i , j , then calculate the reprojection error e i , j between q i , j and the actual matched point, as shown in Equation (7):
e i , j = q ^ i , j q i , j
Set a reprojection threshold τ . If e i , j < τ , the point is marked as an inlier. After calculating the reprojection errors for all matched points, the number of inliers n that meet the conditions is counted, and the inlier ratio of the cluster is calculated as follows:
α = n N
In Equation (8), n is the number of inliers that meet the conditions, N is the total number of matches, and α is the inlier ratio of the cluster. To determine the dynamics of a cluster, an inlier ratio threshold α m i n is set. If the inlier ratio of cluster C i , α > α m i n , the cluster is classified as a static cluster; otherwise, all feature point pairs within the cluster are marked as noise and excluded from the static feature set. This step effectively eliminates the feature point pairs of dynamic clusters while retaining the feature points in static clusters, thereby reducing the impact of dynamic scenes on the localization accuracy of the visual SLAM system. The specific algorithm flow is outlined as follows:
1.
Compute the distance matrix and cluster the point set into C = { C 1 , C 1 , C 2 , , C K } .
2.
Initialize parameters: set the reprojection error threshold τ , the minimum inlier ratio α m i n , the matching set D , and initialize the number of inliers n = 0 .
3.
Select four matching pairs from C i and compute the homography matrix H i .
4.
Using the homography matrix H i , calculate the number of inliers n in which the reprojection error e i , j , between image P A projected to image P B is less than τ .
5.
Calculate the inlier ratio α of the cluster by using the total number of matches N and the number of inliers n that meet the conditions.
6.
If α > α m i n , classify the cluster as a static cluster.
7.
Iterate through all clusters C i in the dataset C based on their labels, repeat steps (3) to (6), and add all static points to the set M .
This algorithm preprocesses the dataset by clustering the points and computes the homography matrix to eliminate dynamic clusters. It considers both the initial positions and motion vectors of feature points to cluster matches exhibiting consistent motion. Since the data have been preprocessed, satisfactory results can be achieved with fewer iterations when calculating the reprojection error. Since there is another selection step later, the reprojection error threshold τ should be set relatively high. If τ is set too low, it may result in insufficient inliers. Generally, τ is chosen within the range of 50–100. To account for specific environments, such as images containing many dynamic features, the minimum inlier ratio α m i n should be set lower. If set too high, static clusters may be misclassified as dynamic, thereby affecting the accuracy of the method. Experimental results suggest that setting α m i n to a value between 0.3 and 0.5 produces better results. It is worth noting that the traditional RANSAC method requires at least four inliers to compute the transformation matrix, while the clusters obtained through clustering already satisfy the minimum inlier requirement. Therefore, there is no need to evaluate the inlier count when selecting points from the clusters. The algorithm flow of DSSAC is outlined as follows:

3.3. Image Feature Mismatch Elimination Method Combining DSSAC and RANSAC

After the DSSAC method eliminates dynamic features and significant mismatches, a small number of mismatches may remain in the dataset, potentially affecting the quality of matched points. Although the RANSAC method is robust in the feature matching and effectively handles outliers, it relies on predefined mathematical models, which can reduce its efficiency and accuracy in environments with numerous dynamic features or a large number of outliers. The dataset processed by the DSSAC method eliminates most dynamic features and outliers, thereby significantly reducing interference during the subsequent model fitting process. This enables RANSAC to rapidly and accurately estimate geometric models on high-quality datasets. By combining DSSAC with RANSAC, the DSSAC method effectively removes dynamic features and significant mismatches through clustering, providing more accurate matched point data for subsequent processing. RANSAC further eliminates residual mismatches with fewer iterations, thereby ensuring the robustness and accuracy of geometric model fitting. The combination of these two methods significantly enhances computational efficiency and matching accuracy in dynamic environments. The flow chart of the algorithm in this paper is shown in Figure 3.
In Figure 4, to enhance the observation of details, yellow rectangles are used to highlight regions with variations in feature points, and certain areas are further magnified for clarity. First, perform ORB feature extraction on the input images, followed by brute-force matching of the extracted features (as illustrated in Figure 4a). Second, apply the proposed DSSAC method for global clustering of feature point pairs (with different colors representing different clusters) and initial elimination of mismatches (as shown in Figure 4b), followed by the removal of dynamic feature points (as shown in Figure 4c). Finally, the RANSAC method is employed for local refinement, yielding precise matches (as depicted in Figure 4d).
As illustrated in Figure 4a, brute-force matching methods typically produce a large number of mismatches and dynamic feature points. If RANSAC is directly applied through random sampling iterations, the high probability of selecting mismatched points means that some mismatches may repeatedly appear during multiple iterations, requiring numerous iterations to find the optimal inlier set. Furthermore, each iteration necessitates the calculating of the reprojection error for all matched points, which significantly impacts the algorithm’s efficiency. Therefore, the algorithm proposed in this paper preprocesses the sample set using clustering methods, treating outliers as noise and removing most mismatched points. It classifies feature points based on spatial consistency, effectively identifying and excluding dynamic feature points. However, DSSAC faces challenges in fully removing matched points with small motion vectors that are classified into static clusters. Combining the two algorithms can more effectively eliminate outliers. In the DSSAC algorithm, to ensure the inclusion of all potential static feature point clusters, the reprojection error threshold is set within a relatively large range of 50–100. This relatively loose threshold setting allows for more comprehensive coverage of static features, thereby enhancing the robustness of the data preprocessing stage. The DSSAC algorithm significantly optimizes the initial dataset by effectively eliminating dynamic features. The RANSAC reprojection error threshold is set to 25–50, with the number of iterations limited to 10–20. Since DSSAC preprocessing has significantly reduced the proportion of residual outliers, this configuration with fewer iterations achieves a good balance between accuracy and efficiency.

4. Experiments

4.1. Datasets and Evaluation Metrics

This study utilizes the TUM [20] and KITTI [21] datasets to conduct experiments on inter-frame image feature matching, with a focus on the DSSAC method in both indoor and outdoor dynamic scenarios to analyze and compare its accuracy and runtime performance. The accuracy tests employ two metrics: the mean error ( e m e a n ) and the error variance ( e V a r ) , defined as:
e m e a n = i = 1 n q ^ i , j q i , j N
e V a r = i = 1 n ( e m e a n q ^ i , j q i , j ) 2 N
In Equation (9), the mean error e m e a n represents the average magnitude of all feature point errors and is used to measure the overall fitting accuracy. In Equation (10), the error variance e V a r is used to characterize the dispersion of the data, reflecting the stability of the fitting results. These two metrics collectively represent the matching accuracy and robustness of the method in different environments.
I m p r v ( E ) = 1 E o u r E o t h e r × 100 %
In Equation (11), E represents the evaluation metric, which is used to calculate the percentage improvement of the proposed algorithm relative to other algorithms, thereby quantifying the performance enhancement.

4.2. Mismatch Elimination in Different Scenarios

As shown in Figure 5, Figure 6 and Figure 7, two experimental environments were established: an indoor dynamic environment and an outdoor dynamic environment. Yellow rectangles are used to delineate dynamic objects, with specific regions magnified to enhance the comparison of matching results produced by different algorithms. Figure 5 depicts an indoor dynamic environment where two individuals exhibit gesture variations and overall movement. Figure 6 illustrates vehicles moving on an outdoor road. Figure 7 presents self-collected images depicting a queue of individuals on a playground.
ORB feature points were extracted for each experimental scenario. The RANSAC threshold was set to 3, with 50 iterations performed. In GMS-RANSAC, the GMS grid size parameter was set to the default value of 20, with the RANSAC threshold unchanged, and the iteration count reduced to 30 to improve algorithm speed. In GMSATRANSAC, the GMS parameters were kept unchanged, and the internal matching rate was set to 40%. The self-collected dataset shown in Figure 7 was captured using the Intel RealSense D435i camera, with a resolution of 1920 × 1080 and a frame rate of 30 FPS. To enhance image quality, both brightness and contrast were increased by 5%. Monocular mode was employed for data collection to meet the algorithm’s requirements.
As shown in Figure 5, in the indoor dynamic environment, individuals exhibit relatively small movements, and the complex background causes dynamic feature points to be easily confused with the static scene. The RANSAC, GMS-RANSAC and GMS-ATRANSAC algorithms still exhibit feature mismatches on dynamic individuals, where as the DSSAC-RANSAC algorithm more effectively eliminates mismatches in these dynamic features. The DSSAC-RANSAC algorithm uses a clustering method that combines density and distance, leveraging the spatial similarity of dynamic features in indoor scenes to more accurately distinguish dynamic objects from the static background, thereby more effectively removing dynamic feature points. As shown in Figure 6, in outdoor dynamic scenarios, the first two algorithms still produce mismatches on dynamic objects. However, DSSAC-RANSAC, with its dynamic feature recognition mechanism, effectively distinguishes and identifies moving objects, removing their associated feature points during data processing. Therefore, DSSAC-RANSAC more effectively eliminates mismatched feature points in dynamic environments. In the self-collected dataset shown in Figure 7, due to the similarity between dynamic features and the background, the RANSAC and GMS-RANSAC algorithms fail to fully eliminate mismatches related to dynamic individuals. In contrast, DSSAC-RANSAC effectively eliminates all mismatches related to dynamic individuals.
Experimental results show that the DSSAC-RANSAC algorithm exhibits superior robustness in dynamic scenarios, effectively eliminating image feature mismatches in both indoor and outdoor dynamic environments, as well as in the self-collected dataset.
Figure 8 illustrates the mean error and error variance of four algorithms, RANSAC, GMS-RANSAC, GMS-ATRANSAC, and DSSAC-RANSAC, across different numbers of feature points. It is clear that DSSAC-RANSAC consistently maintains low mean error and variance, regardless of the feature point quantity. This demonstrates that DSSAC-RANSAC not only effectively removes outliers but also maintains high fitting accuracy, bringing the matching results closer to the true values.
Figure 9 shows the runtime of four algorithms with different numbers of feature points. RANSAC relies on extensive random sampling to find the optimal inlier set, requiring repeated computations of the model fitting performance. In scenarios with many outliers, the number of samples increases significantly, and each iteration must traverse all feature points, resulting in longer computation times as the number of feature points grows. The GMS-RANSAC and GMS-ATRANSAC methods, although introducing multi-scale matching strategies with multiple scale divisions and iterative evaluations for feature points, require intensive neighborhood matching computations for each scale division and iteration. All three algorithms mentioned above lack a mechanism for removing dynamic features. Feature points on dynamic objects are repeatedly computed, increasing the time cost with each calculation. The DSSAC-RANSAC method uses a dynamic cluster elimination mechanism to effectively identify and remove mismatched clusters on dynamic objects, avoiding redundant computations on dynamic objects. Additionally, uniform down-sampling of samples is performed before clustering, significantly reducing computation time during the matching process and demonstrating a notable speed advantage.
Table 1, Table 2 and Table 3 show the performance of three algorithms under different ORB feature points, including the mean error, error variance, and computation time in both classic datasets and self-collected datasets.
Table 4, Table 5 and Table 6 respectively show the percentage improvement in error methods, mean error difference, time, and other metrics for the DSSAC-RANSAC algorithm with 1500, 2000, and 2500 ORB feature points compared to three other algorithms. From Table 4, Table 5 and Table 6, it can be seen that the DSR algorithm shows significant advantages in all three evaluation metrics for different numbers of feature points.
Table 7 shows the average percentage improvement in error variance, mean error difference, and runtime, among other metrics, for DSSAC-RANSAC compared to the other three algorithms at different feature point quantities. Compared to RANSAC, DSSAC-RANSAC reduced the mean error difference by 58.5% on average and the error variance by 65.2%. Compared to GMS-RANSAC, the mean error difference decreased by 49.2% on average, and the error variance decreased by 63.0%. Compared to GMS-ATRANSAC, the mean error difference decreased by 36.2% on average, and the error variance decreased by 50.6%. In terms of runtime, DSSAC-RANSAC reduced the time by 69.4% on average compared to RANSAC; by 31.5% compared to GMS-RANSAC; and by 13.7% compared to GMS-ATRANSAC. Therefore, DSSAC-RANSAC is a fast and accurate method for eliminating image feature mismatches.

4.3. Improvements to the ORB-SLAM3 Feature Matching Module Using the DSSAC-RANSAC Algorithm

The DSSAC-RANSAC algorithm was applied to ORB-SLAM3 to eliminate inter-frame mismatches, improving the monocular initialization thread of ORB-SLAM2 and the tracking thread of ORB-SLAM3. Tests were conducted using the rgbd_dataset_freiburg3_walking_xyz sequence from the TUM dataset.

4.3.1. Monocular Initialization of ORB-SLAM2

Monocular initialization is a critical module in SLAM visual systems, primarily responsible for completing the initial pose estimation. Accurate monocular initialization is crucial for improving the precision and quality of map construction. Since ORB-SLAM3 introduces IMU support, monocular initialization utilizes IMU data for assistance. To focus on the improvement effects of pure vision algorithms, the experiments adopted the monocular initialization framework of ORB-SLAM2. This study applies the DSSAC-RANSAC algorithm to the initialization thread of ORB-SLAM2 and compares it with the default initialization algorithm to evaluate the improvements in inter-frame feature matching.
In the experimental image sequence, the characters wearing a plaid shirt and a black top are both moving in the current frame.In Figure 10, the two ends of the green line segments in the image represent the successfully matched ORB feature points between the previous and current frames, reflecting the motion relationship of feature points between frames. Figure 10a shows the default initialization algorithm of ORB-SLAM2. Due to the movement of the characters and the presence of many similar features on the surface of the plaid shirt, a large number of mismatched points appear on the back. Additionally, the dynamic feature points on the character in a black top on the right were not effectively removed. Keeping these mismatched points during initialization not only significantly affects the accuracy of the initial pose estimation but also negatively impacts the subsequent map building process. Figure 10b shows the improved initialization method proposed in this paper, where the dynamic feature matches on both moving characters were successfully removed, and no mismatched points appeared.

4.3.2. Tracking Thread of ORB-SLAM3

The proposed algorithm was applied to the tracking thread of ORB-SLAM3 to verify its capability to eliminate mismatches in the SLAM system.
In Figure 11a,b, the individual on the left exhibits upper-body and leg movements, while the individual on the right rotates their upper body, the green small rectangles denote the tracked feature points. During the feature matching stage, the DSSAC-RANSAC algorithm effectively distinguishes dynamic features from static background elements, preventing dynamic objects from negatively affecting the camera’s initial pose estimation and map construction. In Figure 11c,d, the red points denote 3D map points derived from feature points preserved through matching across multiple image frames. Figure 11c shows the point cloud map generated by ORB-SLAM3’s default method. It can be observed that the number of near-point clouds is higher and more chaotic due to the influence of dynamic people. Figure 11d presents the depth point cloud map with effective ORB feature points retained, where the features of dynamic individuals near the seats on both sides of the near points are effectively removed. Experimental results show that the algorithm proposed in this paper can robustly and accurately remove dynamic feature points in ORB-SLAM3’s tracking thread, significantly improving the accuracy of local map construction and loop closure detection.

5. Conclusions

This paper proposes a density-based improved RANSAC method (DSSAC), which is combined with the traditional RANSAC method. The DSSAC method demonstrates robust outlier elimination capabilities in environments with a high proportion of outliers, effectively distinguishing between dynamic and static clusters to generate high-quality datasets. Building on this, the RANSAC method is employed to further filter a small number of residual outliers. Experimental results show that the DSSAC-RANSAC method significantly reduces mismatches and dynamic feature point interference, achieving higher accuracy and computational efficiency compared to traditional methods. Experiments on different dynamic scenarios demonstrate that the proposed method exhibits superior robustness and accuracy and has been validated on self-collected datasets. Additionally, the DSSAC-RANSAC method was applied to the initialization thread of ORB-SLAM2 and the tracking thread of ORB-SLAM3 for experiments. The results indicate that the method effectively removes dynamic feature points in keyframes, improves the quality of keyframe feature points, and enhances the accuracy and reliability of robot pose estimation and environmental map construction. Future research will focus on further optimizing clustering methods for dynamic features to minimize their impact on static environment features. Additionally, after removing dynamic features, the generated static feature clusters will undergo in-depth analysis, treating each cluster as an independent visual word and integrating its information into the Bag of Words (BoW) model. This improvement is expected to further enhance the processing capability, robustness, and accuracy of visual SLAM systems in dynamic environments.

Author Contributions

Conceptualization, Z.Y.; methodology, Z.Y.; software, Z.Y.; formal analysis, S.Y. and C.Z.; investigation, S.Y. and Y.X.; resources, S.Y. and L.D.; writing—original draft preparation, K.Z.; writing—review and editing, K.Z.; visualization, K.Z. and S.Y.; project administration, D.Z.; funding acquisition, D.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Hubei Provincial Science and Technology Innovative Talent Program Project grant number 2023DJC092, Hubei Provincial Department of Education Key Project grant number D20221404, National Natural Science Foundation of China grant number 52075152, Hubei Provincial Natural Science Foundation Innovation Group Project grant number 2023AFA037, National Natural Science Foundation of China grant number 51907055, Hubei Province Key Research and Development Program, 2022BBA0016, Development and Application of Digital Sorting Technology and Equipment for Crayfish.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Abaspur Kazerouni, I.; Fitzgerald, L.; Dooly, G.; Toal, D. A Survey of State-of-the-Art on Visual SLAM. Expert Syst. Appl. 2022, 205, 117734. [Google Scholar] [CrossRef]
  2. Fuentes-Pacheco, J.; Ruiz-Ascencio, J.; Rendón-Mancha, J.M. Visual Simultaneous Localization and Mapping: A Survey. Artif. Intell. Rev. 2015, 43, 55–81. [Google Scholar] [CrossRef]
  3. Cadena, C.; Carlone, L.; Carrillo, H.; Latif, Y.; Scaramuzza, D.; Neira, J.; Reid, I.; Leonard, J.J. Past, Present, and Future of Simultaneous Localization and Mapping: Toward the Robust-Perception Age. IEEE Trans. Robot. 2016, 32, 1309–1332. [Google Scholar] [CrossRef]
  4. Raguram, R.; Chum, O.; Pollefeys, M.; Matas, J.; Frahm, J.-M. USAC: A Universal Framework for Random Sample Consensus. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 2022–2038. [Google Scholar] [CrossRef]
  5. Peng, L.; Zhang, Y.; Zhou, H.; Lu, T. A Robust Method for Estimating Image Geometry With Local Structure Constraint. IEEE Access 2018, 6, 20734–20747. [Google Scholar] [CrossRef]
  6. Chum, O.; Matas, J. Matching with PROSAC—Progressive Sample Consensus. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05), San Diego, CA, USA, 20–25 June 2005; IEEE: San Diego, CA, USA, 2005; Volume 1, pp. 220–226. [Google Scholar]
  7. Chum, O.; Matas, J.; Kittler, J. Locally Optimized RANSAC. In Pattern Recognition; Michaelis, B., Krell, G., Eds.; Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 2003; Volume 2781, pp. 236–243. ISBN 978-3-540-40861-1. [Google Scholar]
  8. Yang, Z.; He, Y.; Zhao, K.; Lang, Q.; Duan, H.; Xiong, Y.; Zhang, D. Research on Inter-Frame Feature Mismatch Removal Method of VSLAM in Dynamic Scenes. Sensors 2024, 24, 1007. [Google Scholar] [CrossRef] [PubMed]
  9. Myatt, D.R. Robust Estimation in High Noise and Highly Dimensional Data Sets with Applications to Machine Vision. arXiv 2024, arXiv:2409.12805. [Google Scholar]
  10. Dubrofsky, E. Homography Estimation. Master’s Thesis, University of British Columbia, Vancouver, BC, Canada, 2009. [Google Scholar]
  11. Fischler, M.A.; Bolles, R.C. Random sample consensus: A paradigm for model fitting with applications to image analysis and automated cartography. Commun. ACM 1981, 24, 381–395. [Google Scholar] [CrossRef]
  12. Ester, M.; Kriegel, H.-P.; Xu, X. A Density-Based Algorithm for Discovering Clusters in Large Spatial Databases with Noise. In Proceedings of the 2nd International Conference on Knowledge Discovery and Data Mining (KDD-96), Portland, OR, USA, 2–4 August 1996; Volume 96, pp. 226–231. [Google Scholar]
  13. Mumtaz, K. An Analysis on Density Based Clustering of Multi Dimensional Spatial Data. Indian J. Comput. Sci. Eng. 2010, 1, 8–12. [Google Scholar]
  14. Jiang, X.; Ma, J.; Jiang, J.; Guo, X. Robust Feature Matching Using Spatial Clustering with Heavy Outliers. IEEE Trans. Image Process. 2020, 29, 736–746. [Google Scholar] [CrossRef] [PubMed]
  15. Liu, P.; Zhou, D.; Wu, N. VDBSCAN: Varied Density Based Spatial Clustering of Applications with Noise. In Proceedings of the 2007 International Conference on Service Systems and Service Management, Chengdu, China, 9–11 June 2007. [Google Scholar]
  16. Liu, B. A Fast Density-Based Clustering Algorithm for Large Databases. In Proceedings of the 2006 International Conference on Machine Learning and Cybernetics, Dalian, China, 13–16 August 2006; IEEE: Piscataway, NJ, USA, 2006; pp. 996–1000. [Google Scholar]
  17. Hu, L.; Zuo, W.; Zhang, J. A Mismatch Elimination Method Based on Reverse Nearest Neighborhood and Influence Space. J. Comput. Aided Des. Comput. Graph. 2022, 34, 449–458. [Google Scholar] [CrossRef]
  18. Rublee, E.; Rabaud, V.; Konolige, K.; Bradski, G. ORB: An Efficient Alternative to SIFT or SURF. In Proceedings of the 2011 International Conference on Computer Vision, Barcelona, Spain, 6–13 November 2011; IEEE: Piscataway, NJ, USA, 2011; pp. 2564–2571. [Google Scholar]
  19. Hadrovic, E.; Osmankovic, D.; Velagic, J. Aerial Image Mosaicing Approach Based on Feature Matching. In Proceedings of the 2017 International Symposium ELMAR, Zadar, Croatia, 18–20 September 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 177–180. [Google Scholar]
  20. Tenorth, M.; Bandouch, J.; Beetz, M. The TUM Kitchen Data Set of Everyday Manipulation Activities for Motion Tracking and Action Recognition. In Proceedings of the 2009 IEEE 12th International Conference on Computer Vision Workshops, ICCV Workshops, Kyoto, Japan, 27 September–4 October 2009; IEEE: Piscataway, NJ, USA, 2009; pp. 1089–1096. [Google Scholar]
  21. Geiger, A.; Lenz, P.; Stiller, C.; Urtasun, R. Vision Meets Robotics: The KITTI Dataset. Int. J. Robot. Res. 2013, 32, 1231–1237. [Google Scholar] [CrossRef]
Figure 1. DSSAC-RANSAC algorithm flowchart; the images are derived from the KITTI dataset.
Figure 1. DSSAC-RANSAC algorithm flowchart; the images are derived from the KITTI dataset.
Sensors 25 00622 g001
Figure 2. Flowchart of the DSSAC method.
Figure 2. Flowchart of the DSSAC method.
Sensors 25 00622 g002
Figure 3. Methodology flowchart of this study.
Figure 3. Methodology flowchart of this study.
Sensors 25 00622 g003
Figure 4. ORB feature matching in the TUM dataset using DSSAC-RANSAC with 1500 feature points. (a) Brute-force matching, (b,c) clustering and dynamic cluster elimination process of DSSAC, and (d) results after RANSAC removes a small number of mismatches and low-quality points.
Figure 4. ORB feature matching in the TUM dataset using DSSAC-RANSAC with 1500 feature points. (a) Brute-force matching, (b,c) clustering and dynamic cluster elimination process of DSSAC, and (d) results after RANSAC removes a small number of mismatches and low-quality points.
Sensors 25 00622 g004
Figure 5. Comparison of feature matching performance between RANSAC, GMS-RANSAC (G-R), GMS-ATRANSAC (G-ATR) and DSSAC-RANSAC (D-R) in TUM indoor dynamic scenes.
Figure 5. Comparison of feature matching performance between RANSAC, GMS-RANSAC (G-R), GMS-ATRANSAC (G-ATR) and DSSAC-RANSAC (D-R) in TUM indoor dynamic scenes.
Sensors 25 00622 g005aSensors 25 00622 g005b
Figure 6. Comparison of feature matching performance between RANSAC, GMS-RANSAC (G-R), GMS-ATRANSAC (G-ATR) and DSSAC-RANSAC (D-R) in KITTI outdoor dynamic scenes.
Figure 6. Comparison of feature matching performance between RANSAC, GMS-RANSAC (G-R), GMS-ATRANSAC (G-ATR) and DSSAC-RANSAC (D-R) in KITTI outdoor dynamic scenes.
Sensors 25 00622 g006
Figure 7. Comparison of feature matching performance between RANSAC, GMS-RANSAC (G-R), GMS-ATRANSAC (G-ATR) and DSSAC-RANSAC (D-R) in self-collected outdoor scenes.
Figure 7. Comparison of feature matching performance between RANSAC, GMS-RANSAC (G-R), GMS-ATRANSAC (G-ATR) and DSSAC-RANSAC (D-R) in self-collected outdoor scenes.
Sensors 25 00622 g007
Figure 8. Comparison of the matching mean error (a) and error variance (b) of RANSAC, GMS-RANSAC, GMS-ATRANSAC and DSSAC-RANSAC at different numbers of feature points.
Figure 8. Comparison of the matching mean error (a) and error variance (b) of RANSAC, GMS-RANSAC, GMS-ATRANSAC and DSSAC-RANSAC at different numbers of feature points.
Sensors 25 00622 g008
Figure 9. Runtime of four algorithms with different numbers of feature points.
Figure 9. Runtime of four algorithms with different numbers of feature points.
Sensors 25 00622 g009
Figure 10. ORB-SLAM2 default initialization method (a) and proposed initialization method (b) in the TUM indoor dynamic dataset.
Figure 10. ORB-SLAM2 default initialization method (a) and proposed initialization method (b) in the TUM indoor dynamic dataset.
Sensors 25 00622 g010
Figure 11. DSSAC-RANSAC applied to feature tracking between adjacent keyframes in ORB-SLAM3, as shown in (a,b). (c) Depth point cloud map of the indoor scene constructed using the default method of ORB-SLAM3. (d) Depth point cloud map of the indoor scene constructed by integrating the method proposed in this paper.
Figure 11. DSSAC-RANSAC applied to feature tracking between adjacent keyframes in ORB-SLAM3, as shown in (a,b). (c) Depth point cloud map of the indoor scene constructed using the default method of ORB-SLAM3. (d) Depth point cloud map of the indoor scene constructed by integrating the method proposed in this paper.
Sensors 25 00622 g011
Table 1. Accuracy and speed comparison of RANSAC (RAN), GMS-RANSAC (G-R), GMS-ATRANSAC (G-ATR) and the proposed algorithm (D-R) when processing 1500 ORB feature points.
Table 1. Accuracy and speed comparison of RANSAC (RAN), GMS-RANSAC (G-R), GMS-ATRANSAC (G-ATR) and the proposed algorithm (D-R) when processing 1500 ORB feature points.
SequencesMetricsRANG-RG-ATRD-R
TUM
(Indoor dynamic)
e m e a n 2.4371.9131.2371.084
e V a r 1.2370.8130.6840.284
time (ms)185736767
KITTI
(Outdoor dynamic)
e m e a n 2.3721.8971.3510.957
e V a r 1.1130.9010.7050.253
time (ms)187827066
Self-collected
(Indoor dynamic)
e m e a n 2.6752.1251.4111.021
e V a r 1.3521.0210.8250.275
time (ms)178786858
Self-collected
(Outdoor dynamic)
e m e a n 2.7712.0951.3821.102
e V a r 1.5311.1030.8330.301
time (ms)189796354
Table 2. Accuracy and speed comparison of RANSAC (RAN), GMS-RANSAC (G-R), GMS-ATRANSAC (G-ATR) and the proposed algorithm (D-R) when processing 2000 ORB feature points.
Table 2. Accuracy and speed comparison of RANSAC (RAN), GMS-RANSAC (G-R), GMS-ATRANSAC (G-ATR) and the proposed algorithm (D-R) when processing 2000 ORB feature points.
SequencesMetricsRANG-RG-ATRD-R
TUM
(Indoor dynamic)
e m e a n 2.4551.7011.1141.048
e V a r 1.1550.8010.5010.248
time (ms)2401118373
KITTI
(Outdoor dynamic)
e m e a n 2.3511.7381.1251.106
e V a r 1.0380.7860.5220.283
time (ms)2371098371
Self-collected
(Indoor dynamic)
e m e a n 2.5361.8351.2210.973
e V a r 1.2570.8970.6130.375
time (ms)2211077969
Self-collected
(Outdoor dynamic)
e m e a n 2.7021.8441.4321.124
e V a r 1.2150.9110.6710.331
time (ms)2171098071
Table 3. Accuracy and speed comparison of RANSAC (RAN), GMS-RANSAC (G-R), GMS-ATRANSAC (G-ATR) and the proposed algorithm (D-R) when processing 2500 ORB feature points.
Table 3. Accuracy and speed comparison of RANSAC (RAN), GMS-RANSAC (G-R), GMS-ATRANSAC (G-ATR) and the proposed algorithm (D-R) when processing 2500 ORB feature points.
SequencesMetricsRANG-RG-ATRD-R
TUM
(Indoor dynamic)
e m e a n 2.6251.5151.1871.138
e V a r 1.3250.7150.5780.314
time (ms)29913110282
KITTI
(Outdoor dynamic)
e m e a n 2.5131.4191.2321.056
e V a r 1.2870.7490.6130.289
time (ms)28713410281
Self-collected
(Indoor dynamic)
e m e a n 2.6331.5331.3191.121
e V a r 1.3970.8130.6470.348
time (ms)27512910083
Self-collected
(Outdoor dynamic)
e m e a n 2.6711.6011.2811.005
e V a r 1.4330.7810.6340.482
time (ms)2731299977
Table 4. Percentage improvement of DSSAC-RANSAC over RANSAC, GMS-RANSAC and GMS-ATRANSAC in error variance, mean error, and runtime when processing 1500 ORB feature points.
Table 4. Percentage improvement of DSSAC-RANSAC over RANSAC, GMS-RANSAC and GMS-ATRANSAC in error variance, mean error, and runtime when processing 1500 ORB feature points.
Sequences I m p r v ( e m e a n )
For RAN
(%)
I m p r v ( e m e a n )
For G-R
(%)
I m p r v ( e m e a n )
For G-ATR
(%)
I m p r v ( e V a r )
For RAN
(%)
I m p r v ( e V a r )
For G-R
(%)
I m p r v ( e V a r )
For G-ATR
(%)
I m p r v ( t i m e )
For RAN
(%)
I m p r v ( t i m e )
For G-R
(%)
I m p r v ( t i m e )
For G-ATR
(%)
TUM
(Indoor
dynamic)
55.5%77.0%12.4%43.3%65.0%58.5%63.8%8.2%0%
KITTI
(Outdoor dynamic)
59.7%77.2%29.2%49.6%77.3%64.1%64.7%19.5%5.7%
Self-
collected
(Indoor
dynamic)
61.8%79.7%27.6%52.0%73.1%66.7%71.4%25.6%14.7%
Self-
collected
(Outdoor dynamic)
60.2%80.3%20.2%47.4%72.7%63.7%71.4%31.6%14.3%
AVG.59.3%78.6%22.4%48.1%72.0%63.3%67.8%21.2%8.7%
Table 5. Percentage improvement of DSSAC-RANSAC over RANSAC, GMS-RANSAC and GMS-ATRANSAC in error variance, mean error, and runtime when processing 2000 ORB feature points.
Table 5. Percentage improvement of DSSAC-RANSAC over RANSAC, GMS-RANSAC and GMS-ATRANSAC in error variance, mean error, and runtime when processing 2000 ORB feature points.
Sequences I m p r v ( e m e a n )
For RAN
(%)
I m p r v ( e m e a n )
For G-R
(%)
I m p r v ( e m e a n )
For G-ATR
(%)
I m p r v ( e V a r )
For RAN
(%)
I m p r v ( e V a r )
For G-R
(%)
I m p r v ( e V a r )
For G-ATR
(%)
I m p r v ( t i m e )
For RAN
(%)
I m p r v ( t i m e )
For G-R
(%)
I m p r v ( t i m e )
For G-ATR
(%)
TUM
(Indoordynamic)
57.3%38.4%6%78.5%69.0%50.0%69.6%34.2%12.0%
KITTI
(Outdoor dynamic)
53.0%36.4%2%72.7%64.0%45.8%70.0%34.9%14.5%
Self-
collected
(Indoor
dynamic)
61.6%47.0%20.3%70.2%58.2%38.8%68.8%35.5%12.3%
Self-
collected
(Outdoor dynamic)
58.4%39.0%21.5%72.8%63.7%50.7%67.3%34.9%11.2%
AVG.57.6%40.2%12.5%73.6%63.7%46.3%68.9%34.9%12.5%
Table 6. Percentage improvement of DSSAC-RANSAC over RANSAC, GMS-RANSAC and GMS-ATRANSAC in error variance, mean error, and runtime when processing 2500 ORB feature points.
Table 6. Percentage improvement of DSSAC-RANSAC over RANSAC, GMS-RANSAC and GMS-ATRANSAC in error variance, mean error, and runtime when processing 2500 ORB feature points.
Sequences I m p r v ( e m e a n )
For RAN
(%)
I m p r v ( e m e a n )
For G-R
(%)
I m p r v ( e m e a n )
For G-ATR
(%)
I m p r v ( e V a r )
For RAN
(%)
I m p r v ( e V a r )
For G-R
(%)
I m p r v ( e V a r )
For G-ATR
(%)
I m p r v ( t i m e )
For RAN
(%)
I m p r v ( t i m e )
For G-R
(%)
I m p r v ( t i m e )
For G-ATR
(%)
TUM
(Indoor
dynamic)
56.6%24.9%4.1%76.3%56.1%45.7%72.6%37.4%19.6%
KITTI
(Outdoor dynamic)
58.0%25.6%14.3%77.5%61.4%52.9%71.8%39.6%20.6%
Self-
collected
(Indoor
dynamic)
57.4%26.9%15.0%75.1%57.2%46.2%69.8%35.7%17%
Self-
collected
(Outdoor dynamic)
62.4%37.2%21.5%66.4%38.3%24.0%71.8%40.3%22.2%
AVG.58.6%28.7%13.7%73.853.3%42.2%71.5%38.319.9%
Table 7. The average percentage improvement in error variance, mean error, and runtime for DSSAC-RANSAC compared to RANSAC, GMS-RANSAC, and GMS-ATRANSAC when processing 1500, 2000, and 2500 ORB feature points.
Table 7. The average percentage improvement in error variance, mean error, and runtime for DSSAC-RANSAC compared to RANSAC, GMS-RANSAC, and GMS-ATRANSAC when processing 1500, 2000, and 2500 ORB feature points.
ORB Feature Points I m p r v ( e m e a n )
For RAN
(%)
I m p r v ( e m e a n )
For G-R
(%)
I m p r v ( e m e a n )
For G-ATR
(%)
I m p r v ( e V a r )
For RAN
(%)
I m p r v ( e V a r )
For G-R
(%)
I m p r v ( e V a r )
For G-ATR
(%)
I m p r v ( t i m e )
For RAN
(%)
I m p r v ( t i m e )
For G-R
(%)
I m p r v ( t i m e )
For G-ATR
(%)
150059.3%78.6%22.4%48.1%72.0%63.3%67.8%21.2%8.7%
200057.6%40.2%12.5%73.6%63.7%46.3%68.9%34.9%12.5%
250058.6%28.7%13.7%73.853.3%42.2%71.5%38.3%19.9%
AVG.58.5%49.2%16.2%65.2%63.0%50.6%69.4%31.5%13.7%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yang, Z.; Zhao, K.; Yang, S.; Xiong, Y.; Zhang, C.; Deng, L.; Zhang, D. Research on a Density-Based Clustering Method for Eliminating Inter-Frame Feature Mismatches in Visual SLAM Under Dynamic Scenes. Sensors 2025, 25, 622. https://doi.org/10.3390/s25030622

AMA Style

Yang Z, Zhao K, Yang S, Xiong Y, Zhang C, Deng L, Zhang D. Research on a Density-Based Clustering Method for Eliminating Inter-Frame Feature Mismatches in Visual SLAM Under Dynamic Scenes. Sensors. 2025; 25(3):622. https://doi.org/10.3390/s25030622

Chicago/Turabian Style

Yang, Zhiyong, Kun Zhao, Shengze Yang, Yuhong Xiong, Changjin Zhang, Lielei Deng, and Daode Zhang. 2025. "Research on a Density-Based Clustering Method for Eliminating Inter-Frame Feature Mismatches in Visual SLAM Under Dynamic Scenes" Sensors 25, no. 3: 622. https://doi.org/10.3390/s25030622

APA Style

Yang, Z., Zhao, K., Yang, S., Xiong, Y., Zhang, C., Deng, L., & Zhang, D. (2025). Research on a Density-Based Clustering Method for Eliminating Inter-Frame Feature Mismatches in Visual SLAM Under Dynamic Scenes. Sensors, 25(3), 622. https://doi.org/10.3390/s25030622

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop