Next Article in Journal
Long-Term Ground Deformation Monitoring and Quantitative Interpretation in Shanghai Using Multi-Platform TS-InSAR, PCA, and K-Means Clustering
Previous Article in Journal
DC-Mamba: A Novel Network for Enhanced Remote Sensing Change Detection in Difficult Cases
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Global Semantic Localization from Abstract Ellipse-Ellipsoid Model and Object-Level Instance Topology

State Key Laboratory of Robotics and Systems (HIT), School of Mechatronics Engineering, Harbin Institute of Technology, Harbin 150001, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2024, 16(22), 4187; https://doi.org/10.3390/rs16224187
Submission received: 9 September 2024 / Revised: 4 November 2024 / Accepted: 8 November 2024 / Published: 10 November 2024

Abstract

:
Robust and highly accurate localization using a camera is a challenging task when appearance varies significantly. In indoor environments, changes in illumination and object occlusion can have a significant impact on visual localization. In this paper, we propose a visual localization method based on an ellipse-ellipsoid model, combined with object-level instance topology and alignment. First, we develop a CNN-based (Convolutional Neural Network) ellipse prediction network, DEllipse-Net, which integrates depth information with RGB data to estimate the projection of ellipsoids onto images. Second, we model environments using 3D (Three-dimensional) ellipsoids, instance topology, and ellipsoid descriptors. Finally, the detected ellipses are aligned with the ellipsoids in the environment through semantic object association, and 6-DoF (Degree of Freedom) pose estimation is performed using the ellipse-ellipsoid model. In the bounding box noise experiment, DEllipse-Net demonstrates higher robustness compared to other methods, achieving the highest prediction accuracy for 11 out of 23 objects in ellipse prediction. In the localization test with 15 pixels of noise, we achieve A T E (Absolute Translation Error) and A R E (Absolute Rotation Error) of 0.077 m and 2 . 70 in the f r 2 _ d e s k sequence. Additionally, DEllipse-Net is lightweight and highly portable, with a model size of only 18.6 MB, and a single model can handle all objects. In the object-level instance topology and alignment experiment, our topology and alignment methods significantly enhance the global localization accuracy of the ellipse-ellipsoid model. In experiments involving lighting changes and occlusions, our method achieves more robust global localization compared to the classical bag-of-words based localization method and other ellipse-ellipsoid localization methods.

1. Introduction

Robustly and accurately estimating a robot’s 6-DoF pose in complex environments from a single image is a critical capability for visual SLAM (Simultaneous Localization and Mapping) and navigation [1,2,3]. However, seasonal changes, weather variations, differences between natural and artificial lighting, and continuous alterations in both static and dynamic objects can result in substantial discrepancies between the robot’s observed appearance and the previously constructed map [4,5]. These discrepancies present significant challenges for robot localization tasks.
The localization component of traditional visual SLAM models the environment as a set of keyframe databases during the mapping process and matches the query image with keyframes through local feature extraction and comparison [6,7]. When the appearance difference between the query image and keyframes is small, it can achieve high localization accuracy. However, changes in illumination or object occlusion in the environment can significantly alter the appearance of images captured in the same scene, which may negatively affect the accuracy and robustness of localization [8,9]. Real-world environments contain many semantic objects whose 3D structure and semantic information remain time-invariant under certain conditions [10]. For example, a television in a room retains the same structure and semantics regardless of time, and its spatial location tends to remain stable. Robots can recognize such objects in the environment using deep learning models like Faster R-CNN [11], or YOLO [12,13], which creates opportunities for robots to localize themselves, much like humans, by relying on these reliable semantic signposts. For instance, Li et al. [14] modeled objects as cuboids and used Faster R-CNN [11] to examine the observed bounding box of an object, aiming to construct a model for pose estimation based on the projection of the cuboid and the detected bounding box. However, the cuboid projection is not rectangular, requiring numerous assumptions for accurate pose estimation.
Abstracting objects in the environment as ellipsoids provides a more efficient solution. When a 3D ellipsoid is projected onto a 2D (Two-dimensional) image using the pinhole model, its center is projected to a pixel in the image. If the camera observes more than two ellipsoids, the center-pixel pairs form a classic PnP (Perspective-n-Point) problem. When the number of observed ellipsoids is exactly two, Vincent et al. [15] can approximately localize them based on their projections and rotation parameters. In the image, the projection of an ellipsoid appears as a quadratic curve (ellipse), and the pixel corresponding to the center of this curve represents the pixel projected from the ellipsoid’s center. Therefore, the accuracy of ellipse prediction in 2D images is directly related to the overall localization precision. The above may only be achieved when ellipsoids and ellipses are accurately matched. YOLO et al. [12,13] can detect object bounding boxes and categories, where the bounding boxes can approximate the projections of ellipsoids, and the object categories can be matched with ellipsoids in the environment. However, in practical applications, noise in the bounding boxes and the presence of multiple objects from the same category may affect localization accuracy.
In this paper, we propose a pipeline based on ellipse-ellipsoid pairs for global localization. First, we designed an object-level instance topology and alignment method that accurately matches detected targets with objects. Second, we developed a robust ellipse prediction model, DEllipse-Net, which uses segmented RGB-D blocks to reliably estimate the projection of ellipsoids. Finally, we use the aligned ellipse-ellipsoid pairs to achieve precise global localization. Our main contributions are:
  • We combine the semantic, 3D coordinates and topology information of objects to propose an object-level instance topology and alignment method, which can accurately align ellipses and ellipsoids and effectively improve the localization accuracy.
  • We propose an ellipse prediction network (DEllipse Net) capable of accurately predicting the projections of 3D ellipsoids on images, even in the presence of bounding box noise. Additionally, this network is lightweight, easily portable, and can be adapted to multiple objects with a single model.
  • We propose a global semantic localization system based on an ellipse-ellipsoid model and object-level instance topology, and evaluate its robustness to significant changes in appearance (light changes and object occlusion).
Our global localization system is shown in Figure 1 and contains the following parts. Object detection: Targets of interest in the query image are detected using YoloV8, and the RGB-D image are cropped according to the bounding box of the target. Ellipse prediction: Taking the cropped RGB-D image pairs as input, the projection (ellipse) of the object is estimated using DEllipse-Net. Object-level instance topology: The objects (ellipses) detected in the query image are subjected to 3D instance topology and the descriptors of each ellipse are extracted. Global map: Consists of a set of ellipsoids with category, size, and 3D coordinates representing different objects in the environment. The ellipsoids are associated through an object-level instance topology. Pose estimation: Position estimation based on matched ellipse-ellipsoid pairs.

2. Related Work

Visual global localization is a widespread concern in visual SLAM. Based on the appearance, we categorize them into feature-based and semantic-based.

2.1. Global Localization Based on Feature Point

Global localization using BoW(bag-of-words) [16] method with points as features is the most common approach. FAB-MAP [17] uses BoW [16] and Chow-Liu trees [18] to infer the correspondence between the feature points in the query image and the 3D map points, and to perform camera pose estimation. DBoW2 [19] constructs an image database based on inverse-order file structure using ORB (Oriented FAST and Rotated BRIEF), SIFT (Scale Invariant Feature Transform), etc. as descriptors for indexing images and fast querying. Visual SLAMs such as ORB-SLAM [20,21,22] and Vins-mono [23] use DBoW2 for relocation and loop-closing. However, this approach relies on feature matching and is not robust enough when there are strong changes in appearance.
To better cope with changes in appearance, recent research has turned to the use of more advanced visual features. HF-Net [24] proposes a hierarchical localization method based on CNN to predict local and global features. Following a coarse-to-fine localization paradigm, global retrieval is first performed to obtain a set of database images, and then these maps are clustered into positions using the covariance of the 3D-SfM (Structure from Motion) model, and 2D-3D matching of local features at these candidate positions is performed to obtain the camera’s accurate 6-DoF position. LoFTR [25] proposes a new method for local image feature matching, using the global receptive field provided by Transformer [26] makes it possible to produce dense matches in low texture regions based on the feature descriptors of the two images. The global receptive field provided by the Transformer allows LoFTR to produce dense matches in less textured regions where it is often difficult for the feature detector to produce usable points of interest. VS-Net [27] proposes a voting network that uses segmentation to partition pixels into distinct landmark blocks and uses landmark location voting branches to estimate landmark locations within each block. The framework establishes a correspondence between a query image and a specific scene with a learnable set of landmarks by landmark matching to predict accurate bit positions.

2.2. Global Localization Based on Semantic Object

Semantic feature based localization is mainly dependent on the development of image detection techniques. [28] utilizes object dense semantics and environment understanding for global localization. The environment with three-bit dense semantics, semantic graph and its topology is first modeled. Then, semantic objects are associated using random walks to use this object-level representation for position recognition, followed by 6-DoF pose estimation via semantic-level point alignment. [29] modeled object-level features in the environment using voxels and rectangles to represent the environment as a semantic graph with topological information. Based on this, an effective edit-distance based graph matching method is proposed to realize the alignment of the detection target and the objects in the environment for cyclic correction. Although this method has good localization ability and robustness to the environment, it is not able to localize accurately from a single image.
Recent work has proposed a solution for global localization using ellipse-ellipsoid pairs from a single image. Vincent et al. [30] introduced a method to infer camera translation from a pair of ellipsoids when the rotation matrix is known, addressing the issue of failed repositioning under significant appearance changes. This method leverages inertial sensors to provide necessary camera rotation parameters. Later, Vincent et al. improved on [15] by developing a method to restore the full pose through the correspondence between two ellipses. Refs. [15,30] involves constructing error functions based on the projection relationship between 2D ellipses and 3D ellipsoids. Consequently, accurate ellipse prediction becomes crucial for positioning accuracy. In both [15,30], the inscribed ellipse of the YOLO’s bounding box is directly used as the ellipse projection. However, since the bounding box is axis-aligned and may contain errors in both size and position, this reduces localization accuracy. To address this, Zins et al. [31] developed an ellipse estimation network that infers ellipses from the RGB image within the bounding box, overcoming the limitations of directly fitting ellipses to axis-aligned bounding boxes, and demonstrating robustness to changes in bounding boxes. Nevertheless, to ensure accurate ellipse predictions, Zins et al. [31] constructed a separate network model for each object in the environment, which is clearly impractical. Moreover, the above methods did not accurately match ellipses to ellipsoid pairs in the early stages of pose estimation, instead relying on rough IoU-based (Intersection over Union) judgments between the projected and detected ellipses. This approach introduces unnecessary errors when there are multiple objects of the same category present in the environment.
Therefore, to enhance the localization accuracy of ellipsoids, we first designed an object-level topology and alignment method to accurately match detected ellipses with the corresponding ellipsoids in the environment (Section 3). Next, we introduced depth information into the ellipse prediction network to improve target-background separation, thereby enhancing the network’s robustness and accuracy in predicting ellipses (Section 4). Finally, we combine these two approaches to achieve precise and robust global localization.

3. Global Ellipsoid Mapping and Object-Level Instance Topology and Alignment

When applying the ellipsoid model for global localization, matching detected ellipses with their corresponding environmental ellipsoids is critical. Ideally, recognized targets in the image could be matched directly to environmental objects based on their categories. However, it is common for multiple objects of the same category to coexist in the environment. To address this, we not only establish an abstract ellipsoid map for localization, but also design object-level instance topology and alignment methods. These approaches associate environmental objects with image targets, enabling accurate matching.

3.1. Global Ellipsoid Mapping Construction

To solve the global localization problem using the ellipse-ellipsoid model, a global ellipsoid map is required. In this work, the ellipsoid map is constructed once and remains static to support long-term localization. We use ORB-SLAM3 [22] to compute the camera pose and generate a point cloud from the RGB-D data of the keyframes through camera interpolation. Voxblox [32] then uses the point cloud and keyframe poses as input, constructing a dense map using its “simple” mode.
Figure 2a shows a dense map constructed with ORB-SLAM3 [22] and Voxblox [32]. We select objects of interest in the map and segment them. As shown in Figure 2b, we use a suitable bounding box to enclose each object. In Figure 2c, the segmented objects are assigned corresponding semantic information. The bounding box simulates the size, location, and rotation of each object in the environment. Then, we represent each object using the outer ellipsoid of the bounding box and associate it with its semantic label. The ellipsoid O is expressed as:
4 γ · ( x 2 W 2 + y 2 H 2 + z 2 D 2 ) = 1
where, [ W , H , D ] are the three axes of the bounding box. The center and rotation of the ellipsoid are consistent with the bounding box. γ = 0.64 is used to control the size of the ellipsoid to ensure that it can completely wrap around the object.

3.2. Object-Level Instance Topology

When multiple objects of the same type exist in the environment, matching them by category is challenging. Humans often distinguish these objects based on their spatial relationships and surrounding context. Therefore, it is essential not only to use ellipsoids to represent the size and position of objects in the environment, but also to correlate them with nearby objects.
We represent the ellipsoid map with semantic labels introduced in Section 3.1 as nodes, using the ellipsoid’s center as the node’s coordinates in the world coordinate system. As shown in Figure 3, nodes of different colors are used to represent ellipsoids with different semantics, and similar objects may appear many times in the environment. Nodes are connected by undirected edges based on Euclidean distance, forming a topological graph that incorporates both distance and semantic information. To describe the distance relationships between node O i and other nodes, the environment is uniformly divided into multiple spheres with radius r i = r i centered at O i . r is a fixed radius, and r i is the radius of the i _ t h sphere. If the distance d between O i + 1 and O i falls within [ r i , r i + 1 ] , the distance weight of O i + 1 relative to O i is W i = e r i . By centering at different nodes, we calculate the distance weights between each central node and the surrounding nodes. We assume that the closer other nodes are to O i , the stronger their intrinsic connections, and thus, the greater the distance weight.
Then take O i as the root node of the random walk, record the category and distance weight of the nodes passed by the random walk, and use it as a descriptor for the node in the environment. Ellipsoids in the environment are represented as { S k o } M , M is the number of ellipsoids, k is the ID of ellipsoids. Descriptor of each object S k o = ( C k o , W k o ) , where C k o is the category matrix, and W k o is the weight matrix. C k o and W k o are both T × D matrices, T is the step size of random walk and D is the depth of random walk. T and D are set based on the number of ellipsoids in the environment. In this paper, T = 4 and D = 50 .
As shown in Figure 3, starting from the root node O 0 , we perform a random walk with a step size of 4 in the topology graph. In the first step, we record the category 0 and weight w 0 of the root node. In the second step, the walk moves to O 2 , and we record its corresponding category 6 and weight w 2 . In the third step, the walk proceeds to O 4 , where we record category 4 and weight w 4 . In the fourth step, the walk continues to O 3 , recording category 5 and weight w 3 . At this point, we have completed one random walk with a step size of 4. We then repeat the above steps, starting a new walk from the root node and recording the category and weight labels at each step. After completing D walks, we have thoroughly recorded the category and weight descriptors S k o = ( C k o , W k o ) for O 0 within the environment.

3.3. Descriptor Extraction of Query Image and Object Association of Environment

For querying images, we first use predicted ellipses to crop the depth image and then project the cropped image into point cloud segments. The centroid of each point cloud segment is used as the center of the corresponding object. We then apply the method described in Section 3.2 to associate the objects and extract their descriptors. The object in the query image is represented as { S k q } N , where S k q = ( C k q , W k q ) , with N being the number of detected ellipses and k is the ID of each detected ellipse.
When ellipses and ellipsoids belong to different categories, they do not match. If there are multiple ellipsoids of the same type in the environment, we use similarity scores to distinguish them. For query ellipse S k q and ellipsoid in the global map S k o , the similarity is calculated as:
S c o r e ( S k q , S k o ) = i = 1 D j = 1 D C i j · W i j T D 2
C i j [ n ] = 1 if C i q [ n ] = C j o [ n ] 0 if C i q [ n ] C j o [ n ]
W i j [ n ] = e W i q [ n ] W j o [ n ]
where C i o is the category vector in the i _ t h row of C k o , C j q is the category vector in the j _ t h row of C k q , n denotes the position of elements in the vector. C i j is the A N D operation between C i o and C j q on the elements, used for filtering descriptions of the same category. W i o is the distances vector in the i _ t h row of W k o , W j q is the distances vector in the j _ t h row of W k q . W i j is used to calculate the difference in distance at the same position. The higher the score, the higher the similarity.

4. Ellipse Prediction Network and Global Pose Estimation

The key to ellipse-ellipsoid localization lies in predicting the projection of a 3D ellipsoid onto the image, as accurate ellipse prediction can significantly improve localization accuracy. In this section, we propose an ellipse prediction method that combines depth information and employs the methods of Vincent et al. [15] and Zins et al. [31] for global localization.

4.1. Ellipse Estimation Network

Ellipse estimation is a method for predicting the projection of a 3D ellipsoid in an image, which relies on the bounding box from target detection. A approach is to use the inscribed ellipse of the bounding box as the projection of the ellipsoid. However, since the bounding box is aligned with the coordinate axes and contains noise, this can result in errors in the predicted ellipse. Another approach leverages the powerful capability of neural networks to extract intrinsic information from images, accurately predicting ellipses based on differences between the target and the background. In the real world, most objects exhibit geometric structures with convex surfaces. Based on the assumption of convex bodies, targets and backgrounds in the environment can be segmented. Depth information provides the geometric structure of objects, and using it can help distinguish objects within the environment. Therefore, we combine RGB and depth information to construct an ellipse prediction model.
The neural network part of the system is shown in Figure 1, consisting of YOLOv8 [13] for object detection and our ellipse prediction network. First, the camera’s intrinsic parameters are used to calculate the 3D spatial coordinates corresponding to each pixel in the depth image, generating a point cloud of dimensions W × H × 3 . Next, the bounding boxes provided by YOLOv8 are used to crop both the RGB image and the point cloud. Our ellipse prediction network, DEllipse-Net, takes the target’s RGB image and a square subset of the point cloud as inputs, and resizes them to 256 × 256 using interpolation.
Our network, shown in Figure 4, consists of a base convolutional network for feature extraction, followed by MLP (Multi-layer Perceptron), several fully connected layers, and finally three parallel branches predict the ellipse parameters: center, axis and angle. The sizes of the center and axis are normalized relative to the input data dimensions 256 × 256 . To achieve this, the S i g m o i d activation function is applied to constrain both values within the range ( 0 , 1 ) . When calculating the difference between the predicted ellipse and the actual ellipsoid projection, the center and axes are scaled back to their original image dimensions using ( W b b , H b b ) / 256 . Where W b b and H b b are the width and height of the bounding box. We define the right half of the ellipse as the direction of the ellipse, which can be fully described by the sine value. In the network, we use the T a n h function to restrict the direction value to the range ( 1 , 1 ) , with the output representing the positive rotation of the angle. The detailed network structure is shown in Table 1.

4.2. Loss Function

In Section 4.1, we obtained three parameters representing ellipses: center, axis, and angle. Based on these three parameters, the expression for an ellipse is
Ψ ( x ) = ( x c ) T c o s θ s i n θ s i n θ c o s θ 1 a 2 0 0 1 b 2 c o s θ s i n θ s i n θ c o s θ T ( x c ) .
where, x is the point on the ellipse, c is the center coordinate, a and b are the two axes of the ellipse, θ is the angle. But, this representation may not effectively integrate the axis and position components with the azimuth angle due to the discontinuity of the directional angle. Even a slight rotation will lead to significant differences in ellipses that share the same axis and center. To mitigate this effect, we adopt the new ellipse representation proposed in [31]:
Ψ ( x ) = ( x c ) T c o s θ s i n θ s i n θ c o s θ a 0 0 b c o s θ s i n θ s i n θ c o s θ T ( x c ) ,
for ellipse representation. This representation of the ellipse allows for a natural handling of the discontinuity in the angle parameter.
To measure the difference between the predicted ellipse and the true ellipse, we discretize the continuous ellipse as follows:
Ψ ^ ( x ) = D ( x ) if x inside Ψ ( x ) D ( x ) if x outside Ψ ( x ) 0 if x Ψ ( x )
where, D ( x ) is the distance from coordinate x to the ellipse curve. A 25 × 25 sampling grid is then employed to sample the ellipse. The loss of the network is
L o s s = i = 1 N ( Ψ ^ p r e d ( p i ) Ψ ^ g t ( p i ) ) ,
where, p i is the coordinate of the sampling point.

4.3. Pose Estimation from Ellipse to Ellipsoid

The 2D/3D object association list serves as the input for pose estimation. In Section 3.3, we match the ellipse in the query image with the ellipsoid in the environment to obtain a set of corresponding ellipse and ellipsoid pairs.
Ψ k q : { O i o , O j o , }
Ψ k q denotes the ellipse in the image, while O k o represents the ellipsoids in the environment that belong to the same category as Ψ k q , listed in descending order according to their scores.
When there are three or more ellipse-ellipsoid pairs in the environment, the centers of the ellipses and the ellipsoids form corresponding point pairs. At this stage, we employ the P3P (Perspective-3-Point) method to estimate the pose and calculate the projection of each ellipsoid onto the image, as well as the Intersection over Union (IoU) of the ellipse under this pose. We use the IoU to select the optimal pose solution. Unlike [31], which uses an elliptic cone to infer the center, we directly utilize the center of the estimated ellipse and assign an initial depth from a depth image to accelerate convergence.
When there are two ellipse-ellipsoid pairs, we utilize the method from [15] for pose estimation. Assuming that the camera roll is negligible, this method simplifies the 6-DoF problem into a more manageable problem with only one remaining degree of freedom, corresponding to an angular parameter.

5. Experimental and Results

In this section, we evaluate our method using a publicly available RGB-D dataset. Our approach, which incorporates an ellipse-ellipsoid model and object topology alignment, demonstrates robust global localization performance, even under challenging conditions such as lighting variations and object occlusions. All experiments were conducted on a computer equipped with an Intel Core i5-13400K CPU running at 2.5 GHz and a GeForce RTX 4060 Ti GPU. The network was implemented using the PyTorch framework https://pytorch.org/.

5.1. Ellipse Prediction

We used the TUM dataset [33] (the f r 2 _ d e s k sequence) and the 7-Scenes dataset [34] (the C h e s s sequence) to evaluate DEllipse-Net for ellipse prediction. In the f r 2 _ d e s k sequence, we selected 11 categories and 12 objects to construct the ellipsoid environment. A total of 1000 images were selected for training DEllipse-Net and YOLOv8, while 500 images were reserved for testing. In the C h e s s sequence, we selected 7 categories and 11 objects to construct the ellipsoid environment. Six sequences totaling 6000 images were randomly selected, with 3000 images used for training and the remaining 3000 images for testing. During network training, the batch size was set to 16 for 200 epochs, using the Adam optimizer with an initial learning rate of 5 × 10 5 .
To demonstrate the superiority of our proposed ellipse network, we compared it with two other methods: one based on a classical mathematical ellipse (Mathematical) [15], and the other, an ellipse prediction network proposed by Zins et al. (Ellipse-Net) [31]. We evaluated the performance using four metrics: I o U (Intersection over Union), I o U a v g , I o U E r r o r , A T E (Absolute Translation Error), and A R E (Absolute Rotation Error). I o U measures the overlap between the predicted ellipse and the ground truth ellipse, with higher values indicating closer alignment between the predicted and actual ellipsoid projection. I o U a v g is the average I o U of all predicted objects of the same category. I o U E r r o r represents the difference between the predicted I o U and the ideal I o U , with smaller errors indicating better accuracy in predicting the ellipsoid projection. A T E assesses the translation error, with lower values indicating that the estimated translation is closer to the true translation. A R E measures the rotational error, with lower values reflecting smaller discrepancies between the predicted and ground truth rotation.
I o U = Ell g t Ell p r e Ell g t Ell p r e
I o U a v g = i = 1 M I o U i M
I o U e r r o r = 1 Ell g t Ell p r e Ell g t Ell p r e
A T E = ( 1 N i = 1 N T p r e T G T 2 ) 1 2
A R E = ( 1 N i = 1 N 180 π cos 1 ( t r a c e ( R g t 1 R p r e 1 ) 2 ) 2 ) 1 2
where, Ell p r e denotes the predicted ellipse, and Ell g t represents the ground truth ellipse projected from the ellipsoid. T p r e and T g t are the estimated translation vector and the true translation vector, respectively. R p r e and R g t denote the estimated rotation matrix and the true rotation matrix, respectively.
We applied noise of 5 pixels, 10 pixels, and 15 pixels to the bottom left and top right vertices of the bounding box to evaluate the robustness of the three methods for ellipse prediction. Table 2 presents the I o U a v g variation of ellipse predictions and corresponding ellipsoid projections for different objects under bounding box noise across two datasets. The table shows that when the detection box is undisturbed, the I o U a v g values of all three methods achieve good results of 0.9 or above. However, as the noise increases, the I o U a v g values for the three methods decrease. Notably, among the 23 objects in the two datasets, our method did not attain the highest I o U a v g value for 5 of these objects. Additionally, our method did not achieve the highest I o U a v g value in the prediction of 7 objects under partial disturbance. Nevertheless, for the remaining 15 objects, our method produced the best results. This indicates that our approach effectively addresses the inaccuracies in ellipse prediction caused by boundary detection errors. Figure 5 illustrates the performance of the three methods across the two datasets with a noise of 15 pixels. The white ellipse represents the true ellipsoid projection, while the white bounding box indicates the area with added noise. All three methods of ellipse prediction were performed within this region. It is evident that the added disturbance has the greatest impact on the Math method, while Ellipse-Net demonstrates some degree of robustness against interference. Our method, however, maintains relatively strong performance even after noise, for instance, in the f r 2 _ d e s k dataset, our predictions almost completely overlap with the true ellipsoid.
With the enhancement of the ellipsoid prediction’s anti-interference ability, there is a corresponding improvement in localization capability. Table 3 illustrates the impact of three ellipsoid prediction methods on localization performance under various noise. It is important to note that our testing was conducted using object-level topology and alignment, with the variables in this test limited to the three different ellipsoid prediction methods. The table demonstrates that, under different disturbances, the stable ellipsoid predictions generated by DEllipse-Net significantly enhance the robustness of ellipsoid-based localization. In eight localization experiments, we achieved the best results in seven of them.
More specifically, we present the statistical results of the I o U e r r o r for different objects under undisturbed conditions, as shown in Figure 6 and Figure 7, respectively. These line graphs illustrate the percentage of ellipses at various thresholds of I o U e r r o r . In the case of the f r 2 _ d e s k dataset (Figure 6), eight objects achieved optimal results at different thresholds. For instance, the “TV” had all predicted I o U e r r o r below 0.08. When predicting the “Cup”, the threshold for I o U e r r o r was set at 0.1, where our method achieved a 100 % success rate, while Ellipse-Net accounted for 82 % and Mathematical accounted for 80 % . Furthermore, there are notable differences in ellipse prediction accuracy for different objects within the same category. For example, in the prediction of “Cola”, the Mathematical and Ellipse-Net exhibit varying predictive capabilities, while our approach demonstrates relatively high robustness. Similarly, as shown in Figure 7, our method also achieved the best results for C h e s s .
Finally, we compared the model sizes of the ellipses across the three methods and the average prediction time per frame, as shown in Table 4. Since the Mathematical relies on mathematical techniques for ellipse prediction, it does not have a traditional model, resulting in minimal time required for predicting ellipses from image frames. In contrast, the single model size of Ellipse-Net reaches 82.6 MB, and each object requires corresponding model training during actual operation. Consequently, when dealing with a large number of objects in the environment, the space occupied becomes substantial, and it also takes the most time to predict ellipses for a single frame. However, DEllipse-Net only requires the training of one model to handle ellipse predictions for all objects, and our ellipse model is only 18.3 MB in size, with a running time that is shorter than that of Ellipse-Net. This demonstrates that DEllipse-Net is both lightweight and portable.

5.2. Localization Accuracy with Object-Level Instance Topology-Alignment

Localization prediction involves matching the predicted ellipse in the image with the corresponding ellipsoid in the environment. In previous works, such as [15,31], ellipses and ellipsoids are matched by category, where a single ellipse may correspond to the projection of multiple ellipsoids, leading to pose estimation errors. By applying our object-level instance topology to ellipse-ellipsoid localization, we achieve more accurate matching between ellipses and ellipsoids.
We use the C h e s s dataset from 7-Scenes [34] for our experiments. The C h e s s dataset contains 6 sequences, each with 1000 images. Since our method requires the detection of two or more ellipse-ellipsoid pairs for pose estimation, we use this criterion for image filtering. The primary aim of this experiment is to evaluate the impact of our instance topology on pose estimation. Therefore, we conducted localization tests using three different ellipse estimation models under two different ellipse-ellipsoid matching modes. A T E and A R E are used as metrics to assess localization accuracy.
As shown in Table 5, “×” indicates that instance topology is not used, while “√” indicates the use of our instance topology. In the “×” mode, Ours effectively reduces both translation and rotation errors. In the translation test, our method achieved the lowest A T E in four sequences, while it ranked second in seq-04 and seq-06. In the rotation test, our method produced suboptimal results in seq-01 and seq-02, but achieved the lowest A R E in the remaining sequences. In the “√” mode, our method demonstrated similar performance.
Compared with the “×” mode, after using our instance topology, the translation and rotation errors of the camera are greatly reduced. In seq-04 sequence, Mathematical reduces the translation error from 0.208 m to 0.036 m and the corresponding rotation error from 8 . 88 to 1 . 03 after using descriptor matching. The translation error of Ellipse-Net decreased from 0.164 m to 0.037 m, and the rotation error decreased from 7 . 62 to 1 . 09 . The translation error of our method decreased from 0.165 m to 0.032 m, and the rotation error decreased from 7 . 31 to 0 . 95 . This shows that our instance topology greatly improves the probability of object matching, thereby improving the localization accuracy.

5.3. Pose Accuracy in Different Lighting Environments

When performing localization in indoor environments, the accuracy is affected by various factors due to the time gap between map creation and subsequent localization. Among these factors, lighting variation has the most significant impact on visual localization. In this section, we compare a feature-based method, OVS (OpenVSLAM) [35], with two Ellipse-Ellipsoid-based methods (Ellipse-Net and ours) to evaluate the localization performance of these three approaches under different lighting conditions. We use the f r 2 _ d e s k , f r 3 _ o f f i c e _ h o u s e h o l d , C h e s s , and R e d k i t c h e n sequences as test data and adjust the HSV values in the images to simulate low-light conditions, as shown in Figure 8. Under normal lighting conditions, we use the RGB-D model of the full OVS system for map creation, but only its relocalization module during the localization phase. Ellipse-Net follows the approach provided by [31], while our method is based on DEllipse-Net and object-level instances topology-alignment localization. A T E and A R E are used as metrics to evaluate localization accuracy.
As shown in Table 6, the localization accuracy of all three methods decreases in dark environments compared to normal lighting. In the f r 2 _ d e s k sequence, under normal lighting, OVS has translation and rotation errors of 0.102 m and 8 . 73 , Ellipse-Net has errors of 0.024 m and 0 . 89 , and our method achieves 0.012 m and 0 . 56 . In dark conditions, OVS’s errors rise to 0.119 m and 6 . 39 , Ellipse-Net’s to 0.065 m and 2 . 44 , while our method shows only 0.022 m and 0 . 89 . Overall, our method consistently outperforms OVS and Ellipse-Net across all lighting conditions, with similar advantages in the other three environments.
Despite both Ellipse-Net and our approach being based on the Ellipse-Ellipsoid model, our method is more robust to lighting variations. Across all four sequences, Ellipse-Net shows significant localization accuracy changes under different lighting conditions. For example, in the f r 3 _ o f f i c e _ h o u s e h o l d sequence, its translation and rotation errors increase by 0.039 m and 0 . 84 in dark conditions, while OVS’s increase by 0.023 m and 0 . 45 . In contrast, our method sees only minimal increases of 0.009 m and 0 . 19 , demonstrating its robustness against lighting variations.

5.4. Pose Accuracy When Objects Are Occluded

This experiment tests the localization ability of classical point-based methods and ellipse-ellipsoid methods under object occlusion. In four sequences, to simulate occlusion, we applied a black mask to occlude corresponding objects in the RGB images. We sequentially occluded 1, 2, and 3 objects in the query images to evaluate how both methods handle object occlusion, as shown in Figure 9. To ensure the ellipse-ellipsoid model can still predict poses, we selected query images containing five or more objects from the sequences. We used images without occlusions for map construction and then performed localization testing using the occluded images. The localization performance of both methods under occlusion was evaluated using A T E and A R E .
As shown in Table 7, in the f r 2 _ d e s k sequence, when the number of occluded objects increased from one to two, OVS’s translation error rose slightly from 0.238 m to 0.241 m, and its rotation error from 17 . 0 to 17 . 08 . When the number of occlusions increased from two to three, the translation error increased to 0.260 m, and the rotation error to 18 . 19 . Ellipse-Net showed a similar trend, with translation errors increasing by 0.002 m and 0.016 m, and rotation errors by 0 . 01 and 0 . 29 . In contrast, our method showed no increase in errors during the first stage and only a small rise of 0.017 m in translation and 0 . 49 in rotation errors during the second stage. These results suggest that localization accuracy decreases with more occlusions for OVS, Ellipse-Net, and our method. This trend was also observed in the f r 3 _ o f f i c e _ h o u s e h o l d , R e d k i t c h e n , and C h e s s sequences.
The Ellipse-Ellipsoid model relies on object detection for ellipse prediction, matching, and global localization. When landmarks are occluded, both Ellipse-Net and our method fail to detect them, leading to localization failure. In contrast, OVS uses ORB feature points, matched with keyframes via BoW [16] and applies PnP for pose estimation, making it less sensitive to occlusion after successful matching. During testing, we used black ellipses to occlude objects, preventing feature detection in darkened areas and reducing incorrect matches. Although our method is more affected by occlusions, it still outperforms OVS, demonstrating superior localization accuracy under occlusion conditions.

6. Conclusions

We propose an object-based global localization method that estimates the 6-DoF camera pose using an Ellipse-Ellipsoid model and object-level instance topology alignment. First, we designed a neural network that predicts ellipses using RGB and depth images. Second, the instance topology and alignment model allows for successful matching between detected ellipses and those in the environment. Finally, we achieve accurate and robust localization in a 3D ellipsoid environment that we constructed. We evaluated our method on publicly available and pre-processed datasets. Compared to existing approaches, our method can robustly predict ellipses in images and, combined with object-level topology alignment, maintains high localization accuracy even in environments with significant changes in lighting and occlusion.
However, our method has some limitations. For instance, we have not conducted experiments in real-world environments for lighting and occlusion, and the ellipsoid map still requires manual annotation. In future work, when experimental facilities are fully equipped, we plan to conduct tests in real-world settings and further optimize our algorithm. Additionally, we are working on the construction of instance and ellipsoid maps, aiming to simultaneously build both instance and ellipsoid maps.

Author Contributions

Methodology, H.W.; Project administration, Y.L.; Software, H.W.; Supervision, Y.L.; Writing & original draft, H.W.; Writing—review & editing, C.W. and Y.W. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by Key Special Projects of Heilongjiang Province’s Key R&D Program (No. 2023ZX01A01) and Heilongjiang Province’s Key R&D Program: ‘Leading the Charge with Open Competition’ (No. 2023ZXJ01A02).

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Chen, K.; Zhang, J.; Liu, J.; Tong, Q.; Liu, R.; Chen, S. Semantic Visual Simultaneous Localization and Mapping: A Survey. arXiv 2022, arXiv:2209.06428. [Google Scholar] [CrossRef]
  2. Yin, Z.; Wen, H.; Nie, W.; Zhou, M. Localization of Mobile Robots Based on Depth Camera. Remote Sens. 2023, 15, 4016. [Google Scholar] [CrossRef]
  3. Huang, Y.; Xie, F.; Zhao, J.; Gao, Z.; Chen, J.; Zhao, F.; Liu, X. ULG-SLAM: A Novel Unsupervised Learning and Geometric Feature-Based Visual SLAM Algorithm for Robot Localizability Estimation. Remote Sens. 2024, 16, 1968. [Google Scholar] [CrossRef]
  4. Wang, Y.; Jiang, C.; Chen, X. GOReloc: Graph-Based Object-Level Relocalization for Visual SLAM. IEEE Robot. Autom. Lett. 2024, 9, 8234–8241. [Google Scholar] [CrossRef]
  5. Zhao, X.; Li, Q.; Wang, C.; Dou, H.; Liu, B. Robust Depth-Aided RGBD-Inertial Odometry for Indoor Localization. Measurement 2023, 209, 112487. [Google Scholar] [CrossRef]
  6. Rosinol, A.; Abate, M.; Chang, Y.; Carlone, L. Kimera: An Open-Source Library for Real-Time Metric-Semantic Localization and Mapping. In Proceedings of the 2020 IEEE International Conference on Robotics and Automation (ICRA), Paris, France, 31 May–31 August 2020; pp. 1689–1696. [Google Scholar]
  7. Abate, M.; Chang, Y.; Hughes, N.; Carlone, L. Kimera2: Robust and Accurate Metric-Semantic SLAM in the Real World. In Springer Proceedings in Advanced Robotics; Springer: Cham, Switzerland, 2024. [Google Scholar] [CrossRef]
  8. Li, M.; Ma, Y.; Qiu, Q. SemanticSLAM: Learning based Semantic Map Construction and Robust Camera Localization. In Proceedings of the 2023 IEEE Symposium Series on Computational Intelligence (SSCI), Mexico City, Mexico, 5–8 December 2023; pp. 312–317. [Google Scholar] [CrossRef]
  9. Adkins, A.; Chen, T.; Biswas, J. ObVi-SLAM: Long-Term Object-Visual SLAM. IEEE Robot. Autom. Lett. 2024, 9, 2909–2916. [Google Scholar] [CrossRef]
  10. Guo, X.; Hu, J.; Chen, J.; Deng, F.; Lam, T.L. Semantic Histogram Based Graph Matching for Real-Time Multi-Robot Global Localization in Large Scale Environment. IEEE Robot. Autom. Lett. 2021, 6, 1–8. [Google Scholar] [CrossRef]
  11. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef] [PubMed]
  12. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You Only Look Once: Unified, Real-Time Object Detection. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar] [CrossRef]
  13. Jocher, G. Ultralytics YOLO. arXiv e-prints 2023. Available online: https://github.com/ultralytics/ultralytics (accessed on 22 February 2024).
  14. Li, J.; Meger, D.; Dudek, G. Context-coherent scenes of objects for camera pose estimation. In Proceedings of the 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Vancouver, BC, Canada, 24–28 September 2017. [Google Scholar]
  15. Gaudillière, V.; Simon, G.; Berger, M.-O. Perspective-2-Ellipsoid: Bridging the Gap Between Object Detections and 6-DoF Camera Pose. IEEE Robot. Autom. Lett. 2020, 5, 5189–5196. [Google Scholar] [CrossRef]
  16. Mccallum, A.K. Bow: A toolkit for statistical language modeling. In Proceedings of the ICASSP, Atlanta, GA, USA, 7–10 May 1996. [Google Scholar]
  17. Cummins, M.; Newman, P. FAB-MAP: Probabilistic Localization and Mapping in the Space of Appearance. Int. J. Robot. Res. 2008, 27, 647–665. [Google Scholar] [CrossRef]
  18. Chow, C.; Liu, C. Approximating discrete probability distributions with dependence trees. IEEE Trans. Inf. Theory 1968, 14, 462–467. [Google Scholar] [CrossRef]
  19. Galvez-Lopez, D.; Tardos, J.D. Bags of Binary Words for Fast Place Recognition in Image Sequences. IEEE Trans. Robot. 2012, 28, 1188–1197. [Google Scholar] [CrossRef]
  20. Mur-Artal, R.; Tardos, J.D. ORB-SLAM2: An Open-Source SLAM System for Monocular, Stereo, and RGB-D Cameras. IEEE Trans. Robot. 2017, 33, 1255–1262. [Google Scholar] [CrossRef]
  21. Mur-Artal, R.; Montiel, J.M.M.; Tardos, J.D. ORB-SLAM: A Versatile and Accurate Monocular SLAM System. IEEE Trans. Robot. 2015, 31, 1147–1163. [Google Scholar] [CrossRef]
  22. Campos, C.; Elvira, R.; Gomez Rodriguez, J.J.; Montiel, J.M.M.; Tardos, J.D. ORB-SLAM3: An Accurate Open-Source Library for Visual, Visual-Inertial, and Multimap SLAM. IEEE Trans. Robot. 2021, 37, 1874–1890. [Google Scholar] [CrossRef]
  23. Qin, T.; Li, P.; Shen, S. VINS-Mono: A Robust and Versatile Monocular Visual-Inertial State Estimator. IEEE Trans. Robot. 2018, 34, 1004–1020. [Google Scholar] [CrossRef]
  24. Sarlin, P.-E.; Cadena, C.; Siegwart, R.; Dymczyk, M. From Coarse to Fine: Robust Hierarchical Localization at Large Scale. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 16–17 June 2019; pp. 12708–12717. [Google Scholar] [CrossRef]
  25. Sun, J.; Shen, Z.; Wang, Y.; Bao, H.; Zhou, X. LoFTR: Detector-Free Local Feature Matching with Transformers. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 8918–8927. [Google Scholar] [CrossRef]
  26. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention Is All You Need. arXiv 2017, arXiv:1706.03762. [Google Scholar]
  27. Huang, Z.; Zhou, H.; Li, Y.; Yang, B.; Xu, Y.; Zhou, X.; Bao, H.; Zhang, G.; Li, H. VS-Net: Voting with Segmentation for Visual Localization. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 6097–6107. [Google Scholar] [CrossRef]
  28. Liu, Y.; Petillot, Y.; Lane, D.; Wang, S. Global Localization with Object-Level Semantics and Topology. In Proceedings of the 2019 International Conference on Robotics and Automation (ICRA), Montreal, QC, Canada, 20–24 May 2019; pp. 4909–4915. [Google Scholar] [CrossRef]
  29. Lin, S.; Wang, J.; Xu, M.; Zhao, H.; Chen, Z. Topology Aware Object-Level Semantic Mapping Towards More Robust Loop Closure. IEEE Robot. Autom. Lett. 2021, 6, 7041–7048. [Google Scholar] [CrossRef]
  30. Gaudillière, V.; Simon, G.; Berger, M.-O. Camera Relocalization with Ellipsoidal Abstraction of Objects. In Proceedings of the 2019 IEEE International Symposium on Mixed and Augmented Reality (ISMAR), Beijing, China, 14–18 October 2019; pp. 8–18. [Google Scholar] [CrossRef]
  31. Zins, M.; Simon, G.; Berger, M.O. Object-Based Visual Camera Pose Estimation From Ellipsoidal Model and 3D-Aware Ellipse Prediction. Int. J. Comput. Vis. 2022, 130, 1107–1126. [Google Scholar] [CrossRef]
  32. Oleynikova, H.; Taylor, Z.; Fehr, M.; Siegwart, R.; Nieto, J. Voxblox: Incremental 3D Euclidean Signed Distance Fields for on-board MAV planning. In Proceedings of the 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Vancouver, BC, Canada, 24–28 September 2017; pp. 1366–1373. [Google Scholar] [CrossRef]
  33. Sturm, J.; Engelhard, N.; Endres, F.; Burgard, W.; Cremers, D. A Benchmark for the Evaluation of RGB-D SLAM Systems. In Proceedings of the 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Vilamoura-Algarve, Portugal, 7–12 October 2012. [Google Scholar]
  34. Glocker, B.; Izadi, S.; Shotton, J.; Criminisi, A. Real-time RGB-D camera relocalization. In Proceedings of the 2013 IEEE International Symposium on Mixed and Augmented Reality (ISMAR), Adelaide, SA, Australia, 1–4 October 2013; pp. 173–179. [Google Scholar] [CrossRef]
  35. Sumikura, S.; Shibuya, M.; Sakurada, K. OpenVSLAM: A Versatile Visual SLAM Framework. In Proceedings of the 27th ACM International Conference on Multimedia, Nice, France, 21–25 October 2019. [Google Scholar]
Figure 1. Overview of global semantic localization. Before using a single-frame RGB-D image for localization, it is essential to construct a “Global Map” that includes the 3D ellipsoid abstraction and the topological relationships of each object instance. Upon receiving each RGB-D image as input, we use YOLOv8 [13] to detect the semantic information and bounding boxes of each object within the RGB image. We then crop the corresponding object regions from both the RGB and depth images for ellipse prediction using DEllipse-Net. Afterward, the ellipse and semantic information are combined to construct a topology map, which is aligned with the “Global Map”. Finally, the aligned ellipse-ellipsoid pairs are used for pose estimation.
Figure 1. Overview of global semantic localization. Before using a single-frame RGB-D image for localization, it is essential to construct a “Global Map” that includes the 3D ellipsoid abstraction and the topological relationships of each object instance. Upon receiving each RGB-D image as input, we use YOLOv8 [13] to detect the semantic information and bounding boxes of each object within the RGB image. We then crop the corresponding object regions from both the RGB and depth images for ellipse prediction using DEllipse-Net. Afterward, the ellipse and semantic information are combined to construct a topology map, which is aligned with the “Global Map”. Finally, the aligned ellipse-ellipsoid pairs are used for pose estimation.
Remotesensing 16 04187 g001
Figure 2. The construction process of ellipsoid map. (a) is the dense map constructed by ORB-SLAM3 and Voxblox. (b) is the semantic map of objects. (c) is the ellipsoid map of objects.
Figure 2. The construction process of ellipsoid map. (a) is the dense map constructed by ORB-SLAM3 and Voxblox. (b) is the semantic map of objects. (c) is the ellipsoid map of objects.
Remotesensing 16 04187 g002
Figure 3. Object-level instance topology and descriptor extraction. Top left: undirected topology diagram between nodes and distance weights and random walks diagram with O 0 as the root node. Top right: List of distance weights for each node. Bottom: Category and distance weights obtained through random walks. Nodes of different colors represent objects of different categories.
Figure 3. Object-level instance topology and descriptor extraction. Top left: undirected topology diagram between nodes and distance weights and random walks diagram with O 0 as the root node. Top right: List of distance weights for each node. Bottom: Category and distance weights obtained through random walks. Nodes of different colors represent objects of different categories.
Remotesensing 16 04187 g003
Figure 4. Overview of the DEllipse-Net. “MP” is maximum pooling, “AP” is average pooling, “RL” is Relu.
Figure 4. Overview of the DEllipse-Net. “MP” is maximum pooling, “AP” is average pooling, “RL” is Relu.
Remotesensing 16 04187 g004
Figure 5. Different methods for ellipse prediction in bounding box with 15 pixel noise. The top is the predicted result of an image in the f r 2 _ d e s k , and the bottom is the predicted result of an image in the C h e s s . (ac) are comparisons between the three methods (Mathematical, Ellipse-Net and DEllipse-Net (Ours)) and ground truth, respectively. The white rectangle is the bounding box after adding 15 pixels of noise. “White ellipse” is the ground truth, “Blue ellipse” is Mathematical, “Red ellipse” is Ellipse-Net, and “Green ellipse” is DEllipse-Net.
Figure 5. Different methods for ellipse prediction in bounding box with 15 pixel noise. The top is the predicted result of an image in the f r 2 _ d e s k , and the bottom is the predicted result of an image in the C h e s s . (ac) are comparisons between the three methods (Mathematical, Ellipse-Net and DEllipse-Net (Ours)) and ground truth, respectively. The white rectangle is the bounding box after adding 15 pixels of noise. “White ellipse” is the ground truth, “Blue ellipse” is Mathematical, “Red ellipse” is Ellipse-Net, and “Green ellipse” is DEllipse-Net.
Remotesensing 16 04187 g005
Figure 6. The percentage of images in f r 2 _ d e s k with different I o U e r r o r thresholds as a proportion of the total number of images.
Figure 6. The percentage of images in f r 2 _ d e s k with different I o U e r r o r thresholds as a proportion of the total number of images.
Remotesensing 16 04187 g006
Figure 7. The percentage of images in C h e s s with different I o U e r r o r thresholds as a proportion of the total number of images.
Figure 7. The percentage of images in C h e s s with different I o U e r r o r thresholds as a proportion of the total number of images.
Remotesensing 16 04187 g007
Figure 8. A comparison of images from the f r 2 _ d e s k , f r 3 _ o f f i c e _ h o u s e h o l d , C h e s s , and R e d k i t c h e n sequences under normal and dark conditions.
Figure 8. A comparison of images from the f r 2 _ d e s k , f r 3 _ o f f i c e _ h o u s e h o l d , C h e s s , and R e d k i t c h e n sequences under normal and dark conditions.
Remotesensing 16 04187 g008
Figure 9. The image performance of the f r 2 _ d e s k , f r 3 _ o f f i c e _ h o u s e h o l d , C h e s s , and R e d k i t c h e n sequences with one, two, and three occlusions.
Figure 9. The image performance of the f r 2 _ d e s k , f r 3 _ o f f i c e _ h o u s e h o l d , C h e s s , and R e d k i t c h e n sequences with one, two, and three occlusions.
Remotesensing 16 04187 g009
Table 1. Detailed network parameters of DEllipse-Net.
Table 1. Detailed network parameters of DEllipse-Net.
LaryInputOutputSizeStridePaddingActivation
FeatureConv2d 6 × 256 × 256 64 × 256 × 256 311Relu
Conv2d 64 × 256 × 256 64 × 256 × 256 311Relu
MaxPool2d 64 × 256 × 256 64 × 256 × 256 220-
Conv2d 64 × 256 × 256 128 × 256 × 256 311Relu
MaxPool2d 128 × 256 × 256 128 × 256 × 256 220-
Conv2d 128 × 256 × 256 256 × 256 × 256 311Relu
MaxPool2d 256 × 256 × 256 256 × 256 × 256 220-
Conv2d 256 × 256 × 256 512 × 256 × 256 311Relu
MaxPool2d 512 × 256 × 256 512 × 256 × 256 220-
Conv2d 512 × 256 × 256 512 × 256 × 256 311Relu
AdaptiveAvgPool2d 512 × 256 × 256 512 × 2 × 2 ----
MLPLinear2048256----
BatchNorm1d256256----
Relu256256----
Linear256256----
BatchNorm1d256256----
Relu256256----
Linear25664----
BatchNorm1d6464----
Relu6464----
AxisLinear6432----
BatchNorm1d3232----
Relu3232----
Linear322----
Sigmoid22----
CenterLinear6432----
BatchNorm1d3232----
Relu3232----
Linear322----
Sigmoid22----
AngleLinear6432----
BatchNorm1d3232----
Relu3232----
Linear321----
Tanh11----
Table 2. The I o U a v g values of each objects in C h e s s and f r 2 _ d e s k under different bounding box noise. “Red”, “Blue”, “Purple”, and “Orange” represent the best values under the noise of 0 pixel, 5 pixel, 10 pixel, and 15 pixel noise, respectively.
Table 2. The I o U a v g values of each objects in C h e s s and f r 2 _ d e s k under different bounding box noise. “Red”, “Blue”, “Purple”, and “Orange” represent the best values under the noise of 0 pixel, 5 pixel, 10 pixel, and 15 pixel noise, respectively.
DatasetObjectMathematicalEllipse-NetOurs
0 pix5 pix10 pix15 pix0 pix5 pix10 pix15 pix0 pix5 pix10 pix15 pix
f r 2 _ d e s k TV0.9600.8540.8210.7870.9650.9030.8870.8680.9790.9120.8960.878
Book0.9320.7460.6960.6570.9430.8500.8220.7960.9580.8560.8270.805
Mouse0.8130.5400.4690.4040.9020.7010.6340.5580.8960.6600.5920.518
Keyboard0.7850.7100.6820.6470.9360.8920.8780.8620.9300.8830.8650.840
Cup0.9160.6480.6550.5110.9290.7610.7110.6510.9660.7820.7900.674
Tape0.9440.7140.7400.6120.9400.8190.8350.7380.9430.8290.8460.752
Telephone0.9320.7810.5030.7060.9380.8620.6330.8060.9550.8690.7520.823
Cola-00.7710.5650.5210.4320.8840.6780.6030.5730.9000.7870.7680.700
Cola-10.9400.8890.8560.8160.8940.8450.8130.7850.9700.9210.8950.865
Bowl0.9410.9060.8950.8510.9780.9350.8860.8320.9670.9280.8650.816
Dice0.8400.8150.7860.7210.9250.8910.7630.7260.9710.8990.8430.816
Teddy beer0.8960.7980.7680.6650.8930.7920.7730.7120.9510.8840.8430.752
C h e s s TV-00.9540.9430.9140.8900.9710.9220.9130.9040.9760.9340.9240.915
TV-10.9530.9390.9190.8950.9370.9250.9140.8990.9810.9380.9270.918
Xbox-00.8670.8150.7720.7270.9300.8980.8810.8580.9570.8930.8780.859
Xbox-10.8390.8350.7810.7320.9390.8780.8630.8400.9600.8870.8710.851
Chair-00.9020.8790.8400.8020.9550.9110.8990.8820.9760.9220.9080.890
Chair-10.8450.8100.7930.7740.8790.8630.8160.7860.9580.8460.8380.827
Chair-20.8840.8820.8580.8320.9480.8730.8660.8550.8950.8630.8530.842
Timer0.9330.8390.7890.7420.9580.8780.8510.8170.9350.8670.8370.802
Games0.9130.8230.7710.7230.9270.8790.8460.8090.9350.8680.8380.802
Interruptor0.9340.6950.6000.5310.9410.7780.6950.6260.9470.7850.6970.625
Gamepad0.9130.7560.6770.6100.9240.8340.7430.6840.9540.8040.7710.708
Table 3. A T E ( m ) and A R E ( ) of three ellipse prediction models under different bounding box noises. “Red” and “Blue” represent the best values of A T E and A R E , respectively.
Table 3. A T E ( m ) and A R E ( ) of three ellipse prediction models under different bounding box noises. “Red” and “Blue” represent the best values of A T E and A R E , respectively.
ModelErrors fr 2 _ desk Chess
0 pix5 pix10 pix15 pix0 pix5 pix10 pix15 pix
Mathematical [15] A T E ( m ) 0.0370.0530.1000.1900.0780.0870.1620.303
A R E ( ) 1.361.893.477.102.542.735.1410.32
Ellipse-Net [31] A T E ( m ) 0.0150.0280.0390.1630.0750.0680.1020.149
A R E ( ) 0.771.141.524.252.582.153.224.71
Ours A T E ( m ) 0.0120.0210.0460.0770.0550.0660.0950.144
A R E ( ) 0.560.901.722.701.862.032.894.48
Table 4. The size and runtime of different models in the C h e s s sequence. Runtime is the average running time of an image.
Table 4. The size and runtime of different models in the C h e s s sequence. Runtime is the average running time of an image.
ModelModel Size (MB)Model NumberTotal Size (MB)Runtime (s)
Mathematical [15]---0.00014
Ellipse-Net [31]82.611908.60.01688
DEllipse-Net18.3118.30.01087
Table 5. The impact of different ellipse prediction methods and alignment methods on localization. “√” used object-level instance topology and alignment methods, while “×” did not use them. Translation error is ATE (m), Rotation error is ARE (°).
Table 5. The impact of different ellipse prediction methods and alignment methods on localization. “√” used object-level instance topology and alignment methods, while “×” did not use them. Translation error is ATE (m), Rotation error is ARE (°).
Ellipse ModelErrorsTopologyseq-01seq-02seq-03seq-04seq-05seq-06
Mathematical [15] A T E ( m ) ×0.1470.1520.2230.2080.1620.082
0.0660.0530.0370.0360.0350.045
A R E ( ) ×5.456.5910.518.886.482.96
2.182.041.071.031.191.55
Ellipse-Net [31] A T E ( m ) ×0.1360.1240.1650.1640.1080.048
0.0800.0480.0520.0370.0410.046
A R E ( ) ×6.444.687.797.624.421.60
2.802.351.711.091.461.60
DEllipse-Net (Ours) A T E ( m ) ×0.1250.1230.1620.1650.1050.049
0.0500.0420.0360.0320.0490.042
A R E ( ) ×5.775.127.507.314.321.65
1.671.921.080.951.751.47
Table 6. The A T E and A R E of OVS, Ellipse-Net, and our method across four sequences under normal and dark conditions.
Table 6. The A T E and A R E of OVS, Ellipse-Net, and our method across four sequences under normal and dark conditions.
MethodErrors fr 2 _ desk fr 3 _ office _ household Chess Redkitchen
NormalDarkNormalDarkNormalDarkNormalDark
OVS [35] A T E  (m)0.1020.1190.0440.0670.0810.1290.0400.050
A R E ( ) 8.736.391.151.709.5910.651.592.10
Ellipse-Net [31] A T E  (m)0.0240.0650.0490.0880.1430.2060.0430.107
A R E ( ) 0.892.441.572.416.8810.131.323.23
Ours A T E  (m)0.0120.0220.0130.0210.0550.1100.0370.043
A R E ( ) 0.560.890.330.521.863.931.612.02
Table 7. A T E and A R E of OVS, Ellipse-Net, and Ours across four sequences with varying numbers of occlusions.
Table 7. A T E and A R E of OVS, Ellipse-Net, and Ours across four sequences with varying numbers of occlusions.
MethodCover fr 2 _ desk fr 3 _ office _ household Chess Redkitchen
Number ATE  (m) ARE  (°) ATE  (m) ARE  (°) ATE  (m) ARE  (°) ATE  (m) ARE  (°)
OVS [35]One0.23817.000.0461.040.0531.670.0220.96
Two0.24117.080.0450.990.0622.000.0261.08
Three0.26018.190.0511.390.0692.230.0291.15
Ellipse-Net [31]One0.0301.320.0290.860.0952.940.0230.98
Two0.0321.330.0371.010.0983.040.0261.04
Three0.0481.620.0441.490.1123.330.0421.31
OursOne0.0170.830.0210.460.0310.800.0110.32
Two0.0170.830.0220.490.0381.310.0130.43
Three0.0341.320.0370.880.0421.420.0261.28
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wu, H.; Liu, Y.; Wang, C.; Wei, Y. Global Semantic Localization from Abstract Ellipse-Ellipsoid Model and Object-Level Instance Topology. Remote Sens. 2024, 16, 4187. https://doi.org/10.3390/rs16224187

AMA Style

Wu H, Liu Y, Wang C, Wei Y. Global Semantic Localization from Abstract Ellipse-Ellipsoid Model and Object-Level Instance Topology. Remote Sensing. 2024; 16(22):4187. https://doi.org/10.3390/rs16224187

Chicago/Turabian Style

Wu, Heng, Yanjie Liu, Chao Wang, and Yanlong Wei. 2024. "Global Semantic Localization from Abstract Ellipse-Ellipsoid Model and Object-Level Instance Topology" Remote Sensing 16, no. 22: 4187. https://doi.org/10.3390/rs16224187

APA Style

Wu, H., Liu, Y., Wang, C., & Wei, Y. (2024). Global Semantic Localization from Abstract Ellipse-Ellipsoid Model and Object-Level Instance Topology. Remote Sensing, 16(22), 4187. https://doi.org/10.3390/rs16224187

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop