Next Article in Journal
Low-Slow-Small (LSS) Target Detection Based on Micro Doppler Analysis in Forward Scattering Radar Geometry
Next Article in Special Issue
A Robust Indoor Positioning Method based on Bluetooth Low Energy with Separate Channel Information
Previous Article in Journal
Autonomous RFID Sensor Node Using a Single ISM Band for Both Wireless Power Transfer and Data Communication
Previous Article in Special Issue
Surface Correlation-Based Fingerprinting Method Using LTE Signal for Localization in Urban Canyon
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Reliable and Fast Localization in Ambiguous Environments Using Ambiguity Grid Map

School of Mechanical Science and Engineering, Huazhong University of Science and Technology, Wuhan 430074, China
*
Author to whom correspondence should be addressed.
Sensors 2019, 19(15), 3331; https://doi.org/10.3390/s19153331
Submission received: 28 June 2019 / Revised: 21 July 2019 / Accepted: 26 July 2019 / Published: 29 July 2019
(This article belongs to the Collection Positioning and Navigation)

Abstract

:
In real-world robotic navigation, some ambiguous environments contain symmetrical or featureless areas that may cause the perceptual aliasing of external sensors. As a result of that, the uncorrected localization errors will accumulate during the localization process, which imposes difficulties to locate a robot in such a situation. Using the ambiguity grid map (AGM), we address this problem by proposing a novel probabilistic localization method, referred to as AGM-based adaptive Monte Carlo localization. AGM has the capacity of evaluating the environmental ambiguity with average ambiguity error and estimating the possible localization error at a given pose. Benefiting from the constructed AGM, our localization method is derived from an improved Dynamic Bayes network to reason about the robot’s pose as well as the accumulated localization error. Moreover, a portal motion model is presented to achieve more reliable pose prediction without time-consuming implementation, and thus the accumulated localization error can be corrected immediately when the robot moving through an ambiguous area. Simulation and real-world experiments demonstrate that the proposed method improves localization reliability while maintains efficiency in ambiguous environments.

1. Introduction

For an autonomous robot, reliable and fast localization is a critical prerequisite during navigation. However, ambiguous environments may cause perception aliasing of the external sensors [1,2], which makes it a challenging task for a robot to accurately determine its pose or even leads to localization failure. As demonstrated in Figure 1, the ambiguous environments are specified as the environments that contain ambiguous areas, including long corridors, empty space, and tangled thicket. In such environments, however, there still exist some unambiguous areas that can help the robot to locate itself. For this end, constructing a localization method on the basis of the modeled ambiguity of the environment can increase the localization reliability without loss of efficiency.
For localization procedure of a robot, global localization is firstly performed to obtain the initial pose by using the observation data from external sensors, followed by the pose tracking which enables the robot to rapidly track its pose [3]. To be more specific, pose tracking consists of two steps, i.e., prediction and update. In the prediction step, based on the odometry motion model, the pose increment obtained by internal sensors is incorporated into the results of the previous localization moment to predict the current pose. As for the update, the localization error of the predicted pose is corrected by the observation data of external sensors, which respects the observation model to produce the final localization results of the current moment.
Recently, most of the existing localization methods [4,5,6,7,8,9] only rely on the external sensors in the update step, and it remains a tough job to satisfy the practical requirements when rectifying the localization error in an ambiguous environment. This is because the external sensors may suffer from perceptual aliasing, which implies that the observation data captured at different pose are difficult to distinguish due to the environmental ambiguity. Consequently, the localization error of the prediction step may accumulate as the robot moves through an ambiguous area of the environment. Even if the robot reaches an unambiguous area, the accumulated localization error is often too large to be corrected by those methods.
To solve the localization problem in an ambiguous environment, additional facilities such as artificial reflectors [10,11] and wireless sensor networks [12] have attracted increasing attention. With additional facilities, these methods require additional maintenance costs and need to modify the existing environment, which are not unreasonable for some practical usage. Assuming that a robot will leave the area where perceptual aliasing occurs and enters an unambiguous area, the pose of the robot can be recovered by a global localization method. In an unambiguous area, although global localization method can be used to estimate the pose of a robot according to the observation data captured by external sensors, the operating time of global localization is much longer than that of pose tracking. In addition, it is necessary to solve the detection problem of when the robot enters the unambiguous area. Additional information [13,14] except for the readings of internal sensors is utilized in the prediction step of pose tracking to allow more reliable pose prediction. However, their methods are impractical to solve the robot localization problem in ambiguous environments due to the fact that they do not explicitly consider the ambiguity of an environment and such environment property contains useful information for robot localization.
For robot navigation, the adaptive Monte Carlo localization (AMCL) method is able to achieve effective and fast robot localization in different environments [6,7,8,9] and the particle representation used in AMCL has included some effects of ambiguity. However, due to the limitation of particle number, the inaccurate motion model and observation model, it cannot cope with long term perception aliasing. In this paper, the AGM which explicitly models the ambiguity of an environment is utilized to improve the standard AMCL so that it can correct the accumulated localization error after robot moving through an ambiguous area and ultimately achieve a fast and reliable robot localization. The main contributions of this paper are summarized as follows:
  • The AGM is achieved by proposing a new localizability evaluation method called average ambiguity error (AAE) so that the possible localization error at a given pose can be estimated, and those unambiguous areas of an environment can be identified.
  • By integrating the AGM, the standard Dynamic Bayes network (DBN) for robot localization is improved to model the localization problem in ambiguous environments. Moreover, a new motion model referred to as portal motion model is implemented to obtain more reliable pose prediction in ambiguous areas.
  • Based on the improved DBN, the AGM-based adaptive Monte Carlo localization (AGM-AMCL) method is derived to achieve fast probability inference.
  • Simulation and real-world experiments validate the effectiveness of the AGM and the AGM-AMCL method, which can reliably locate a robot with guaranteed efficiency in three different ambiguous environments.
The remainder of this paper is constructed as follows: Section 2 presents a range of related work in terms of localizability evaluation and reliable localization. Section 3 provides the AGM and its detailed implementation. In Section 4, we describe the improved DBN, and then we derive the portal motion model and the AGM-AMCL. Finally, in Section 5, simulations and real-world experiments are conducted to evaluate the performance of the AGM and the AGM-AMCL method. Finally, some conclusions are offered in Section 6.

2. Related Work

Considering the key issue we attempt to solve in this paper, we will focus on the related work of localizability evaluation and reliable localization in ambiguous environments. To the best of our knowledge, there are few articles explicitly dealing with the localization problem in ambiguous environments. Global localization methods and optimized pose tracking methods can be used to relocate a robot after the robot leaving an ambiguous area and entering an unambiguous area. Therefore those methods were introduced in the reliable localization part.

2.1. Localizability Evaluation

Evaluating the localizability in an environment is to quantitatively estimate the achievable localization accuracy of a given pose. In general, localizability evaluation methods are not related to the property of an environment, therefore they are applicable in both ambiguous environments and unambiguous environments. Fisher information matrix (FIM) is a useful indicator for evaluating environmental localizability. Based on the Cramér-Rao lower bound, Censi [15,16] used FIM to represent the theoretical accuracy limit of a localization method employing range-finder data and attempted to use FIM to analyze the achievable accuracy of pose tracking. FIM has no analytical solution if an environment is represented by the occupancy grid map. To solve this problem, a discrete FIM [17] was proposed, which was obtained by simulated scan rays from an occupancy grid map. Afterwards, Irani et al. [18] presented a path planning method for a mobile manipulator based on discrete FIM. The localizability evaluation methods based on FIM can be used to evaluate the theoretical localization accuracy limit of a certain position, however, it is difficult for FIM to estimate the localization accuracy in a real localization process. This is because, under a given environment, the FIM is only related to the robot’s pose, without considering the environmental similarity of different poses and the observation model used in the actual localization process.
In addition to the localizability evaluation based on the FIM, there have been some related studies in recent years. In [19], a novel strategy for incremental sampling-based motion planning was proposed, which introduced a new evaluation of localizability using the trace change rate of covariance matrix from the prediction step to the update the step of an extended Kalman filter. Zhen et al. [20] proposed a method to evaluate the localizability based on the constraints provided by a sensor measurement. The above methods can evaluate the environmental localizability relatively while the evaluation value itself is meaningless for estimating localization error. In [21], to evaluate the localizability, a vehicle was driven through the designated environment and an error model was learned. However, their method takes a long time to experiment and is not suitable for large scale environments. Javanmardi et al. [22] adopted four types of criteria to evaluate the capability of maps for localization. Based on their criteria, the localizability of a map can be comprehensively evaluated. Our method differs from theirs in the sense that the observation model adopted in the real localization process and the sensor measurement uncertainties are involved in our evaluate method.

2.2. Reliable Localization

In recent years, many localization methods have been proposed to enhance localization reliability. One of the commonly used methods is adding additional facilities. Our previous work [10] proposed a new observation model for MCL, which improves the robustness of localization by reliable reflector prediction in the ambiguous environments. Beinhofer et al. [23] addressed the data association problem by deploying a limited number of uniquely identifiable artificial landmarks. A practical approach to landmark placement was presented in [24] that minimizes the localization errors. Besides, wireless sensor networks [12], radio frequency identification [25,26], etc. were also applied to achieve reliable localization. However, the additional facilities are costly for placement and maintenance, which limits the abovementioned method for potential applications.
Once localization fails, re-localization can be achieved through global localization, thereby improving the reliability of localization. A particle filter improved by a particle swarm optimization algorithm was proposed in [27]. With their method, more robust global localization can be achieved using fewer particles. A laser scan matching with an improved probabilistically-motivated Maximum Likelihood Estimation algorithm is adopted to obtain the global optimum position [2], which can reach an updating rate of 2 Hz in a feature-poor environment. A novel branch-and-bound multiresolution global localization method using a multiresolution Gaussian mixture map was proposed in [28] which was used for real-time localization in urban environments. By finding correspondences between a set of sparse features detected in two occupancy grid maps, a map to map matching method was designed [29], which can be used for global localization. In [30], a differential evolution approach was tested for global localization in an environment with different types of occlusions, which used three different cost functions based on a non-symmetric fitness function. The above methods attempt to achieve reliable and fast global localization. However, due to the need for global optimization, their running time is still longer than pose tracking. In [31] and [32], classifiers for pose recognition were trained by the machine learning method. Through such classifiers, the global pose of a robot can be identified quickly. However, those methods rely on reliable feature extraction methods, and wrong feature extraction may result in large localization errors.
By optimizing the prediction step of pose tracking, it can generate more accurate pose samples, also called particles, and improve the reliability of localization. Gutmann et al. [33] used sensor resetting method to generate some pose particles in the prediction step. The number of particles relies on the ratio between the short term and the long term average observation likelihood. Building on the basic Monte Carlo localization algorithm, a robust localization algorithm called Mixture-MCL was developed [34], which attempted to generate pose samples according to the observation model using a set of learned kd-trees in the prediction step to complement the odometry motion model. However, this method takes a large amount of memory, and as the size of the environment grows, the amount of memory required grows rapidly. Zhang et al. [13] proposed the concept of similar energy region. Extra pose samples can be generated in similar energy region to obtain a reliable pose prediction. Oh et al. [14] developed an Augmented Motion model, which used the semantic information available in maps to bias the motion model toward areas of higher probability. In [35], by using a global hypothetical pose set to represent an environment, a novel qualitative motion model is proposed to generate extra pose hypothesis. In [13,14,33,34,35], additional information is used in the prediction step to generate extra pose samples which may closer to the robot true pose. Once the accumulated localization error is too large, extra pose samples may compensate for such error and improve the reliability of localization. However, their methods do not explicitly consider the ambiguous nature of an environment, it is very difficult for these methods to achieve reliable localization in ambiguous environments. Inspired by the abovementioned methods, additional information from the AGM is exploited to improve the prediction step of the AMCL method, which allows our method to generate some extra particles at a proper time with proper distribution and effectively improves the reliability of localization.

3. AGM

AGM is used for modeling the ambiguity of an environment, such that we can figure out which positions suffer from perceptual aliasing and how much localization error will be caused by perceptual aliasing at those positions. Analogical to the occupancy grid map, AGM has multiple cells, each of which measures the localization error resulting from perceptual aliasing. In this paper, we refer the aforesaid localization error to the AAE, and it is crucial to build the AGM. Practically, it is noticed that a robot cannot distinguish two different poses effectively due to the similar surroundings of these poses or the considerable sensor noise or poor observation model [36]. Considering that, the formulation of AAE should be related to the environment as well as the external sensor and the constructed observation model.
Define x = ( p , θ ) as a pose with p = ( x , y ) used to denote a position in two-dimension. Let z x be the observation data obtained by the external sensor of a robot at the pose x . Using the observation data z x and an input pose y , the likelihood function of an observation model is denoted by:
L x ( y ) P ( z x | y )
From Equation (1), we conclude that L x ( x ) represents the likelihood value when the input pose y is equal to the robot’s true pose. For a general pose y if the maximum value of L x ( y ) is identical with L x ( x ) , the robot’s pose can be accurately obtained through maximum likelihood estimation. However, in an ambiguous environment, there are other values of y that makes L x ( y ) equal or bigger than L x ( x ) . In this situation, it is difficult to distinguish pose x and pose y based on this likelihood function. Thus, in this paper, the probability of a robot’s true pose x confused with some other pose y can be defined as:
P y | x A P ( L x ( y ) + ε L x ( x ) )
where ε is a small threshold to compensate the unmodeled factors of the likelihood computation.
Given the observation data z x , the likelihood value of L x ( y ) can be uniquely determined. Thus probability P y | x A is rely on the distribution of z x and Equation (2) can be expressed as:
P y | x A =   z x 1 ( L x ( y ) + ε L x ( x ) | z x ) ( z x ) P ( z x ) d z x
where P ( z x ) is the distribution of observation data at the pose x and 1 ( L x ( y ) + ε L x ( x ) | z x ) ( z x ) denote an indicator function, i.e.:
1 ( L x ( y ) + ε L x ( x ) | z x ) ( z x ) = 1   L x ( y ) + ε L x ( x ) | z x   0   L x ( y ) + ε < L x ( x ) | z x
Since P y | x A cannot be determined analytically, we employ Monte Carlo integral to achieve its approximation as:
P y | x A   1 N i = 1 N 1 ( L x ( y ) + ε L x ( x ) | z x i ) ( z x i )
where z x i denotes the i-th sample of the observation data, N denotes the total number of samples.
Now we define the AAE of a given pose x as:
a a e ( x ) = Δ i M d ( x + Δ i , x ) P x + Δ i | x A Δ i M P x + Δ i | x A
where M is a set of the incremental pose, Δ i ( Δ x i , Δ y i , Δ θ i ) denotes the i-th element in M , d ( x + Δ i , x ) denotes the difference between x + Δ i and x , which is expressed by:
d ( x + Δ i , x ) = Δ x i 2 + Δ y i 2 + ξ | Δ θ i |
where ξ is a coefficient factor used to convert the orientation difference into the distance difference.
From Equation (6), it is known that the AAE of a pose represents the weighted average localization error caused by the ambiguity between the true pose and the other poses which are contained by M . Large AAE value of a given pose means that a large localization error may be induced after a robot moving through that pose. Moreover, by exploiting the AAE value of a trajectory, the accumulated localization can be estimated in some way and this property is used in our localization method in the following section.
Then, based on Equation (6), the AAE of a position can be derived directly:
a a e ( p x ) = 1 n θ θ = 0 ° 360 ° a a e ( x ( p , θ ) )
where n θ denotes the number of discrete θ .
Figure 2 shows a schematic diagram of M . Defining R d ( M ) and R θ ( M ) as the distance range and the orientations range of the set M , respectively, we have:
M = { Δ i ( Δ x i , Δ y i , Δ θ i ) | Δ x i 2 + Δ y i 2 R d ( M ) | Δ θ i | R θ ( M ) }
In this paper, in accordance with occupancy grid map, Δ x i and Δ y i are pre-selected as an integer multiple of the grid map resolution and Δ θ i is selected as an integer multiple of 3° to make the trade-off between the efficiency and accuracy.
To derive a solvable solution for AAE determined by Equation (8), we need to focus on the sampling method of P ( z x ) and the configuration of ξ , R d ( M ) and R θ ( M ) . In addition, considering that a 2-D laser range finder is used as the external sensor in this paper, the following method for computing P ( z x ) is based on the 2-D laser range finder. First, sampling with a large amount of data is time-consuming and unrealistic in a large-scale environment. As studied in [37], the 2-D laser range finder and the beam model are suggested to be used as the external sensor and observation model for the practical implementation, respectively. Then we can adopt the following feasible and effective method to obtain observation data samples: (1) build the occupancy grid map of an environment using SLAM method; (2) determine the parameters of the beam model by real observation data using learning method presented in [37]; (3) obtain the observation data samples at various poses through simulation. Second, the ξ is expressed by ξ = χ 1 / r a d ( χ 2 ) , where χ 1 and χ 2 denotes the maximum tolerable distance error and the maximum tolerable orientation error for reliable localization, separately. Finally, by using the inequality a a e ( x p ) R d ( M ) + ξ R θ ( M ) , which can be derived from Equations (6) and (7), we can obtain the range constraints for M , i.e., χ 1 R d ( M ) ,   χ 2 R θ ( M ) . More accuracy localizability evaluation can be obtained through a larger range of M while it is time-consuming. Therefore, we select R d ( M ) = χ 1 and R θ ( M ) = χ 2 .

4. AGM-AMCL

As shown in Figure 3, the standard AMCL [38] mainly contains three steps: resampling step, prediction step, and update step. In the resampling step, particles are sampled according to the likelihood of each particle. Then new particles are generated in the prediction step based on the odometry motion model. After that, the likelihood of each particle is evaluated by the observation model in the update step. In addition to those three steps, our method also includes a step for estimating the accumulated ambiguity error using the AGM. Moreover, in the prediction step of our method, the portal motion model is used to generate extra particles in an adjacent unambiguous area if a robot is in an ambiguous area. Due to the extra particles, large localization error can be corrected as soon as the robot enters the adjacent unambiguous areas.
In fact, our method can be derived from a DBN which models the robot localization problem in unambiguous environments. In the following part, the derivation of our method, as well as the portal motion model and the estimation method for the accumulated localization error using the AGM, will be detailed introduced.

4.1. Standard DBN for Localization

Given a prior occupancy grid map m g , the robot localization can be modeled as a probabilistic inference problem:
B e l ( x t ) = P ( x t | z 1 : t , u 1 : t , m g )
where x t denotes the robot pose at a time t , z 1 : t is the sequence of observations, u 1 : t denotes a sequence of motion control commands or odometry measurements.
Figure 4 demonstrates the standard DBN [39] for robot localization which contains two importance independent assumptions: Markov independence of the odometry and Markov independence of the observation.
According to the independent assumptions embedded in the standard DBN, we can simplify the robot pose belief that presented in Equation (10) as:
B e l ( x t ) = η P ( z t | x t , m g ) x t 1 P ( x t | x t 1 , u t ) B e l ( x t 1 ) d x t 1
where η is a normalization coefficient, P ( x t | x t 1 , u t ) and P ( z t | x t , m g ) are corresponding to the motion model and observation model, respectively. To fast infer the robot’s pose belief in Equation (11), AMCL [38] can be used with the KLD-sampling method regulating the particle set size.
Because of perception aliasing, the localization error accumulates rapidly when a robot moves through an ambiguous area. However, the DBN, shown in Figure 4, does not explicitly consider the accumulated localization error caused by perception aliasing so that it is not suitable for modeling the robot localization problem in ambiguous environments. For this reason, an improved DBN is proposed.

4.2. Improved DBN

The improved DBN is shown in Figure 5, where m g stands for the occupancy grid map and m a is the corresponding AGM, e t is the cumulative ambiguity error at time t, which reflects the accumulated localization error caused by perception aliasing.
To infer the robot’s pose, as well as the accumulated localization error, the robot localization in ambiguous environments is expressed as:
B e l ( x t , e t ) = P ( x t , e t | z 1 : t , u 1 : t , m a , m g )
Also, based on the independent assumptions embedded in the improved DBN, we have:
B e l ( x t , e t ) = P ( x t , e t | z 1 : t , u 1 : t , m a , m g ) = η P ( z t | x t , m g ) P ( x t , e t | z 1 : t 1 , u 1 : t , m a , m g ) = η P ( z t | x t , m g ) P ( x t , e t , x t 1 , e t 1 | z 1 : t 1 , u 1 : t , m a , m g ) d x t 1 d e t 1 = η P ( z t | x t , m g )   P ( x t , e t | x t 1 , e t 1 , u t , m a ) P ( x t 1 , e t 1 | z 1 : t 1 , u 1 : t 1 , m a , m g ) d x t 1 d e t 1   = η P ( z t | x t , m g ) P ( e t | x t , e t 1 , m a )   P ( x t | x t 1 , e t 1 , u t , m a ) B e l ( x t 1 , e t 1 )
As compared to Equation (11), we conclude that Equations (11) and (13) share the same observation model P ( z t | x t , m g ) . A different motion model P ( x t | x t 1 , e t 1 , u t , m a ) considering the cumulative ambiguity error e t 1 and the m a are imposed in the improved DBN. Meanwhile, Equation (13) has an additional term P ( e t | x t , e t 1 , m a ) to estimate the cumulative ambiguity error. It is noted that the multiple integrations in Equation (13) cannot be derived analytically. In our paper, the AMCL is adopted to inference the robot’s pose as well as the accumulated localization error, which utilized a set of weighted particles to estimate the probability distribution. Therefore, the probability distribution P ( x t , e t , x t 1 , e t 1 | z 1 : t 1 , u 1 : t , m a , m g ) can be reformulated according to [40] as:
P ( x t , e t , x t 1 , e t 1 | z 1 : t 1 , u 1 : t , m a , m g ) = i = 1 n P ω t i δ ( [ x t , e t , x t 1 , e t 1 ] [ x t i , e t i , x t 1 i , e t 1 i ] )
where [ x t i , e t i , x t 1 i , e t 1 i ] denotes a sample of robot state also called particle, ω t i stands for the i-th particle weight, δ is the Dirac delta function and n p is the number of particles.
Substituting Equation (14) into the second part of Equation (13) yields:
P ( x t , e t | z 1 : t , u 1 : t , m a , m g ) = P ( x t , e t , x t 1 , e t 1 | z 1 : t 1 , u 1 : t , m a , m g ) d x t 1 d e t 1 = i = 1 n P ω t i δ ( [ x t , e t , x t 1 , e t 1 ] [ x t i , e t i , x t 1 i , e t 1 i ] ) d x t 1 d e t 1 = i = 1 n P ω t i δ ( [ x t , e t ] [ x t i , e t i ] )

4.3. Portal Motion Model for Improved DBN

Intuitively, the prediction step of the AMCL generates some particles in an unambiguous area which is adjacent to the ambiguous area, the update and resampling steps of the AMCL method will make the particles immediately converge to the robot’s true pose when the robot leaves the ambiguous area. To accommodate that, we rewrite the motion model P ( x t | x t 1 , e t 1 , u t , m a ) as:
P ( x t | x t 1 , e t 1 , u t , m a ) = ( 1 α t ) P o d o ( x t | x t 1 , u t ) + α t P p o r t a l ( x t | x t 1 , e t 1 , m a )
where P o d o ( x t | x t 1 , u t ) is the odometry motion model, α t is a weighting factor and P p o r t a l ( x t | x t 1 , e t 1 , m a ) is the portal motion model to generate the particles in an adjacent unambiguous area, which is defined as:
P p o r t a l ( x t | x t 1 , e t 1 , m a ) = { η h i t e 1 2 ( x t μ h i t ) T Σ h i t 1 ( x t μ h i t ) , f o r   a a e ( p x t ) t a 0 ,   f o r   a a e ( p x t ) > t a
where η h i t is a normalization coefficient, a a e ( p x t ) denotes the AAE of the position p x t for the pose x t , t a is a threshold and a a e ( x t ) t a implies the robot is in an ambiguous area. t a is set to χ 1 which denote the maximum tolerable distance error. It is concluded from Equation (17) that given x t 1 , e t 1 and m a , x t obeys a truncated Gaussian distribution. Specifically, in an ambiguous area, the probability of x t is zero whereas it obeys a Gaussian distribution with a mean μ h i t and a covariance matrix Σ h i t in an unambiguous area.
The portal motion model describes a state transition that a robot may be immediately moved to someplace in the adjacent unambiguous area regardless of the odometer measurement. Despite the physically impossible, it can be used to obtain more reliable pose prediction in an ambiguous environment so that the accumulated localization error can be corrected immediately when the robot reaches the adjacent unambiguous area. As a complement to the odometer motion model, the portal motion model explicitly indicates the effect of cumulative localization error caused by perception aliasing on prediction robot pose.
A schematic diagram for the portal motion model is shown in Figure 6. We define d r a n g e e t 1 as the cumulative ambiguity error at the time t 1 , and its length is shown with a blue line. Assuming the robot pose at the time t 1 is x 1 which is in the ambiguous area, μ h i t can be generated by a simulated beam with ray casting method [41]. Practically, the simulated beam continues unless the endpoint of the beam hits an unambiguous area or an obstacle or the length of the beam reaches d r a n g e . As shown in Figure 6, the red line shows the simulated beam generated from x 1 and the length of the beam is represented as d h i t . if the beam ends in an unambiguous area, the position and the orientation of μ h i t can be determined by the position and orientation of the beam, respectively. And if the beam ends because of the other two conditions, the portal motion model is not enabled. Define σ p and σ θ representing the adjustable standard deviation of the position and orientation, respectively. Large σ p and σ θ means large uncertainties of the truncated Gaussian distribution. Then, we can calculate Σ h i t by:
Σ h i t = [ σ p 2 0 0 0 σ p 2 0 0 0 σ θ 2 ] [ ν 1 , ν 2 , ν 3 ] [ 1 0 0 0 k h i t 0 0 0 1 ] [ ν 1 , ν 2 , ν 3 ] 1
where ν 1 is a unit vector with the same orientation as x 1 , ν 2 is the unit vector perpendicular to ν 1 , ν 3 = ( 0 , 0 , 1 ) T , k h i t is a stretch factor that used to adjust the ratio between the feature vectors of the covariance matrix Σ hit given by:
k h i t = ( k max ,   f o r d r a n g e d h i t k max d r a n g e d h i t ,   f o r   e l s e
where k max is the parameter to limit the ratio between the feature vectors of the covariance matrix Σ h i t .
Figure 7a–c depict the changing process of Σ h i t . It can be seen that as the pose of the robot changes from x 1 to x 3 , d h i t gradually decreases ( d h i t 1 > d h i t 2 > d h i t 3 ). According to Equation (18), the uncertainties of the truncated Gaussian distribution in the direction perpendicular to the robot’s orientation gradually increase due to the decreasing of d h i t , as represented by the blue oval area. We adopt the KLD-sampling method such that few and relatively concentrated particles will be generated by the portal motion model when the robot is in an ambiguous area and is far away from an unambiguous area. Moreover, as the distance between the robot and the unambiguous area getting shorter, more dispersed particles will be generated by the portal motion model due to large uncertainties of Σ h i t . With an adaptive Σ h i t , we can reasonably add uncertainties for the portal motion model and improves the reliability of localization.

4.4. AGM-AMCL Implementation

For implementation, P ( e t | x t , e t 1 , m a ) is designed using a deterministic function:
e t = { e t 1 + a a e ( p x t ) ρ e t 1 + 1 f o r   a a e ( p x t ) > t a 0   f o r   a a e ( p x t ) t a
where ρ 0 and is a parameter to control the changing rate of e t . By Equation (20), we show that e t in an ambiguous area and an unambiguous area is derived as e t = e t 1 + a a e ( p x t ) / ( ρ e t 1 + 1 ) and e t = 0 , respectively.
Using the length of the beam d h i t , we schedule the α t adaptively as:
α t = ( 0 f o r   ( d h i t = d r a n g e ) ( d h i t = 0 ) 1 d h i t d r a n g e f o r 0 < d h i t < d r a n g e
As demonstrated in Equation (21), if the beam ends due to the length restriction or the zero- length of d h i t which means the beam end because of obstacle or the pose x t in an unambiguous area, α t is assigned to zero, otherwise, it is regulated in real-time where 0 < d h i t < d r a n g e indicates that the robot is in an ambiguous area and an unambiguous area is found in the scope of d r a n g e .
A detailed overview of the AGM-AMCL algorithm can be found in Algorithm 1. The output particle set is initialized as an empty set in Step 1. Step 2 declares a variable C t to store the ray casting result to avoid repetitive computation and is initialized from Step 3 to Step 5. Step 7 to Step 23 show the prediction step of our localization method which is derived from Equation (16). μ h i t and d h i t are obtained by the ray casting method at step 10, α t is calculated according to Equation (21) at step 15. At step 16, c is a binary variable sampled from a Bernoulli distribution with parameter α t . c = 1 means that the prediction pose will be sampled from the portal motion model while c = 0 means the prediction pose will be sampled from the odometry motion model. At step 18, Σ h i t is computed according to Equation (18). Through step 19 and 20, the portal model is used to sample a prediction pose. According to Equation (20), the cumulative ambiguity error e t is updated at step 24. In order to improve the efficiency of the particles, the KLD sampling method presented in [38] is adopted at step 28.
Algorithm 1 AGM-AMCL
Input: S t 1 = { x t 1 [ i ] , e t 1 [ i ] , ω t 1 [ i ] i = 1 N t 1 } , u t , z t , m g , m a
Output:   S t = { x t [ i ] , e t [ i ] , ω t [ i ] i = 1 N t }
1:   S t , N x 0 , n 0 , β 0    // Initialization
2:   C t = { x h i t [ i ] , d h i t [ i ] , tag [ i ] i = 1 N t 1 }    //For storing RAYCAST result
3:  for i 1 , , N t 1 do   //Initialize C t
4:    tag [ i ] 0
5:  end for
6:  do
7:    j   RESAMPLING   ( S t 1 )   //Resampling according to particle weight
8:    d range = e t 1 [ j ]
9:   if tag [ j ] = = 0 then
10:     ( μ h i t , d h i t ) RAYCAST ( x t 1 [ j ] , d range , m a )
11:     C t [ j ] ( x h i t , d h i t , 1 )
12:   else
13:     ( μ h i t , d h i t ) C t [ j ]
14:   end if
15:    α   COMPUTEALPHA   ( d h i t , d range )
16:   sample c according to Bernoulli_Distribution( α )
17:   if   c = = 1
18:     Σ h i t   COMPUTESIGMAHITT   ( d range , d h i t , x t 1 [ j ] )
19:    Create P portal ( x t | x t 1 [ j ] , e t 1 [ i ] , m a ) according to N ( μ h i t , Σ h i t )
20:    Sample     x t [ n ] according to P portal ( x t | x t 1 [ j ] , e t 1 [ i ] , m a )
21:   else
22:    Sample x t [ n ] according to P o d o ( x t | x t 1 [ j ] , u t )
23:   end if
24:    e t [ n ]   PREDICTAMBIGUITYERROR   ( x t [ n ] , e t 1 [ j ] , m a )
25:    ω t [ n ] P ( z t | x t [ n ] , m a ) //Compute weight according to observation
26:    β β + ω t
27:    S t S t { x t [ n ] , e t [ n ] , ω t }
28:    N x ADPTINGSAMPLESIZE ( x t [ n ] )
29:    n n + 1
30:  while n < N x and n < N min
31:  for i 1 , , n do   //Normalize particle weight
32:    ω t [ i ] ω t [ i ] / β
33:  end for
34:  return S t

5. Experiments and Results

In this section, we will verify the effectiveness of the AGM and the performance of the AGM-AMCL algorithm severally. Note that, if not particularly indicated, AAE represents the AAE of a position in this section.

5.1. Platform and Environments

The experimental platform and environments are shown in Figure 8, respectively. The used mobile robot named Turtlebot2 is equipped with a Hokuyo UTM-30LX LIDAR and an LPMS-RS232AL2 IMU, which serve as the external and internal sensor, separately. As depicted in Figure 8a,b, three different ambiguous environments with long corridors, open square and tangled thicket are considered for the real-world experiments.

5.2. AGM of Artificial Environments

To verify the effectiveness of the AGM, we draw several grid maps, in which the localizability of different position can be easily distinguished manually. In our research, the log-likelihood is used for observation likelihood computation. And is set to 0.5 which is a very small value for log-likelihood and is useful to compensate for some numerical calculation error. χ 1 and χ 2 were selected as 0.25 m and 5° respectively in this experiment, therefore the maximum AAE value is 0.5 m. As demonstrated in Figure 9, the top row and bottom row present three artificial occupancy grid maps and their corresponding AGM, respectively. Low AAE value (blue) of a cell means small localization error while high AAE value (red) means large localization error. When building these AGMs, the range error of the laser scan simulator was set to 0 m to simulate the laser scan data. The map shown in Figure 9(a1) is a long corridor. It can be seen from its AGM that the AAE value in both ends of the corridor is small, and the ε AAE value in the middle of the corridor is high, which is according with human experience. The map shown in Figure 9(b1), is a circle. When a robot is at the center of the circle, it will obtain the same sensor readings regardless of the orientation of the robot. Therefore, as shown in Figure 9(b2), the AAE value at the center of the circle is very high. An open space environment is shown in Figure 9(c1). We can see from its AGM that the farther a position is away from the obstacle, the higher the AAE value. For those cells where the AAE value is higher than one-half of the maximum AAE value (In this AGM experiment for artificial environments, the maximum AAE value is 0.5 m), the AAE value shown in Figure 9(c2) is significantly higher than the AAE value shown in Figure 9(a2,b2). This is because in the open space of Figure 9(c1), a robot actually cannot recognize anything and it cannot localization itself. But in Figure 9(a1,b1), the robot is only confusing in some directions and the sensor reading obtained from those positions can be used to reduce some of the localization uncertainties.
Figure 10 shows the performance of the AGM of an unambiguous environment under the condition that the simulated laser scan data have different range error. Figure 10a shows the occupancy grid map. Figure 10b,c are the corresponding AGM of Figure 10a when the range error of the laser scan simulator is set to 0 m and 0.5 m, respectively. By comparing Figure 10b with Figure 10c, we can clearly see that even in the same environment, large range error will lead to large AAE value. Although the environment shown in Figure 10a does not seem to be an ambiguous environment, experiment results show that large range error of the laser scan data may still lead to perception aliasing, which demonstrates the importance of sensor noise for localizability evaluation.

5.3. AGM of Real Environments

The SLAM method proposed in [42] was used to build the occupancy grid maps of scene A, B and C which are shown in Figure 11(a1–c1), respectively. In fact, it is difficult to build the occupancy grid maps of these three environments because of their ambiguity. However, with the help of those vehicles parked on the side of the road, localization can be achieved during mapping. Moreover, loop closure and offline optimization adopted by the SLAM method enable us to build consistent occupancy grid maps. After mapping, those vehicles are manually removed from the occupancy grid maps due to the fact that the position of those parked vehicles had been changed when we carried out our localization experiment. The resolution of our occupancy grid maps was set as 0.2 m. In addition, χ 1 and χ 2 are selected as 0.5 m and 10°, respectively, in our experiments. The true noise parameters of the Hokuyo UTM-30LX 2D Laser range finder are adopted to generate more realistic laser scan data for computing AAE.
It can be seen from the AGM, shown in Figure 11(a2–c2) that those positions which are in the middle of a long corridor, in the center of the open square, near a straight-line wall and tangled thicket get high AAE value. Those positions near a corner get low AAE. The results are in line with human experience.

5.4. Localization Simulation

The aim of the simulation experiments is to comprehensively evaluate the performance of our proposed AGM-AMCL method and compare it with the AMCL method and the augmented MCL method proposed in [37]. In fact, the augmented MCL is very similar to the method proposed in [33]. The difference is that extra random particles are added in [37] while sensor resetting method was adopted to generate extra particles in [33], which is not suitable in large scale and ambiguous environments. The trajectories (Figure 11) which are selected to pass through an ambiguous area are used for all simulation experiments. Laser scan data used for localization simulation can be generated at each pose on the simulation trajectories with the help of the occupancy grid maps. t a is set to 0.5 which is the maximum tolerable distance error. k max is selected to 2 and ρ is set to 0.5. These three parameters were used in all of our experiments. In addition, the localization methods are initialized around the true position in our simulation experiments only to verify the pose tracking performance.
Figure 12 presents the evolution process of particles during localization simulation in scene B. The color of a particle visualizes the particle weight just like the color used in Figure 10 except that the range represented by the color is 0 to 1.
In the beginning, those particles (Figure 12(a1)) generated by AMCL can follow the true pose of the robot but as the robot entering an ambiguous area, the particles (Figure 12(b1)) became divergent and there were fewer and fewer particles (Figure 12(c1)) near the true pose of the robot during the robot moving through an ambiguous area. Even the robot entered an unambiguous area, the particles (Figure 12(d1)) generated by AMCL could not correct the accumulated localization error and converged to a wrong pose. For augmented MCL, extra random particles Figure 12(b2–d2) are added when the average likelihood of particles was decreased. However, those extra random particles led the method to converge to a wrong place in large scale and ambiguous environment. This result shows that it is essential to generate particles in proper locations, otherwise wrong convergence could occur. On the other hand, as the robot entered the ambiguous area, our localization method generate extra particles (Figure 12(c3)) in the adjacent unambiguous area when the accumulated localization error is large enough. Once the robot reached the unambiguous area, those extra particles (Figure 12(d3)) help the robot to correct its accumulated localization error immediately.
Figure 13, Figure 14 and Figure 15 present the localization simulation results of three localization methods using the trajectory shown in Figure 11. Distance error and orientation error are shown in (a) and (b). Number of particles and the running time are shown in (c) and (d). In terms of the distance error, our method has better performance than the other two methods. The distance error of AMCL is divergent when the robot moved through the ambiguous areas while the distance error of augmented MCL fluctuate greatly because of the randomly added particles. This is because the orientation can be easily determined in the long corridor while it will cause large distance error. The computational complexity of those three methods is related to the number of particles. However, the particle number of our method and AMCL is stochastic, which rely on the sensor measurement data during robot moving. For this reason, it is hard to evaluate the efficiency theoretically. From the results of three different environments, the efficiency of our method can be demonstrated in three aspects. First, the running time of our method is similar to AMCL in unambiguous environments by using the AGM. Second, because of the right convergence, our method is more efficiency after the robot moving through an ambiguous environment. Third, our method is slower than AMCL when the robot is in an ambiguous area and near an unambiguous area, but it is still enough for real-time localization.
In order to evaluate the localization reliability of our method, we consider the last localization result after a robot traveling through a trajectory. If the distance error of the last localization result is smaller than 0.5 m and the angular error of the last localization result is smaller than 10°, we assume this localization experiment is successful, otherwise, this localization experiment fails. Then the reliability of our localization method can be verified through the localization success rate of repeated localization experiments on the same trajectory. Various degrees of odometer errors were adopted in our reliability experiment. The true odometry data is computed according to:
u t t r u e = x t 1 t r u e x t t r u e
where and are the pose compounding operator [43], x t 1 t r u e and x t t r u e are the true pose of a robot at time t 1 and time t , respectively. x t 1 t r u e and x t t r u e can be obtained from the simulated trajectory. Taking the true odometer data as the mean and setting a covariance matrix, a Gaussian distribution can be obtained. The odometer data with errors can be obtained by sampling from such Gaussian distribution. In our experiments, the abovementioned covariance matrix is set to a diagonal matrix with one parameter:
Σ s a m p l e = [ m 0 0 0 m 0 0 0 0.2 * m ]
Then we define the odometry error rate η as:
ς = m distance ( x t t r u e , x t 1 t r u e )
By changing ς , the odometer data with different degrees of error can be obtained. In our reliability experiments, eleven different value of ς varying from 0 to 1 were selected to evaluate the localization reliability under different degree of odometry error. In addition, given ς and a trajectory, we carried out one hundred repetitive experiments to get the localization success rate. The results of the reliability experiments with different odometry errors are shown in Figure 16.
As shown in these results, the localization success rate of AGM-AMCL is significantly higher than that of AMCL and augmented-MCL under different odometry error in three scenes. The minimum success rate of AMCL, augmented-MCL and AGM-AMCL are 6%, 0% and 79%, respectively. Moreover, the localization success rate of AMCL and augmented-MCL decrease rapidly with the increase of odometry error in three scenes while the localization success rate of AGM-AMCL is always 100% in scene A and slightly decreases in scene B and C. The results of the reliability experiments show that our localization method is more reliable when locating a robot in ambiguous environments.

5.5. Localization Using Real-World Data

Based on the real-world laser scan data and the odometry data captured by our robot in scene A, B and C, the localization results of three localization methods are obtained, as shown in Figure 17. As we can see from these figures, AMCL and augmented-MCL failed to track the robot pose due to the fact that the registered laser scan data obtained by AMCL and augmented-MCL are inconsistent with the maps. On the contrary, the registered laser scan data obtained by our method is consistent with the maps and the resulting trajectories are closed successfully which are in line with the actual robot trajectories. The inconsistent laser scan data registration obtained by AGM-AMCL is caused by dynamic obstacles and localization error. From the trajectories generated by AGM-AMCL, it can be seen that some discontinuous points appeared on the neighboring locations of an ambiguous area and an unambiguous area. This is because the accumulated localization error was corrected by our localization method after the robot moving through an ambiguous area, which demonstrates that our localization method is more reliable in real-world ambiguous environments.

6. Conclusions

In this paper, an AGM-based AMCL method to achieve fast and reliable localization of a robot in ambiguous environments was presented. Our method was derived from an improved DBN which introduces two nodes to represent the accumulated ambiguity error and the AGM separately. The AGM models the ambiguity of an environment, and it can be used to predict the accumulated localization error during pose reasoning. Moreover, utilizing the AGM, a portal motion model was implemented to obtain more reliable pose prediction in an ambiguous area. We evaluated the effectiveness of the AGM and the AGM-AMCL method through simulation and real-world experiments. The results demonstrated that the proposed AGM is in line with human experience and our localization method can achieve fast and reliable localization in ambiguous environments, as compared with conventional methods.
In the future, we will improve the AGM to model the ambiguity in dynamic environments so as to address the localization problem in an environment where the ambiguity is not only caused by symmetrical or featureless areas but also by dynamic objects.

Author Contributions

Conceptualization, G.L.; Data curation, J.M., X.Z., L.J. and C.L.; Formal analysis, G.L.; Funding acquisition, Y.H.; Methodology, G.L.; Project administration, Y.H.; Resources, Y.H.; Software, G.L.; Supervision, Y.H.; Validation, J.M., Y.X. and L.J.; Visualization, X.Z. and C.L.; Writing—original draft, G.L.; Writing—review & editing, J.M. and Y.X.

Funding

This work was supported in part by Dongguan Innovative Research Team Program under Grant 201536000100031, in part by Science and Technology Planning Project of Guangdong Province under Grant 2017B090913001, in part by National Natural Science Foundation of China under Grant 51805190, in part by China Postdoctoral Science Foundation under Grant 2019M650179 and in part by Guangdong major science and technology project under Grant 2019B090919003.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Lajoie, P.; Hu, S.; Beltrame, G.; Carlone, L. Modeling Perceptual Aliasing in SLAM via Discrete-Continuous Graphical Models. IEEE Robot. Autom. Lett. 2019, 4, 1232–1239. [Google Scholar] [CrossRef]
  2. Tang, J.; Chen, Y.; Jaakkola, A. NAVIS-An UGV Indoor Positioning System Using Laser Scan Matching for Large-Area Real-Time Applications. Sensors 2014, 14, 11805–11824. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Pinto, A.M.; Moreira, A.P.; Costa, P.G. A localization method based on map-matching and particle swarm optimization. J. Intell. Robot. Syst. 2015, 77, 313–326. [Google Scholar] [CrossRef]
  4. Maffei, R.; Jorge, V.A.M.; Rey, V.F.; Kolberg, M.; Prestes, E. Fast Monte Carlo Localization using spatial density information. In Proceedings of the IEEE International Conference on Robotics and Automation, Seattle, WA, USA, 26–30 May 2015; pp. 6352–6358. [Google Scholar]
  5. Saarinen, J.; Andreasson, H.; Stoyanov, T.; Lilienthal, A.J. Normal distributions transform Monte-Carlo localization (NDT-MCL). In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems, Tokyo, Japan, 3–7 November 2013; pp. 382–389. [Google Scholar]
  6. Guan, R.P.; Ristic, B.; Wang, L.; Palmer, J.L. KLD sampling with Gmapping proposal for Monte Carlo localization of mobile robots. Inf. Fusion 2019, 49, 79–88. [Google Scholar] [CrossRef]
  7. Hanten, R.; Buck, S.; Otte, S.; Zell, A. Vector-AMCL: Vector based adaptive monte carlo localization for indoor maps. In Proceedings of the International Conference on Intelligent Autonomous Systems, Shanghai, China, 3–7 July 2016; pp. 403–416. [Google Scholar]
  8. Xu, S.; Chou, W.; Dong, H. A Robust Indoor Localization System Integrating Visual Localization Aided by CNN-Based Image Retrieval with Monte Carlo Localization. Sensors 2019, 19, 249. [Google Scholar] [CrossRef] [PubMed]
  9. Vasiljević, G.; Miklić, D.; Draganjac, I.; Kovačić, Z.; Lista, P. High-accuracy vehicle localization for autonomous warehousing. Robot. Comput. Integr. Manuf. 2015, 42, 1–16. [Google Scholar] [CrossRef]
  10. Li, G.; Meng, J.; Xie, Y.L.; Zhang, X.L.; Jiang, L.Q.; Huang, Y. An Improved Observation Model for Monte-Carlo Localization Integrated with Reliable Reflector Prediction. In Proceedings of the IEEE/ASME International Conference on Advanced Intelligent Mechatronics, Hong Kong, China, 8–12 July 2019. [Google Scholar]
  11. Meyer-Delius, D.; Beinhofer, M.; Kleiner, A.; Burgard, W. Using artificial landmarks to reduce the ambiguity in the environment of a mobile robot. In Proceedings of the IEEE International Conference on Robotics and Automation, Shanghai, China, 9–13 May 2011; pp. 5173–5178. [Google Scholar]
  12. Cheikhrouhou, O.; Bhatti, G.; Alroobaea, R. A Hybrid DV-Hop Algorithm Using RSSI for Localization in Large-Scale Wireless Sensor Networks. Sensors 2018, 18, 1469. [Google Scholar] [CrossRef] [PubMed]
  13. Zhang, L.; Zapata, R.; Lepinay, P. Self-adaptive Monte Carlo localization for mobile robots using range finders. Robotica 2012, 30, 229–244. [Google Scholar] [CrossRef]
  14. Oh, S.M.; Tariq, S.; Walker, B.N.; Dellaert, F. Map-based priors for localization. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems, Sendai, Japan, 28 September–2 October 2004; pp. 2179–2184. [Google Scholar]
  15. Censi, A. On achievable accuracy for range-finder localization. In Proceedings of the IEEE International Conference on Robotics and Automation, Roma, Italy, 10–14 April 2007; pp. 4170–4175. [Google Scholar]
  16. Censi, A. On achievable accuracy for pose tracking. In Proceedings of the IEEE International Conference on Robotics and Automation, Kobe, Japan, 12–17 May 2009; pp. 1–7. [Google Scholar]
  17. Liu, Z.; Chen, W.; Wang, Y.; Wang, J. Localizability estimation for mobile robots based on probabilistic grid map and its applications to localization. In Proceedings of the IEEE International Conference on Multisensor Fusion and Integration for Intelligent Systems, Hamburg, Germany, 13–15 September 2012; pp. 46–51. [Google Scholar]
  18. Irani, B.; Wang, J.; Chen, W. A Localizability Constraint-Based Path Planning Method for Autonomous Vehicles. IEEE Trans. Intell. Transp. Syst. 2018, 20, 2593–2604. [Google Scholar] [CrossRef]
  19. Pilania, V.; Gupta, K. Localization aware sampling and connection strategies for incremental motion planning under uncertainty. Auton. Robot. 2017, 41, 111–132. [Google Scholar] [CrossRef]
  20. Zhen, W.; Zeng, S.; Soberer, S. Robust localization and localizability estimation with a rotating laser scanner. In Proceedings of the IEEE International Conference on Robotics and Automation, Singapore, 29 May–3 June 2017; pp. 6240–6245. [Google Scholar]
  21. Akai, N.; Morales, L.Y.; Takeuchi, E.; Yoshihara, Y.; Ninomiya, Y. Robust localization using 3D NDT scan matching with experimentally determined uncertainty and road marker matching. In Proceedings of the IEEE Intelligent Vehicles Symposium, Los Angeles, CA, USA, 11–14 June 2017; pp. 6240–6245. [Google Scholar]
  22. Javanmardi, E.; Javanmardi, M.; Gu, Y.; Kamijo, S. Factors to Evaluate Capability of Map for Vehicle Localization. IEEE Access 2018, 6, 49850–49867. [Google Scholar] [CrossRef]
  23. Beinhofer, M.; Kretzschmar, H.; Burgard, W. Deploying artificial landmarks to foster data association in simultaneous localization and mapping. In Proceedings of the IEEE International Conference on Robotics and Automation, Karlsruhe, Germany, 6–10 May 2013; pp. 5235–5240. [Google Scholar]
  24. Chen, Y.; Francisco, J.; Trappe, W.; Martin, R.P. A practical approach to landmark deployment for indoor localization. In Proceedings of the 3rd Annual IEEE Communications Society Conference on Sensor Mesh and Ad Hoc Communications and Networks, Reston, VA, USA, 28–28 September 2006; pp. 365–373. [Google Scholar]
  25. Mi, J.; Takahashi, Y. Design of an HF-Band RFID System with Multiple Readers and Passive Tags for Indoor Mobile Robot Self-Localization. Sensors 2016, 16, 1200. [Google Scholar] [CrossRef]
  26. Liu, F.; Zhong, D. GSOS-ELM: An RFID-Based Indoor Localization System Using GSO Method and Semi-Supervised Online Sequential ELM. Sensors 2018, 18, 1995. [Google Scholar] [CrossRef]
  27. Zhang, Q.; Wang, P.; Chen, Z. An improved particle filter for mobile robot localization based on particle swarm optimization. Expert Syst. Appl. 2019, 135, 181–193. [Google Scholar] [CrossRef]
  28. Wolcott, R.W.; Eustice, R.M. Robust LIDAR localization using multiresolution Gaussian mixture maps for autonomous driving. Int. J. Robot. Res. 2017, 36, 292–319. [Google Scholar] [CrossRef]
  29. Blanco, J.L.; González, J.; Fernández-Madrigal, J.-A. A robust multi-hypothesis approach to matching occupancy grid maps. Robotica 2013, 31, 687–701. [Google Scholar] [CrossRef]
  30. Martín, F.; Carballeira, J.; Moreno, L.; Garrido, S.; González, P. Using the Jensen-Shannon, Density Power, and Itakura-Saito Divergences to Implement an Evolutionary-Based Global Localization Filter for Mobile Robots. IEEE Access 2017, 5, 13922–13940. [Google Scholar] [CrossRef]
  31. Park, S.; Roh, K.S. Coarse-to-Fine Localization for a Mobile Robot Based on Place Learning With a 2-D Range Scan. IEEE Trans. Robot. 2016, 32, 528–544. [Google Scholar] [CrossRef]
  32. Rottmann, A.; Mozos, O.M.; Stachniss, C.; Burgard, W. Place classification of indoor environments with mobile robots using boosting. In Proceedings of the National Conference on Artificial Intelligence, Pittsburgh, PA, USA, 9–13 July 2005; pp. 1306–1311. [Google Scholar]
  33. Gutmann, J.S.; Fox, D. An Experimental Comparison of Localization Methods Continued. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems, Lausanne, Switzerland, 30 September–4 October 2002; pp. 454–459. [Google Scholar]
  34. Thrun, S.; Fox, D.; Burgard, W.; Dellaert, F. Robust Monte Carlo localization for mobile robots. Artif. Intell. 2001, 128, 99–141. [Google Scholar] [CrossRef] [Green Version]
  35. Wang, J.; Wang, P.; Chen, Z. A novel qualitative motion model based probabilistic indoor global localization method. Inf. Sci. 2018, 429, 284–295. [Google Scholar] [CrossRef]
  36. Pfaff, P.; Plagemann, C.; Burgard, W. Gaussian mixture models for probabilistic localization. In Proceedings of the IEEE International Conference on Robotics and Automation, Pasadena, CA, USA, 19–23 May 2008; pp. 467–472. [Google Scholar]
  37. Thrun, S.; Burgard, W.; Fox, D. Probabilistic robotics; MIT Press: Cambridge, MA, USA, 2005; pp. 153–169. [Google Scholar]
  38. Fox, D. Adapting the Sample Size in Particle Filters through KLD-Sampling. Int. J. Robot. Res. 2003, 22, 985–1003. [Google Scholar] [CrossRef]
  39. Biswas, J.; Veloso, M.M. Episodic non-Markov localization. Robot. Autom. Syst. 2017, 87, 162–176. [Google Scholar] [CrossRef]
  40. Särkkä, S. Bayesian Filtering and Smoothing; Cambridge University Press: Cambridge, UK, 2013; pp. 120–125. [Google Scholar]
  41. Even, J.; Furrer, J.; Morales, Y.; Ishi, C.T.; Hagita, N. Probabilistic 3-D Mapping of Sound-Emitting Structures Based on Acoustic Ray Casting. IEEE Trans. Robot. 2017, 33, 333–345. [Google Scholar] [CrossRef]
  42. Hess, W.; Kohler, D.; Rapp, H.; Andor, D. Real-time loop closure in 2D LIDAR SLAM. In Proceedings of the IEEE International Conference on Robotics and Automation, Stockholm, Sweden, 16–21 May 2016; pp. 1271–1278. [Google Scholar]
  43. Corke, P. Robotics, Vision and Control: Fundamental Algorithms in MATLAB; Springer: Cham, Switzerland, 2011; pp. 32–40. [Google Scholar]
Figure 1. Ambiguous environments: (a) long corridor; (b) empty space; (c) tangled thicket.
Figure 1. Ambiguous environments: (a) long corridor; (b) empty space; (c) tangled thicket.
Sensors 19 03331 g001
Figure 2. Schematic diagram of M . The black grids and white grids indicate the occupied cells and free space cells, separately. The R d ( M ) and R θ ( M ) are marked on this figure. The position and the direction of the red arrow represent the pose x to be evaluated. Those blue arrows represent those poses derived from the incremental pose set M and x according to x + Δ i .
Figure 2. Schematic diagram of M . The black grids and white grids indicate the occupied cells and free space cells, separately. The R d ( M ) and R θ ( M ) are marked on this figure. The position and the direction of the red arrow represent the pose x to be evaluated. Those blue arrows represent those poses derived from the incremental pose set M and x according to x + Δ i .
Sensors 19 03331 g002
Figure 3. Main steps of AMCL and our method.
Figure 3. Main steps of AMCL and our method.
Sensors 19 03331 g003
Figure 4. Standard DBN for localization problem.
Figure 4. Standard DBN for localization problem.
Sensors 19 03331 g004
Figure 5. Improved DBN for localization problem. The AGM node ( m g ) and the cumulative ambiguity error node ( e t ) are introduced.
Figure 5. Improved DBN for localization problem. The AGM node ( m g ) and the cumulative ambiguity error node ( e t ) are introduced.
Sensors 19 03331 g005
Figure 6. Schematic diagram of the portal motion model. The white area, black area, and gray area represent an ambiguous area, an unambiguous area, and area occupied by obstacles, respectively. x 1 is a pose sample. The blue transparent circle is the Gaussian distribution N ( μ h i t , Σ h i t ) , and the blue opaque portion expresses the truncated Gaussian distribution of Equation (17).
Figure 6. Schematic diagram of the portal motion model. The white area, black area, and gray area represent an ambiguous area, an unambiguous area, and area occupied by obstacles, respectively. x 1 is a pose sample. The blue transparent circle is the Gaussian distribution N ( μ h i t , Σ h i t ) , and the blue opaque portion expresses the truncated Gaussian distribution of Equation (17).
Sensors 19 03331 g006
Figure 7. Changing process of Σ h i t . The uncertainty of N ( μ h i t , Σ h i t ) is increasing as a robot moves in an ambiguous area and more particles will be generated by the portal motion model.
Figure 7. Changing process of Σ h i t . The uncertainty of N ( μ h i t , Σ h i t ) is increasing as a robot moves in an ambiguous area and more particles will be generated by the portal motion model.
Sensors 19 03331 g007
Figure 8. Experimental platform and real-world environments: (a) Turtlebot2 for data capturing; (b) scene A: long corridors; (c) scene B: empty space; (d) scene C: tangled thicket. The red lines in scene A show three parallel long corridors, and each corridor is surrounded by two parallel walls without prominent features. The red square in scene B shows an open square. The red square in scene C show area with tangled thicket.
Figure 8. Experimental platform and real-world environments: (a) Turtlebot2 for data capturing; (b) scene A: long corridors; (c) scene B: empty space; (d) scene C: tangled thicket. The red lines in scene A show three parallel long corridors, and each corridor is surrounded by two parallel walls without prominent features. The red square in scene B shows an open square. The red square in scene C show area with tangled thicket.
Sensors 19 03331 g008
Figure 9. AGMs of virtual ambiguous environments: (a1c1) are the occupancy grid maps of three artificial environments. (a2c2) are the corresponding AGMs. AGMs show that the middle area of (a1), the center area of (b1) and the open area of (c1) get high AAE value because of their ambiguous.
Figure 9. AGMs of virtual ambiguous environments: (a1c1) are the occupancy grid maps of three artificial environments. (a2c2) are the corresponding AGMs. AGMs show that the middle area of (a1), the center area of (b1) and the open area of (c1) get high AAE value because of their ambiguous.
Sensors 19 03331 g009
Figure 10. AGMs with different range error for virtual unambiguous environments: (a) Occupancy grid map of an unambiguous environment. (b) AGM with zero distance error. (c) AGM with 0.5 m distance error. The figure clearly shows that large distance error leads to large AAE value.
Figure 10. AGMs with different range error for virtual unambiguous environments: (a) Occupancy grid map of an unambiguous environment. (b) AGM with zero distance error. (c) AGM with 0.5 m distance error. The figure clearly shows that large distance error leads to large AAE value.
Sensors 19 03331 g010
Figure 11. AGMs of real-world environments: (a1c1) are the occupancy grid maps of scene A, B and C. (a2c2) are the corresponding AGMs. Blue curves show the ground truth for localization simulation. Those positions in the middle of a long corridor, in the center of an open square, near a straight-line wall and tangled thicket get high AAE value.
Figure 11. AGMs of real-world environments: (a1c1) are the occupancy grid maps of scene A, B and C. (a2c2) are the corresponding AGMs. Blue curves show the ground truth for localization simulation. Those positions in the middle of a long corridor, in the center of an open square, near a straight-line wall and tangled thicket get high AAE value.
Sensors 19 03331 g011
Figure 12. Evolution of particles using AMCL (top), Augmented-MCL (middle) and AGM-AMCL (bottom) in scene B. The letters a, b, c and d denote four locations shown in Figure 11(b1) with red dots. The blue square with a red triangle inside represents the true pose and the colored points denote particles. With some particles in the nearby unambiguous area, our method can converge to the true pose after a robot moving through an ambiguous area.
Figure 12. Evolution of particles using AMCL (top), Augmented-MCL (middle) and AGM-AMCL (bottom) in scene B. The letters a, b, c and d denote four locations shown in Figure 11(b1) with red dots. The blue square with a red triangle inside represents the true pose and the colored points denote particles. With some particles in the nearby unambiguous area, our method can converge to the true pose after a robot moving through an ambiguous area.
Sensors 19 03331 g012
Figure 13. Localization results in scene A: (a) distance error; (b) orientation error; (c) number of particles; (d) running time. Note that the orientation errors of AMCL are small in most times. This is because the orientation can be easily determined in the long corridor while it will cause large distance error. The running time of our method is similar to AMCL in unambiguous areas (the zoomed result from step 100 to 150). Our method is more efficient than AMCL after the robot moving through an ambiguous area (the zoomed result from step 450 to 500). Our method needs more particles when the robot is in an ambiguous area and near an unambiguous area (step 300 to 400), therefore it is less efficient than AMCL. Because of the fixed particle number, augmented-MCL takes more time than the other two methods in most time.
Figure 13. Localization results in scene A: (a) distance error; (b) orientation error; (c) number of particles; (d) running time. Note that the orientation errors of AMCL are small in most times. This is because the orientation can be easily determined in the long corridor while it will cause large distance error. The running time of our method is similar to AMCL in unambiguous areas (the zoomed result from step 100 to 150). Our method is more efficient than AMCL after the robot moving through an ambiguous area (the zoomed result from step 450 to 500). Our method needs more particles when the robot is in an ambiguous area and near an unambiguous area (step 300 to 400), therefore it is less efficient than AMCL. Because of the fixed particle number, augmented-MCL takes more time than the other two methods in most time.
Sensors 19 03331 g013
Figure 14. Localization results in scene B: (a) distance error; (b) orientation error; (c) number of particles; (d) running time. From step 700 to 800, our method is more efficient because of the right convergence. The zoomed results from step 0 to 80 show that our localization method has similar running time with AMCL in unambiguous area.
Figure 14. Localization results in scene B: (a) distance error; (b) orientation error; (c) number of particles; (d) running time. From step 700 to 800, our method is more efficient because of the right convergence. The zoomed results from step 0 to 80 show that our localization method has similar running time with AMCL in unambiguous area.
Sensors 19 03331 g014aSensors 19 03331 g014b
Figure 15. Localization results in scene C: (a) distance error; (b) orientation error; (c) number of particles; (d) running time. The zoomed results from step 50 to 150 show that our localization method has similar running time with AMCL. Our method takes less time from step 600 to 800 while more time is need from step 400 to 550.
Figure 15. Localization results in scene C: (a) distance error; (b) orientation error; (c) number of particles; (d) running time. The zoomed results from step 50 to 150 show that our localization method has similar running time with AMCL. Our method takes less time from step 600 to 800 while more time is need from step 400 to 550.
Sensors 19 03331 g015
Figure 16. The success rate of three localization methods with different odometry errors: (a) Results in Scene A. (b) Results in Scene B. (c) Results in Scene C. The results show that our method has a considerable success rate even in the case when the odometry error is high in ambiguous environments.
Figure 16. The success rate of three localization methods with different odometry errors: (a) Results in Scene A. (b) Results in Scene B. (c) Results in Scene C. The results show that our method has a considerable success rate even in the case when the odometry error is high in ambiguous environments.
Sensors 19 03331 g016
Figure 17. Localization results using real-world data. (a1c1) show the localization result using AMCL. (a2c2) show the localization result using augmented-MCL. (a3c3) show the localization result using our method. Blue curves show the robot trajectories obtained by localization methods. Red point clouds represent the registered laser scan data using the trajectories. Based on the trajectories obtained by our method, laser scan data can be successfully registered in three ambiguous environments. The discontinuity of the trajectories obtained by our method is caused by the portal motion model.
Figure 17. Localization results using real-world data. (a1c1) show the localization result using AMCL. (a2c2) show the localization result using augmented-MCL. (a3c3) show the localization result using our method. Blue curves show the robot trajectories obtained by localization methods. Red point clouds represent the registered laser scan data using the trajectories. Based on the trajectories obtained by our method, laser scan data can be successfully registered in three ambiguous environments. The discontinuity of the trajectories obtained by our method is caused by the portal motion model.
Sensors 19 03331 g017

Share and Cite

MDPI and ACS Style

Li, G.; Meng, J.; Xie, Y.; Zhang, X.; Huang, Y.; Jiang, L.; Liu, C. Reliable and Fast Localization in Ambiguous Environments Using Ambiguity Grid Map. Sensors 2019, 19, 3331. https://doi.org/10.3390/s19153331

AMA Style

Li G, Meng J, Xie Y, Zhang X, Huang Y, Jiang L, Liu C. Reliable and Fast Localization in Ambiguous Environments Using Ambiguity Grid Map. Sensors. 2019; 19(15):3331. https://doi.org/10.3390/s19153331

Chicago/Turabian Style

Li, Gen, Jie Meng, Yuanlong Xie, Xiaolong Zhang, Yu Huang, Liquan Jiang, and Chao Liu. 2019. "Reliable and Fast Localization in Ambiguous Environments Using Ambiguity Grid Map" Sensors 19, no. 15: 3331. https://doi.org/10.3390/s19153331

APA Style

Li, G., Meng, J., Xie, Y., Zhang, X., Huang, Y., Jiang, L., & Liu, C. (2019). Reliable and Fast Localization in Ambiguous Environments Using Ambiguity Grid Map. Sensors, 19(15), 3331. https://doi.org/10.3390/s19153331

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop