Next Article in Journal
Mapping Wildfire Ignition Probability Using Sentinel 2 and LiDAR (Jerte Valley, Cáceres, Spain)
Next Article in Special Issue
An RFID-Based Smart Nest Box: An Experimental Study of Laying Performance and Behavior of Individual Hens
Previous Article in Journal
A Data Cleaning Method for Big Trace Data Using Movement Consistency
Previous Article in Special Issue
A Novel Passive Wireless Sensing Method for Concrete Chloride Ion Concentration Monitoring
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Moving Object Localization Based on UHF RFID Phase and Laser Clustering

1
School of Information Engineering, Southwest University of Science and Technology, Mianyang 621010, China
2
Engineering Product Development, Singapore University of Technology and Design, Singapore 487372, Singapore
3
Department of Computer Science, Lasbela University of Agriculture, Water and Marine Sciences, Balochistan 90150, Pakistan
*
Author to whom correspondence should be addressed.
Sensors 2018, 18(3), 825; https://doi.org/10.3390/s18030825
Submission received: 16 January 2018 / Revised: 2 March 2018 / Accepted: 6 March 2018 / Published: 9 March 2018
(This article belongs to the Special Issue RFID-Based Sensors for IoT Applications)

Abstract

:
RFID (Radio Frequency Identification) offers a way to identify objects without any contact. However, positioning accuracy is limited since RFID neither provides distance nor bearing information about the tag. This paper proposes a new and innovative approach for the localization of moving object using a particle filter by incorporating RFID phase and laser-based clustering from 2d laser range data. First of all, we calculate phase-based velocity of the moving object based on RFID phase difference. Meanwhile, we separate laser range data into different clusters, and compute the distance-based velocity and moving direction of these clusters. We then compute and analyze the similarity between two velocities, and select K clusters having the best similarity score. We predict the particles according to the velocity and moving direction of laser clusters. Finally, we update the weights of the particles based on K clusters and achieve the localization of moving objects. The feasibility of this approach is validated on a Scitos G5 service robot and the results prove that we have successfully achieved a localization accuracy up to 0.25 m.

1. Introduction

Recent documents show a growing interest in indoor localization due to the high demand of location-based services (LBS) [1,2], for example asset tracking and indoor guidance. Extensive research have been done regarding vision-based and range-based sensors. Vision-based localization is challenging as it needs the object to be in the line of sight and also requires complex algorithms to recognize the object. The advantages of RFID (Radio Frequency Identification) provide a way to overcome these difficulties, as RFID has some unique qualities like an RFID tag has a unique identification code (ID), which does not require any complex recognition algorithm [3]. Secondly, compared to vision or range-based sensors, RFID can solve the occlusion issue of the object as the communication of the RFID does not to be in the line of sight. These characteristics make it convenient to use RFID for positioning of an object. In addition, RFID positioning also has the advantages of small size and fast position deduction, as compared to Bluetooth or Wi-Fi positioning techniques [4]. Therefore, RFID positioning technology is widely used in warehousing, library, supermarket, logistics assembly, and other places [5,6]. For example, Shaari et al. proposed an intelligent storage system based on RFID that uses a database to store the position and orientation of an object [7]. Lu et al. designed an RFID-based positioning system for an AGV (Automated Guided Vehicle) in a smart factory [8]. Fortinsimard et al. implemented a smart identification system based on passive RFID technology for smart homes [9]. Nur et al. designed a system that provides real-time information about the product by projecting RFID and shelf information onto a panoramic view [10]. Wang et al. proposed an RFID-based positioning approach for connected vehicles [11].
RFID systems also have some limitations like RFID can not directly obtain the distance or bearing of the tag, so one has to rely on other information to achieve the positioning task. Traditional ways of localization by RFID are commonly based on received signal strength (RSS) with the limitation of multi-path propagation issues in indoor environment, which could lead to poor localization accuracy and demands excessive training to build the model [12]. Hähnel et al. used a mobile robot to measure the signal strength of tags at different positions and constructed a probabilistic sensor model of RFID antenna to locate passive RFID tags; however, this method needs to calibrate the RFID sensor model by installing RFID tags in a well-known environment, and it is obviously time consuming [13]. Although the method used in LANDMARC gives comparatively high accuracy, it requires multiple antennas to detect the phase, which is not practical for mobile robots [14]. Many researchers have tried different ways to improve the localization accuracy based on RSS. For example, Xu et al. used the Bayesian approach and K-Nearest Neighbor to filter out the noise based on LANDMARC with a limitation of a large number of reference tag requirements in advance [15]. Yang et al. applied an improved particle filter based on the features of moving direction in a densely tagged RFID environment, but it requires a long time to build the system model [16]. Moreover, the signal’s time-of-arrival (TOA) and angel-of-arrival (AOA) can also be used for localization. Xu et al. used a TOA-based method to track the position of moving people [17]. TOA has an advantage of a simple positioning principle with small errors, but it requires strict time synchronization and high cost hardware. AOA is vulnerable to environmental factors and has limited applications [4,18]. In addition, the phase information provided by the RFID reader can be used to locate the object directly, without any explicit modeling of signal strength. However, there is an ambiguity problem in the phase extraction [19], which increases the positioning error. Sarkka et al. proposed a UHF RFID tracking system based on the phases from multiple spatially distributed antennas [20]. Some researchers also have tried to combine the RSS and phase for the localization of an object. For example, Ma et al. [21] proposed an approach to predict the position of a mobile object using the RSS and phase. Martinelli et al. proposed using the unscented Kalman filter based on the RSS and phase to locate the robot, but it requires a group of passive tags on the ceiling as landmarks [22].
In literature, RFID is also used to combine with other sensors (for example, laser range finders and visual cameras) to improve the positioning accuracy [23,24,25]. For example, Deyle et al. used visual and three-dimensional laser ranging information to construct images and combined signal strength of different angles by rotating the antenna to deduct the location of an object [26]. RFID technology is gradually applied to the field of construction [27]. Valero et al. used a 3D laser scanner and RFID technology for automatic construction of the 3D model of an environment. The fusion technique allows for identifying the main components of a structure, such as walls, floor, and windows [28]. However, this approach requires the modeling of both sensors. Zhou et al. used a new positioning system based on artificial landmarks for mobile robots. They used a laser-activated RFID tag (active tag) that has a bright LED as landmark for the positioning. The airborne laser sensor emits a laser beam to activate the tag when the robot is moving in the environment. At the same time, the robot detects the tag in the environment through the stereo vision and RFID reader, and calculates the relative position of the tag through the perspective geometry. They combined the information from multiple tags to determine the position and orientation of the robot [29]. However, this method requires the tag to be installed in the test environment in advance and the maintenance cost is high. Martin et al. used combined signals (WiFi and Digital Television) to reduce the error [30]. Liu et al. used a three-dimensional sensor model and a pair of RFID antennas at different heights to localize the tags in 3D [31]. Shirehjini et al. used RFID carpets and several peripherals of sensor to build a positioning system based on low-range passive RFID technology [32]. However, this approach requires installing RFID tags on the floor. Song et al. used a two-level extended Kalman filter (EKF) to fuse the data of sensors mounted in vehicles and RFID data for positioning of a vehicle in GPS-denied environment [33]. Xiong et al. presented a hybrid WSN-RFID system for tracking indoor objects [34]. Su et al. proposed an enhanced boundary condition approach to improve the localization accuracy of an RFID system by integrating the information from a GPS system [35].
Due to their long range, high accuracy, and fast response, laser range finders are widely used in obstacle avoidance, localization, and navigation in robotics [36,37,38]. An et al. proposed an approach that allows a robot to detect and avoid obstacles using a Vector Polar Histogram (VPH) [39]. Sun et al. proposed a noise reduction algorithm to improve ranging accuracy of a laser range finder [40]. With a laser range finder, one can directly obtain the distance and angle information of the obstacles [41,42]. For example, Vu et al. used a laser range finder to create a raster map of the environment that can be used to detect moving objects [43]. Dietmayer et al. used a model-based approach to identify and track objects in the environment [44]. However, laser-based sensors need to establish a model and complex recognition algorithms to identify the object, which usually requires a pre-training stage and expensive computational time [44,45].
The uniqueness of RFID tags makes it possible to recognize an object without any additional recognition stage like a laser range finder or visual camera always needs. Therefore, we propose an approach to combine UHF RFID technology and laser ranging information to localize a moving object. RFID tags have unique IDs that can be used to identify an object. The phase reported from the RFID is a periodic function of the distance and has the ambiguity in positioning of an object [46,47]. The phase difference can be used to infer the moving velocity of an object. On the other hand, the laser range finder can be used to measure distance and angle of the obstacles. However, it needs to recognize the object that we want to localize from a large number of objects in the laser view. The integration of RFID and laser information can lead to a more reliable and accurate positioning system. In particular, we achieve the position deduction of moving object by the fusion of RFID phase and laser clustering. We first estimate the phase-based velocity by phase difference of RFID signals. In the meantime, we segment 2d laser ranging data into clusters and estimate their velocities (i.e., distance-based velocities) and moving directions. Then, we calculate the similarity between the phase-based velocity and the distance-based velocities to analyze the similarity of both velocities. Finally, we predict and update the set of particles in a particle filter according to matching results and accomplish the localization of a moving object. Our approach does not need any signal strength information from RFID, nor does it need any modeling of the object from a laser range finder.
We organize the subsequent sections of this paper as follows. An overview of the system is described in Section 2. In Section 3, we present the details of the sensor fusion using a particle filter. We show the experimental results in Section 4 and conclude the paper with possible extensions in Section 5.

2. System Overview

Our approach for localization is based on the particle filter that integrates the RFID phase and laser clustering, without any modeling on signal strength. As shown in Figure 1, the whole system consists of two major parts: (1) collecting information by an RFID reader and a laser range finder on the robot; (2) fusing the RFID and laser ranging information in a particle filter for localization.
In the first part, we use the RFID reader to obtain the RFID phase of a moving object affixed with a passive tag and then calculate the phase-based velocity of the moving object by the phase difference between two successive RFID signals. It is important to note that there is a phase ambiguity during the processing of RFID phase. Therefore, we need to take measures to reduce the influence of phase ambiguity in order to estimate the phase-based velocity. After several experimental tests, we found that the phase difference of a moving object between two adjacent moments increases sharply, but the difference does not exceed ± 90 . Hence, this check helps us to reduce phase ambiguity and we can easily calculate corresponding velocity from phase difference. Meanwhile, we scan the moving object with a 2D laser range finder carried by the robot and segment the ranging data into clusters. After that, we find the neighboring clusters (two neighbors) in the continuous time and treat the two clusters as the same obstacle, which is used to estimate the distance-based velocity. This part of information acquisition and data processing is used for subsequent velocity matching and the particle filtering.
In the other part, we first calculate the similarity between phase-based velocity and the distance-based velocity and then determine the velocity matching according to a defined similarity rule. In addition, we choose the best K clusters with the best velocity similarity score to facilitate the processing of the particle filtering. To improve localization accuracy and robustness of the system, we use different prediction methods for the particle filter based on the velocity matching results: (1) If the velocity matching is successful, we can find the effective laser cluster for prediction, which is referred to as laser prediction in this paper. (2) If the velocity matching is unsuccessful, we choose another method called the random prediction method. After that, we update the particle’s weight in the particle filter using the best K clusters. Finally, we localize the moving object by iteratively performing prediction, update, and resampling of the particle filter. The role of this part is to improve the positioning efficiency and accuracy by complementing the RFID phase with laser ranging information using a particle filter.

3. Moving Object Localization Based on the Particle Filtering

We describe the details of the system in this section. In particular, the computation of RFID phase velocity is described in Section 3.1, the clustering of laser ranging data is presented in Section 3.2, the estimation of the velocity and moving direction of a cluster is detailed in Section 3.3, and velocity matching and the implementation using a particle filter are detailed in Section 3.4 and Section 3.5, respectively. The mathematical symbols and their meanings used in this paper are listed in Table 1.

3.1. Computing RFID Phase-Based Velocity

The phase obtained by RFID is a periodical function, which can be described as:
φ t = 2 π · d t λ · m o d ( 2 π ) ,
where φ t is the signal phase at time t, λ is the wavelength of the receiving signal, and d t is the distance from RFID tag to the antenna. In practice, signal phase is influenced by the transmitter, receiver, and tag’s reflection characteristics [30], which introduce additional phase rotation. Therefore, the phase of RFID signal could be modeled as:
φ t = 4 π · d t λ + φ T + φ R + φ T a g ,
where φ T is the reader’s transmitter phase deflection, φ r is the reader’s receiver phase deflection, and φ T a g is the tag’s phase deflection.
In this paper, the moving object is affixed with a passive tag and moves at a velocity of v t r , the signal phase being received by the RFID reader continuously. We denote the phases at t 1 and t as φ t 1 and φ t , respectively; therefore, we can estimate the moving velocity of the object by the phase difference:
v t r = Δ d t Δ t = d t d t 1 Δ t .
Combined with Equation (3), moving velocity of a tag can be given as:
v t r = c 4 π f Δ t · ( φ t φ t 1 ) , φ t φ t 1 < π , c 4 π f Δ t · ( φ t φ t 1 2 π ) , φ t φ t 1 π , c 4 π f Δ t · ( φ t φ t 1 + 2 π ) , φ t φ t 1 π .
In this way, the influence of phase ambiguity is effectively reduced, where c is the velocity of light, f is signal frequency, and Δ d t is the moving distance of the tag at two successive times, which is less than the half of wavelength (i.e., Δ d t < λ 2 ).

3.2. Clustering Laser Ranging Data

A 2D laser range finder is used to scan the environment and retrieve the distance and angle information of the moving object. We only save the laser ranging data when the reader detects the tag we want to track. Each scan point consists of a distance and angle measurement. In total, there are 450 ranging points in a scan cycle. In one scanning cycle, ( ρ n , ϕ n ) represents the distance and angle of the nth laser beam [48,49] and can be described into Cartesian coordinate system:
x n = ρ n · s i n ϕ n , y n = ρ n · c o s ϕ n ,
where ( x n , y n ) denotes the laser point P n . The distance between the ith and jth laser point can be described as:
d ( P i , P j ) = ( x i x j ) 2 + ( y i y j ) 2 .
For a better localization of the object, we segment the laser ranging data into different clusters. The process of clustering mainly includes: segmenting the raw data into groups (Section 3.2.1), splitting the group into clusters (Section 3.2.2), merging clusters (Section 3.2.3), and filtering out the ones with larger radii (Section 3.2.3). The final clustering results are considered as the obstacles [41,48,49]. The whole process of clustering is shown in Figure 2. Figure 3 is the raw laser ranging data at two nearby timestamps. As it can be seen from this figure, there are many obstacles in one environment. Therefore, we use the distance between two consecutive laser points for simple grouping, which will segment the laser ranging data into a group (the potential objects). We then split the group to improve the clustering accuracy.

3.2.1. Grouping of Laser Ranging Data

To perform clustering procedure, original laser ranging data is segmented into groups with the following rules: when the distance d ( P i , P i 1 ) between ith and the i 1 th laser point satisfies Equation (7), we consider the two points as the same group G k :
d ( P i , P i 1 ) < d g + d i · d p ,
where d g is the threshold of grouping, d i is the distance between the ith laser point and robot, and d p is the distance parameter. When the continuous laser points (i.e., P i and P i 1 ) in one scanning cycle do not satisfy Equation (7), a new group G k + 1 will be created based on point P i . This step will lead to a grouping set G k ( k { 1 N g } ) , where N g is the number of groups. An example of the grouping result is shown in Figure 4.

3.2.2. Splitting of the Group

The grouping set G k ( k = 1 N g ) is obtained by comparing the distance between the neighboring laser points, but two or more obstacles may exist in one group. Thus, we need to split the group in order to separate the obstacles. For this purpose, we use an iterated fitting algorithm. First, we find the two points in group G k with the longest distance and obtain the line connecting these two points. We then find the point with the longest perpendicular distance Δ d k to the line. If Δ d k satisfies Equation (8), then it continues the splitting process or otherwise quits the iteration. The splitting condition is defined as:
Δ d k > d s + d r · d p ,
where d s is the threshold for splitting, and d r is length of the connecting line. If G k satisfies the splitting condition, we split G k into new groups as G k and G k + 1 , as shown in Figure 5a,b. The iterated splitting process stops until there is no increase in the number of groups, We continue the split process until the number of clusters does not change. As a result, we obtain a number of clusters at time t:
O t = { O t ( i ) } = { r t ( i ) , o t ( i ) } i = 1 N t ,
where r t ( i ) = d r 2 is the radius of ith clustering, o t ( i ) = ( x t ( i ) , y t ( i ) ) is the center of ith cluster, and N t is the number of clusters at time t.

3.2.3. Merging and Filtering

To avoid the overlapping between different clusters, we further merge the clusters with the following criteria:
| r t ( i ) r t ( j ) | ( x t ( i ) x t ( j ) ) 2 + ( y t ( i ) y t ( j ) ) 2 .
If Equation (10) is satisfied, the cluster O t ( i ) and O t ( j ) would be merged into one cluster. As our target is to localize common pedestrians, the radius of a cluster is confined by a threshold r m a x . Finally, we use an example to demonstrate the process of laser clustering. Figure 3 is the raw laser scanning data at two consecutive timestamps, and Figure 5c is the clustering results.

3.3. Estimate the Distance-Based Velocity and Moving Direction of a Cluster

For each cluster O t ( i ) at time t, we find the cluster O t 1 ( j ^ ) with a minimum distance at the previous time t 1 . Therefore, ( O t ( i ) , O t 1 ( j ^ ) ) could be considered as the same object at two sequent times t and t 1 . j ^ is given by:
j ^ = a r g min j ( x t ( i ) x t 1 ( j ) ) 2 + ( y t ( i ) y t 1 ( j ) ) 2 ,
where 1 j N t 1 . Therefore, the velocity and moving direction of the ith cluster can be computed as:
v t ( i ) = Δ d t ( i ) Δ t = ( x t ( i ) x t 1 ( j ^ ) ) 2 + ( y t ( i ) y t 1 ( j ^ ) ) 2 Δ t ,
θ t ( i ) = a t a n ( y t ( i ) y t 1 ( j ^ ) x t ( i ) x t 1 ( j ^ ) ) .
To improve the accuracy of algorithm, we filter out the clusters having velocity higher than 1.0 m/s.

3.4. Similarity Computation Using Phase-Based and Distance-Based Velocity

Although the laser range finder can be used to obtain the distance and angle information, there is a singularity problem on the object identification. Due to the unique identification feature, RFID can precisely solve the singularity issue. We compare the similarity between the RFID phase-based velocity and the laser clustering-based velocity to evaluate whether both of them are from the same object or not. The similarity is computed as:
s i m ( v t ( i ) , v t r ) = 1 | v t ( i ) v t r | | v t ( i ) + v t r | .
In Equation (14), we can observe that, with a high similarity, there is more chance that the corresponding laser cluster is the object we want to localize. To improve the robustness of the system, we select the K clusters (i.e., O t m ( 1 ) , O t m ( 2 ) , O t m ( K ) ) with best similarity values as the potential objects for updating weights of particles, which is described in Section 3.5.

3.5. Moving Object Localization with a Particle Filter Based on the K Best Clusters

In this section, we estimate the position of an object (i.e., posterior probability) by the probabilistic-based approach. The Bayesian inference is a probabilistic framework for estimating the probability density function over a state, given the past measurements reported by the sensor and the motions performed by the object. In our case, we use RFID and laser ranging measurement to estimate the state of the object, i.e., posterior probability density function of the state. According to Bayesian inference, we factorize the posterior probability p ( X t | g 1 : t , r 1 : t , u 1 : t ) into:
p ( X t | g 1 : t , r 1 : t , u 1 : t ) = η t · p ( X t | X t 1 , u t ) · p ( g t | X t , r t ) · p ( X t 1 | g 1 : t 1 , r 1 : t 1 , u 1 : t 1 ) ,
where X t is the position of the object, which we want to estimate at time t, g t is the measurement of the laser range finder at time t, r t is the measurement of RFID at time t, u t is the motion information of the object, and η t is a normalizer. p ( X t | X t 1 , u t ) is the motion model, which is used to predict the object position at time t given the previous position X t and motion information u t . p ( g t | X t , r t ) is the observation model, which describes the likelihood of receiving a measurement g t (i.e., laser-based clusters) given the RFID measurement (i.e., RFID phase-based velocity) and the current state X t . Details of the motion model and observation model will be explained in Section 3.5.1 and Section 3.5.2, respectively. p ( X t 1 | g 1 : t 1 , r 1 : t 1 , u 1 : t 1 ) is the state at time t 1 . To solve this nonlinear problem, we choose the particle filter as the implementation, due to its no-parametric features. Particle filter is a basic implementation of Bayesian framework, which uses a set of particles to approximate the posterior probability distribution. It has the advantages of high robustness and high accuracy to deal with nonlinear and non-Gaussian systems. In this paper, particle filter is used to fuse two kinds of information for positioning.
We represent the object location by a number of particles with different weights, i.e., X t = { X t [ n ] , ω t [ n ] } n = 1 N , where N is the number of particles, X t [ n ] = { x t [ n ] , y t [ n ] } denotes the 2D location of the particle, and ω t [ n ] is the associated weight [50]. The particle filter performs the following three steps based on the arrival of the sensor measurements: prediction, update, and resampling.

3.5.1. Prediction

In this stage, we predict the state of a particle X t [ n ] based on the previous state X t 1 [ n ] and the motion information of the object u t . Here, to enhance the prediction performance, we use two predicting forms: random prediction and laser-based prediction. Random prediction is adopted when a matching cluster is not found, which is described as :
x t [ n ] = x t 1 [ n ] + N ( 0 , v t r · Δ t · σ r ) , y t [ n ] = y t 1 [ n ] + N ( 0 , v t r · Δ t · σ r ) ,
where σ r represents the random Gaussian noise. Random prediction can be considered as the complementary measurement when laser-based prediction is not available.
We use laser-based prediction when a valid matching cluster is found. In this way, we first find the nearest cluster l of particle n, then add Gaussian noise to its corresponding velocity v t ( l ) and moving direction θ t ( l ) . Finally, we execute the prediction according to v t ( l ) ˜ and θ t ( l ) ˜ :
θ t ( l ) ˜ = θ t ( l ) + N ( 0 , σ a ) ,
v t ( l ) ˜ = v t ( l ) + N ( 0 , σ v ) ,
x t [ n ] = x t 1 [ n ] + v t ( l ) ˜ · Δ t · c o s ( θ t ( l ) ˜ ) , y t [ n ] = y t 1 [ n ] + v t ( l ) ˜ · Δ t · s i n ( θ t ( l ) ˜ ) ,
where σ a and σ v represent the Gausian noises added to the velocity and the moving direction of the cluster l, respectively. In some cases where RFID and laser velocities do not match during positioning (when no matching cluster can be found), we use Equation (16) for prediction; otherwise, Equation (19) is used for prediction. Using two different prediction methods can improve the positioning accuracy.

3.5.2. Update

The update stage is used to correct previous prediction by current measurement and update the weights of the particles. Based on the observation model p ( g t | X t , r t ) , the weight ω t [ n ] of the particle X t [ n ] is computed as:
ω t [ n ] = η t · ω t 1 [ n ] · p ( g t | X t [ n ] , r t ) .
Combined with the weighted K most similar clusters, the observation model p ( g t | X t , r t ) is approximated as:
p ( g t | X t , r t ) = i = 1 K s i m ( v t m ( i ) , v t r ) · exp ( d 2 ( X t [ n ] , O t m ( i ) ) 2 ) ,
where
d 2 ( X t [ n ] , O t m ( i ) ) = ( x t [ n ] x t m ( i ) ) 2 σ d + ( y t [ n ] y t m ( i ) ) 2 σ d ,
where σ d is a bandwidth parameter added to the distance to the cluster. The impact of σ d on the positioning accuracy is analyzed in the experimental section (Section 4.2.6).

3.5.3. Resampling

We resample the particles to ensure the efficiency of particle filter. Resampling generates a set of new particles depending on the weights of the particles. The weights of the particles change after the particles are updated, in which the particles that are far away from the target have the smaller weight and the ones closer the target have the larger weight. After many iterations, some particle weights become very small, which is the phenomenon of particle degeneration. In order to effectively reduce the phenomenon of particle degeneration, we use resampling to remove the less weighted particles and duplicate the particles with large weights.

4. Experimental Results

4.1. Experimental Setups

The experiments to check the feasibility of our approach were performed on a service robot SCITOS G5 from Metralabs GmbH, Ilmenau, Germany), as shown in Figure 6. The robot is equipped with a 2D laser range finder (SICK S300), a UHF RFID reader (Speedway Revolution R420 from Impinj, Inc., Seattle, WA, USA) with a sampling frequency of 2 Hz, and two circularly polarized antennas (RFMAX SS8688P from Laird Technologies, London, UK) at two sides of the robot with angles of 45 . The RFID reader offers a reading range up to 7 m. During the experiment, we use Dense Reader Mode 8 (DRM8) and a channel of 920.625 MHz for the RFID reader. The RFID antenna is placed at the same height as the tag (1.2 m above the ground). The measuring range for the laser range finder is up to 29 m, within the angle of 270 and resolution of 0.5 . In addition, the laser range finder works at a frequency of 20 HZ. During the experiments, we affixed a tag (Alien Squiggle RFID Wet Inlay from Alien Technology, San Jose, CA, USA) on a moving object.
We set the testing space for positioning by a rectangular area of 4 m × 2 m, and the robot was placed one meter perpendicular to the long side of the rectangle, as shown in Figure 6a. A person carried the tag and walked along the edge of rectangular area for 5 rounds at the velocity of approx. 0.4 m/s to test the positioning accuracy.
The comparison between the estimated path and the true path is depicted in Figure 6b. In addition, we walked the more complex paths in the test area to test the usability of our approach. The results are shown in Figure 7a,b. In order to facilitate comparison, we use the rectangular path as the adjustment of the experimental parameters. From these figures, we can see that the estimated path is basically consistent to the true path, which demonstrates the feasibility of our approach. The final result shows that mean positioning error is 0.25 m. Average positioning error means the RMSE (rooted mean square error). The error is defined as the Euclidean distance between the ground truth and the estimated position.

4.2. Impact of Different Parameters on the Positioning Accuracy

Next, we carried out six different experiments for evaluating the positioning accuracy under the impact of different factors in each experiment, including antenna configurations, predicting forms, factors in clustering (i.e., d g , d p , d s , and r m a x ), the number of particles N, noise settings (i.e., σ d and σ a ), K value, and factor of σ d .

4.2.1. Impact of Different Antenna Configurations

As the detecting range of RFID is directly confined by antennas on robot, in the first series of experiments, we analyzed the positioning accuracy under different antenna combinations. We fixed d s = 0.1 , r m a x = 1.0 , d p = 0.01 , d g = 0.2 , N = 100 , σ a = 0.1 , σ v = 1.0 , K = 4 , and σ d = 0.1 . The result is shown in Table 2. As it can be seen from this table, the positioning accuracy is impacted by the number of antennas used. The accuracy is low while using only one antenna due to its incompleteness on coverage, while we obtain a better positioning accuracy with two antennas as the error decreased from 1.24 m to 0.258 m (i.e., 79 % improvement). Therefore, this paper uses two antennas to locate the moving object.

4.2.2. Impact of Different Prediction Forms

In the next series of experiments, we tested the impact of positioning accuracy under three different predicting forms: the random predicting, laser predicting, and the combination of random predicting and laser predicting. The experimental settings were the same as described previously. The results are shown in Figure 6b and Figure 8. As it can be seen from this figure, the choice of σ r has a significant impact on the positioning accuracy for the random prediction. σ r = 1.0 leads to the best positioning accuracy, while a too small or too large σ r obviously leads to a worse positioning result (1.7 m for σ r = 0.0 and 0.829 m for σ r = 2.0 , as compared to 0.618 m for σ r = 1.0 ). The reason is that, with a small σ r , all particles will converge to a single point after the integration of several RFID and laser measurements, while, with a large σ r , the particle filter is not able to converge, due to the large noise added to the particles.
In addition, as can be seen from Figure 6b and Figure 8, the combining way of prediction was better than the random prediction or the prediction based on laser information, with the locating error diminished from 0.618 m to 0.248 m (59% improvement) and 0.302 m to 0.248 m (18% improvement), respectively. Figure 6b and Figure 8b–c show the track using the three different predicting forms. Figure 8d shows the real-time error of three different prediction forms.

4.2.3. Impact of Different Parameters of Laser Clustering

In the next series of experiments, we compared the positioning accuracy under the impact of different settings of the laser clustering, i.e., grouping threshold d g , distance parameter d p , splitting threshold d s , and the radius threshold r m a x . The results are shown in Figure 9. As can be seen from Figure 9a, grouping threshold d g and the distance parameter d p have high impact on the positioning accuracy. The accuracy decreases when d g and d p are too small or too large. A choice of d g = 0.2 and d p = 0.01 gives the best positioning accuracy. As can be seen from Figure 9b, the radius threshold r m a x also has some impact on the positioning accuracy. A too small r m a x will retain many useless clusters and introduces more error to the positioning, while a too large r m a x will filter out the true clusters and gives a poor positioning accuracy as well. As can also be seen from Figure 9b, the positioning accuracy decreases with a too small or too large splitting threshold d s . When the splitting threshold d s is too small (for example d s = 0.01 ), one obstacle is segmented into several groups and leads to an increase of positioning error, while the accuracy gets worse with a too large d s , as several obstacles are segmented into one group. To keep all these analysis in mind, we choose the parameters as: d p = 0.01 , d g = 0.2 , d s = 0.1 , and r m a x = 1.0 .

4.2.4. Impact of Different Number of Particles N

As the number of particles is directly related to the processing time and positioning accuracy, we carried out experiments to analyse the impact of the number of particles. We used a CPU with Core i3-2330M, 2.2GHz and 6 GB RAM for the processing. Other parameters are set to be the same as previously and the system is checked for different numbers of particles. The results are listed in Table 3. From this table, we can observe that a small number of particles (i.e., N < 100 ) leads to a poor positioning accuracy. The positioning accuracy increases (but not so obviously) with the increase of the number of particles, but the processing time increases significantly at the same time. Therefore, we choose a particle number of N = 100 to achieve the better positioning accuracy and algorithm efficiency.

4.2.5. Comparison of Different Velocity Noise σ v and Moving Direction Noise σ a

In the next series of experiments, we evaluated the impact of the Gaussian noise σ v and σ a added to the velocity and moving direction of laser clusters. We use the same parameter settings as previously except the ones mentioned below and show the results in Figure 10. As can be seen from this figure, a too large or too small σ a will result in a decrease of the positioning accuracy. We get the best positioning accuracy (i.e., 0.259 m) with σ a = 0.1 for a setting of σ v = 1.0 , which is an improvement of 21.5% and 7.5%, as compared to σ a = 0.0 (0.33 m) and σ a = 0.5 (0.28 m), respectively. The same findings apply to the parameter of σ v . Based on the results in Figure 10, we chose σ a = 0.1 and σ v = 1.0 to achieve the best accuracy of the system.

4.2.6. Impact of Different K and the Bandwidth Parameter σ d

In the last series of experiments, we tested the positioning accuracy under the impact of different settings of K and the bandwidth parameter σ d . The results are shown in Figure 11. As it can be seen from this figure, K = 4 gives the best result. A too large or too small K leads to a worse positioning result. Moreover, we also find a suitable value of σ d is at around 0.1 . The positioning accuracy gets worse when σ d is too large (e.g., 1.0) or too small (e.g., 0.01). With the best settings of K and σ d (i.e., K = 4 and σ d = 0.1 ), we achieve a positioning accuracy of 0.25 m.

5. Conclusions

In this paper, we proposed a new approach for moving object localization based on RFID phase and laser clustering. We first computed the RFID phase-based velocity by the phase difference and laser-based velocity from laser clustering. Then, a particle filter is used for the sensor fusion based on the best K matching clusters. We conducted extensive experiments to validate the feasibility of the proposed approach and our results show that a mean positioning accuracy up to 0.25 m can be achieved. In future, we would like to increase the positioning accuracy by incorporating the signal strength information and extend our work to include multiple moving objects.

Acknowledgments

This work is partially supported by the National Natural Science Foundation of China under the grant numbers of 61601381, 61701421, and 61471306, and partially by China’s 13th Five-Year Plan in the Development of Nuclear Energy under the grant number of 2016(1295).

Author Contributions

Yulu Fu analyzed the data and wrote the paper; Changlong Wang revised the paper; Ran Liu proposed the idea and designed the experiments; Gaoli Liang conducted the experiments; Hua Zhang and Shafiq Ur Rehman reviewed the paper.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
UHFUltra High Frequency
LBSLocation-Based Services
RFIDRadio Frequency Identification
Wi-FiWireless Fidelity
RSSReceived Signal Strength
TOATime of Arrival
AOAAngel of Arrival
WSNWireless Sensor Network
EKFExtended Kalman Filter
VPHVector Polar Histogram
RMSERooted Mean Square Error

References

  1. Zhao, Y. Research on Phase Information Based Passive UHF RFID Localization Algorithms. Ph.D. Thesis, TianJin University, TianJin, China, 2015. [Google Scholar]
  2. Chen, Z.; Zou, H.; Jiang, H. Fusion of WiFi, smartphone sensors and landmarks using the Kalman filter for indoor localization. Sensors 2015, 15, 715–732. [Google Scholar] [CrossRef] [PubMed]
  3. Liu, Y. RFID Technical Principle and Its Application. J. Cent. Univ. Natl. 2006, 15, 358–361. [Google Scholar]
  4. Zhou, J.R. Research on Indoor Localization Using RFID System. Ph.D. Thesis, Zhejiang University, Zhejiang, China, 2014. [Google Scholar]
  5. Liu, R.; Koch, A.; Zell, A. Path following with passive UHF RFID received signal strength in unknown environments. In Proceedings of the 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS 2012), Vilamoura, Portugal, 7–12 October 2012; pp. 2250–2255. [Google Scholar]
  6. Xiao, C.; Chen, N.; Li, D. SCRMS: An RFID and Sensor Web-Enabled Smart Cultural Relics Management System. Sensors 2017, 17, 60. [Google Scholar] [CrossRef] [PubMed]
  7. Shaari, A.M.; Nor, N.S.M. Position and Orientation Detection of Stored Object Using RFID Tags. Procedia Eng. 2017, 184, 708–715. [Google Scholar]
  8. Lu, S.; Xu, C.; Zhong, R.Y. A RFID-enabled positioning system in automated guided vehicle for smart factories. J. Manuf. Syst. 2017, 44, 179–190. [Google Scholar] [CrossRef]
  9. Fortinsimard, D.; Bilodeau, J.S.; Bouchard, K. Exploiting Passive RFID Technology for Activity Recognition in Smart Homes. IEEE Intell. Syst. 2015, 30, 7–15. [Google Scholar] [CrossRef]
  10. Nur, K.; Morenzacinos, M.; Carreras, A. Projection of RFID-Obtained Product Information on a Retail Stores Indoor Panoramas. IEEE Intell. Syst. 2015, 30, 30–37. [Google Scholar] [CrossRef]
  11. Wang, J.; Ni, D.; Li, K. RFID-based vehicle positioning and its applications in connected vehicles. Sensors 2014, 14, 4225–4238. [Google Scholar] [CrossRef] [PubMed]
  12. Ren, Y.; Liu, X.; Zhang, X. Research on phase measuring method for ultra high frequency RFID positioning. Technol. Vert. Horiz. 2014, 14, 36-38+42. [Google Scholar]
  13. Hähnel, D.; Burgard, W.; Fox, D.; Fishkin, K.; Philipose, M. Mapping and localization with RFID technology. In Proceedings of the 2004 IEEE International Conference on Robotics and Automation (ICRA 2004), New Orleans, LA, USA, 26 April–1 May 2004; pp. 1015–1020. [Google Scholar]
  14. Ni, L.M.; Liu, Y.; Lau, Y.C.; Patil, A.P. LANDMARC: Indoor Location Sensing Using Active RFID. Wirel. Netw. 2004, 10, 701–710. [Google Scholar] [CrossRef]
  15. Xu, H.; Ding, Y.; Li, P. An RFID Indoor Positioning Algorithm Based on Bayesian Probability and K-Nearest Neighbor. Sensors 2017, 17, 1806. [Google Scholar] [CrossRef] [PubMed]
  16. Yang, P.; Wu, W. Efficient Particle Filter Localization Algorithm in Dense Passive RFID Tag Environment. IEEE Trans. Ind. Electron. 2014, 61, 5641–5651. [Google Scholar] [CrossRef]
  17. Xu, B.; Gang, W. Random sampling algorithm in RFID indoor location system. In Proceedings of the 2006 IEEE International Workshop on Electronic Design, Test and Applicatios, Kuala Lumpur, Malaysia, 17–19 January 2006; p. 6. [Google Scholar]
  18. Fu, Z. Research of Location-Aware Method Based on RFID Technology. Master’s Thesis, Changchun University of Technology, Changchun, China, 2016. [Google Scholar]
  19. Li, X.; Zhang, Y.; Marsic, I.; Burd, R.S. Online People Tracking and Identification with RFID and Kinect. arXiv, 2017; arXiv:1702.03824. [Google Scholar]
  20. Sarkka, S.; Viikari, V.V.; Huusko, M.; Jaakkola, K. Phase-Based UHF RFID Tracking With Nonlinear Kalman Filtering and Smoothing. IEEE Sens. J. 2012, 12, 904–910. [Google Scholar] [CrossRef]
  21. Ma, H.; Wang, K. Fusion of RSS and Phase Shift Using the Kalman Filter for RFID Tracking. IEEE Sens. J. 2017, 17, 3551–3558. [Google Scholar] [CrossRef]
  22. Martinelli, F. A Robot Localization System Combining RSSI and Phase Shift in UHF-RFID Signals. IEEE Trans. Control Syst. Technol. 2015, 23, 1782–1796. [Google Scholar] [CrossRef]
  23. Prinsloo, J.; Malekian, R. Accurate Vehicle Location System Using RFID, an Internet of Things Approach. Sensors 2016, 16, 825. [Google Scholar] [CrossRef] [PubMed]
  24. Song, X.; Li, X.; Tang, W. A hybrid positioning strategy for vehicles in a tunnel based on RFID and in-vehicle sensors. Sensors 2014, 14, 23095–23118. [Google Scholar] [CrossRef] [PubMed]
  25. Germa, T.; Lerasle, F.; Ouadah, N. Vision and RFID data fusion for tracking people in crowds by a mobile robot. Comput. Vis. Image Underst. 2010, 114, 641–651. [Google Scholar] [CrossRef]
  26. Deyle, T.; Reynolds, M.S.; Kemp, C. Finding and Navigating to Household Objects with UHF RFID Tags by Optimizing RF Signal Strength. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Chicago, IL, USA, 14–18 September 2014; pp. 2579–2586. [Google Scholar]
  27. Valero, E.; Adán, A. Integration of RFID with other technologies in construction. Measurement 2016, 94, 614–620. [Google Scholar] [CrossRef]
  28. Valero, E.; Adan, A.; Cerrada, C. Automatic Construction of 3D Basic-Semantic Models of Inhabited Interiors Using Laser Scanners and RFID Sensors. Sensors 2012, 12, 5705. [Google Scholar] [CrossRef] [PubMed]
  29. Zhou, Y.; Liu, W.; Huang, P. Laser-activated RFID-based Indoor Localization System for Mobile Robots. In Proceedings of the IEEE International Conference on Robotics and Automation, Roma, Italy, 10–14 April 2007; pp. 4600–4605. [Google Scholar]
  30. Martin, E. Multimode radio fingerprinting for localization. In Proceedings of the 2011 IEEE International Radio and Wireless Symposium, Phoenix, AZ, USA, 16–19 January 2011; pp. 383–386. [Google Scholar]
  31. Liu, R.; Koch, A.; Zell, A. Mapping UHF RFID Tags with a Mobile Robot using a 3D Sensor Model. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems, Tokyo, Japan, 3–7 November 2013; pp. 1589–1594. [Google Scholar]
  32. Shirehjini, A.A.N.; Yassine, A.; Shirmohammadi, S. An RFID-Based Position and Orientation Measurement System for Mobile Objects in Intelligent Environments. IEEE Trans. Instrum. Meas. 2012, 61, 1664–1675. [Google Scholar] [CrossRef]
  33. Song, X.; Li, X.; Zhang, W. RFID Application for Vehicle Fusion Positioning in Completely GPS-denied Environments. Eng. Lett. 2016, 24, 19–23. [Google Scholar]
  34. Xiong, Z.; Song, Z.; Scalera, A. Hybrid WSN and RFID indoor positioning and tracking system. Eurasip J. Embed. Syst. 2013, 1, 6. [Google Scholar] [CrossRef]
  35. Su, X.; Li, S.; Yuan, C. Enhanced Boundary Condition–Based Approach for Construction Location Sensing Using RFID and RTK GPS. J. Constr. Eng. Manag. 2014, 140, 04014048. [Google Scholar] [CrossRef]
  36. Li, J.; Zhong, R.; Hu, Q.; Ai, M. Feaure-Based Laser Scan Matching and Its Application for Indoor Mapping. Sensors 2016, 16, 1265. [Google Scholar] [CrossRef] [PubMed]
  37. Jung, J.; Yoon, S.; Ju, S.; Heo, J. Development of Kinematic 3D Laser Scanning System for Indoor Mapping and As-Built BIM Using Constrained SLAM. Sensors 2015, 15, 26430–26456. [Google Scholar] [CrossRef] [PubMed]
  38. Tang, J.; Chen, Y.; Jaakkola, A.; Liu, J.; Hyyppä, J.; Hyyppä, H. NAVIS—An UGV indoor positioning system using laser scan matching for large-area real-time applications. Sensors 2014, 14, 11805–11824. [Google Scholar] [CrossRef] [PubMed]
  39. An, D.; Wang, H. VPH: A new laser radar based obstacle avoidance method for intelligent mobile robots. In Proceedings of the Intelligent Control and Automation, Hangzhou, China, 15–19 June 2004; pp. 4681–4685. [Google Scholar]
  40. Sun, J.L.; Sun, G.M.; Ma, P.G. Laser Target Localization Based on Symmetric Wavelet Denoising and Asymmetric Gauss Fitting. Chin. J. Lasers 2017, 44, 178–185. [Google Scholar]
  41. Cai, Z.; Xiao, Z.; Yu, J. Real-time Detection of Dynamic Obstacles Based on Laser Radar. Control Eng. China 2008, 15, 200–203. [Google Scholar]
  42. Zhou, J.J. Research on Key Technology of Lidar Based Object Detection and Tracking for Intelligent Vehicles. Ph.D. Thesis, Beijing University of Technology, Beijing, China, 2014. [Google Scholar]
  43. Vu, T.D.; Aycard, O.; Appentodt, N. Online Localization and Mapping with Moving Object Tracking in Dynamic Outdoor Environments. In Proceedings of the IEEE Intelligent Vehicles Symposium, Istanbul, Turkey, 13–15 June 2007; pp. 190–195. [Google Scholar]
  44. Streller, D.; Dietmayer, K. Object tracking and classification using a multiple hypothesis approach. In Proceedings of the 2004 IEEE Intelligent Vehicles Symposium, Parma, Italy, 14–17 June 2004; pp. 808–812. [Google Scholar]
  45. Zhang, Z.D. Research on Dynamic Navigation of Indoor Service Robot Using Laser Scanner. Master’s Thesis, Harbin Institute of Technology, Harbin, China, 2014. [Google Scholar]
  46. Liu, R.; Huskic, G.; Zell, A. Dynamic objects tracking with a mobile robot using passive UHF RFID tags. In Proceedings of the 2014 IEEE/RSJ International Conference on Intelligent Robots and Systems(IROS 2014), Chicago, IL, USA, 14–18 September 2014; pp. 4247–4252. [Google Scholar]
  47. Han, J.; Ding, H.; Qian, C.; Ma, D.; Xi, W.; Wang, Z.; Jiang, Z.; Shangguan, L. CBID: A Customer Behavior Identification System Using Passive Tags. In Proceedings of the 2014 IEEE International Conference on Network Protocols, Raleigh, NC, USA, 21–24 October 2014; pp. 47–58. [Google Scholar]
  48. Huang, R.; Liang, H.; Chen, J. Method of tracking and recognition of laser radar unmanned vehicle dynamic obstacle detection. Robot 2016, 38, 437–443. [Google Scholar]
  49. Tan, B.; Yang, C. Dynamic Obstacles Detection by Laser Radars. J. Xi’an Technol. Univ. 2015, 35, 205–209. [Google Scholar]
  50. Liu, R.; Yuen, C.; Do, T.N. Indoor Positioning using Similarity-based Sequence and Dead Reckoning without Training. arXiv, 2017; arXiv:1705.04934. [Google Scholar]
Figure 1. System overview.
Figure 1. System overview.
Sensors 18 00825 g001
Figure 2. Overview of laser-based clustering.
Figure 2. Overview of laser-based clustering.
Sensors 18 00825 g002
Figure 3. Laser ranging data at two timestamps.
Figure 3. Laser ranging data at two timestamps.
Sensors 18 00825 g003
Figure 4. Illustration of laser-based grouping.
Figure 4. Illustration of laser-based grouping.
Sensors 18 00825 g004
Figure 5. Comparison of the laser groups before and after splitting and a real example. (a) before splitting; (b) after splitting; (c) clustering results.
Figure 5. Comparison of the laser groups before and after splitting and a real example. (a) before splitting; (b) after splitting; (c) clustering results.
Sensors 18 00825 g005
Figure 6. Setup of the experiment and positioning result. (a) Setup of the experiment; (b) ground truth and estimated track by a combination of two prediction forms.
Figure 6. Setup of the experiment and positioning result. (a) Setup of the experiment; (b) ground truth and estimated track by a combination of two prediction forms.
Sensors 18 00825 g006
Figure 7. Complex paths. (a) 8-shaped path; (b) W-shaped path.complex paths
Figure 7. Complex paths. (a) 8-shaped path; (b) W-shaped path.complex paths
Sensors 18 00825 g007
Figure 8. Comparison of the performance in three different prediction forms. (a) mean positioning accuracy under different prediction forms; (b) ground truth and estimated track of random prediction; (c) ground truth and estimated track of laser prediction; (d) estimation error at different timestamps.Different prediction method.
Figure 8. Comparison of the performance in three different prediction forms. (a) mean positioning accuracy under different prediction forms; (b) ground truth and estimated track of random prediction; (c) ground truth and estimated track of laser prediction; (d) estimation error at different timestamps.Different prediction method.
Sensors 18 00825 g008
Figure 9. Positioning accuracy under the impact of different parameters of laser clustering. (a) Impact of different grouping thresholds d g and distance parameters d p ; (b) Impact of different maximum cluster radius r m a x and splitting threshold d s .
Figure 9. Positioning accuracy under the impact of different parameters of laser clustering. (a) Impact of different grouping thresholds d g and distance parameters d p ; (b) Impact of different maximum cluster radius r m a x and splitting threshold d s .
Sensors 18 00825 g009
Figure 10. Positioning accuracy under the impact of different velocity noise σ v and moving direction noise σ a .
Figure 10. Positioning accuracy under the impact of different velocity noise σ v and moving direction noise σ a .
Sensors 18 00825 g010
Figure 11. Mean positioning error under the impact of different K and bandwidth σ d .
Figure 11. Mean positioning error under the impact of different K and bandwidth σ d .
Sensors 18 00825 g011
Table 1. Mathematical symbols and their meanings.
Table 1. Mathematical symbols and their meanings.
Mathematical SymbolMeaning
φ t Phase of RFID signal at time t
v t r Phase-based velocity of RFID tag at time t
d g Grouping threshold in laser-based clustering
d p Distance parameter in laser-based clustering
d s Splitting threshold in laser-based clustering
r m a x The maximum cluster radius
v t ( i ) The velocity of cluster i at time t
θ t ( i ) The moving direction of cluster i at time t
KThe number of the best matching clusters
NNumber of particles
X t The object position at time t
( x t [ n ] , y t [ n ] ) Location of particle n at time t
ω t [ n ] The weight of particle n at time t
σ r Gaussian noise in random prediction
σ a Gaussian noise added to the moving direction in laser prediction
σ v Gaussian noise added to the velocity in laser prediction
σ d The bandwidth parameter used to control the weight update of the particle filter
θ t ( l ) ˜ Moving direction after adding Gaussian noise σ a to the cluster l at time t
v t ( l ) ˜ Velocity after adding Gaussian noise σ v to cluster l at time t
Table 2. Comparison of positioning accuracy in meters under different antenna combinations.
Table 2. Comparison of positioning accuracy in meters under different antenna combinations.
Antenna CombinationOnly Right AntennaOnly Left AntennaRight and Left Antennas
Positioning accuracy (m)1.241.410.258
Table 3. Mean positioning accuracy and running time of the algorithm under the impact of different number of particles N.
Table 3. Mean positioning accuracy and running time of the algorithm under the impact of different number of particles N.
Number of Particles NAccuracy (m)Running Time (ms)
50.4574.187
100.2954.535
500.2794.655
1000.2564.858
5000.2546.654
10000.2588.594

Share and Cite

MDPI and ACS Style

Fu, Y.; Wang, C.; Liu, R.; Liang, G.; Zhang, H.; Ur Rehman, S. Moving Object Localization Based on UHF RFID Phase and Laser Clustering. Sensors 2018, 18, 825. https://doi.org/10.3390/s18030825

AMA Style

Fu Y, Wang C, Liu R, Liang G, Zhang H, Ur Rehman S. Moving Object Localization Based on UHF RFID Phase and Laser Clustering. Sensors. 2018; 18(3):825. https://doi.org/10.3390/s18030825

Chicago/Turabian Style

Fu, Yulu, Changlong Wang, Ran Liu, Gaoli Liang, Hua Zhang, and Shafiq Ur Rehman. 2018. "Moving Object Localization Based on UHF RFID Phase and Laser Clustering" Sensors 18, no. 3: 825. https://doi.org/10.3390/s18030825

APA Style

Fu, Y., Wang, C., Liu, R., Liang, G., Zhang, H., & Ur Rehman, S. (2018). Moving Object Localization Based on UHF RFID Phase and Laser Clustering. Sensors, 18(3), 825. https://doi.org/10.3390/s18030825

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop