Next Article in Journal
Immobilization Techniques in the Fabrication of Nanomaterial-Based Electrochemical Biosensors: A Review
Previous Article in Journal
Half-Cell Potential Analysis of an Ammonia Sensor with the Electrochemical Cell Au | YSZ | Au, V2O5-WO3-TiO2
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

On Calibrating the Sensor Errors of a PDR-Based Indoor Localization System

1
Department of CSIE, National Cheng Kung University, Tainan 701, Taiwan
2
Department of CS, National Chiao Tung University, Hsinchu 30010, Taiwan
*
Author to whom correspondence should be addressed.
Sensors 2013, 13(4), 4781-4810; https://doi.org/10.3390/s130404781
Submission received: 15 February 2013 / Revised: 27 March 2013 / Accepted: 29 March 2013 / Published: 10 April 2013
(This article belongs to the Section Physical Sensors)

Abstract

: Many studies utilize the signal strength of short-range radio systems (such as WiFi, ultrasound and infrared) to build a radio map for indoor localization, by deploying a large number of beacon nodes within a building. The drawback of such an infrastructure-based approach is that the deployment and calibration of the system are costly and labor-intensive. Some prior studies proposed the use of Pedestrian Dead Reckoning (PDR) for indoor localization, which does not require the deployment of beacon nodes. In a PDR system, a small number of sensors are put on the pedestrian. These sensors (such as a G-sensor and gyroscope) are used to estimate the distance and direction that a user travels. The effectiveness of a PDR system lies in its success in accurately estimating the user's moving distance and direction. In this work, we propose a novel waist-mounted based PDR that can measure the user's step lengths with a high accuracy. We utilize vertical acceleration of the body to calculate the user's change in height during walking. Based on the Pythagorean Theorem, we can then estimate each step length using this data. Furthermore, we design a map matching algorithm to calibrate the direction errors from the gyro using building floor plans. The results of our experiment show that we can achieve about 98.26% accuracy in estimating the user's walking distance, with an overall location error of about 0.48 m.

1. Introduction

The accurate localization of objects and people in an environment has long been considered an important component of ubiquitous networking. There are many studies on indoor localization [1,2], several of which involve the use of radio signal strength indicators (RSSI) [38] and a large number of preinstalled devices, called beacon nodes, which periodically emit signals. By listening and monitoring signal changes [9,10], the user's location can be estimated. While the idea of these existing systems is attractive, they require intensive pre-deployment efforts for every new building. In addition, the signal variations over time could be potentially very large for real-world operations [11].

On the other hand, a personal dead reckoning or pedestrian dead reckoning (PDR) system is a self-contained technique for indoor localization. This technique only requires a couple of sensors to be put on the user, so that it can be used in any building without pre-installing beacon nodes or pre-building RF maps/propagation models based on surveys of the environment. Most PDR systems use inertial sensors (such as an accelerometer, gyroscope, or digital compass) to measure step length and heading direction. Some also use other sensors, such as cameras [12], ultrasound [13], RFID [14,15], or laser [16], to calibrate their results. Generally, depending on where the sensors are placed, previous studies classified PDR systems into two types: foot- and waist-mounted. The foot-mounted method [1727] uses a double integral on horizontal acceleration to estimate distance, and a gyroscope or compass to measure the heading direction. On the other hand, the waist-mounted [2831] method tries to detect each step event to calculate the total number of steps, and then multiplies this by a constant step length which is based on the pedestrian's characteristics (weight, height, and age) to estimate the total moving distance. Some waist-mounted methods use linear regression to find the relationship between the acceleration, walking speed, and step length [32]. However, when the pedestrian's walking pattern is different from the predefined step length model, this may adversely affect the distance estimation. Finally, although a waist-mounted method is generally more feasible to be implemented on a hand-held device, its accuracy in estimating the step length is typically worse than that of a foot-mounted method which, on the other hand, performs poorly with regard to obtaining an accurate orientation [33].

Sensor drift [34] is a well-known problem in PDR systems. Given that the hardware used in such systems is not perfect, the inertial sensors constantly have some small errors when estimating the distance and direction, and signal noise (such as vibrations of the user's body) can further exacerbate this problem [25,26]. Some PDR systems use a map matching mechanism to calibrate these errors [3537], and these can be categorized into two types. One tries to match the user trajectory to the closest junction and road on the map [37], while the other utilizes the map to filter out positions where the user is unlikely to move (e.g., walls, obstacles, and so on) [35,36]. Both techniques require the use of a detailed scaled map of the building, as shown in Figure 1(a), although in practice this is usually difficult to obtain.

In this paper, we consider a scenario in which the user has a smart-phone and can access the floor plan of the building, like the one shown in Figure 1(b), as these are widely available. The system utilizes the sensors on the smart-phone to compute the user's moving distance and direction, and this can be used with the map to estimate their current position in the building, as shown in Figure 2. Our system can be considered as an add-on to a traditional outdoor navigation app (e.g., Google Latitude [38]) so that the starting position of the user in the building can be estimated using the last-recorded GPS position.

The contribution of this paper is threefold. First, using the characteristic of simple harmonic motion (SHM) and the Pythagorean Theorem, we propose a novel waist-mounted PDR method to accurately estimate the step length for measuring the moving distance. The accuracy of our distance estimation is about 98.25%. Second, based on the geometric similarity between the user trajectory and the floor map, we design a map-matching algorithm to calibrate sensor errors using common building floor plans. Finally, we implement our PDR system on a smart-phone, and our results show that the location error is only about 0.48 m in our test scenario.

The rest of this paper is structured as follows: In Section 2, we describe the related work. We discuss the details of our step-length estimation and map-matching algorithms in Section 3 and Section 4, respectively. The results of our experiment are shown in Section 5. Finally, we conclude this paper in Section 6.

2. Related Work

Our work is built on some previous studies of indoor localization, step length estimation, and map matching.

2.1. Indoor Localization

Many studies of indoor localization use wireless signals to estimate the location, and such approaches can generally be divided into three types, those that use triangulation, fingerprinting, and proximity. The triangulation methods include, for example, the use of time of arrival (TOA) [39], time difference of arrival (TDOA) [40], and received signal strength (RSS) [39]. Fingerprinting methods also utilize the RSS, and in these, a survey phase is required to measure and record the RSS of every place in the building to build a radio map. To locate the target's position, one simply measures the RSS at the user's location and uses some pattern-matching algorithms, such as KNN [41] or SVM [42], to find the position in the radio map which has the most similar RSS record. Finally, the proximity methods usually first deploy many antennas (e.g., RFID [4345]) at known positions. The antenna which is closest to the mobile target will generally have the strongest RSS, which can then be used to estimate the location of the target. However, these signal-based approaches suffer a common problem, which is that radio signals are very often suffer from interference in an indoor environment [11]. In addition, the deployment of the necessary infrastructure (e.g., base-stations or antennas) is costly and labor-intensive.

In addition to the signal-based methods, many prior studies use a PDR system to estimate the trajectory of a user by placing some sensors on the user's body. Inertial sensors, such as accelerometers, gyroscopes, and compasses, are commonly used in such systems. Some PDR systems also include GPS sensors [46,47], and use these to calibrate the PDR drift as long as the GPS signal is available. When the GPS signal is obstructed, the system can then be changed to the PDR mode and continue to record the trajectory. Our study is based on the PDR system, which has the advantage of avoiding the deployment overhead of the signal-based methods. On the other hand, the performance of our system can be enhanced by a signal-based system if available. For example, one limitation of our map-matching approach is its need to collect “enough” trajectory data before it can uniquely identify the user's position on the map. When there is not enough trajectory information, we make use of the known locations of existing WiFi base stations in the building to help estimate the user's location [4850].

2.2. Step Length Estimation

A PDR system can be classified into two types depending on where the sensor is mounted: foot-mounted [1727] and waist-mounted [2831]. During walking, when one foot is swinging forward, the other one must stand on the ground to support the weight of the body, and both methods [1727] use these movement characteristics to detect a step event. To calculate the step length, the foot-mounted methods typically perform a double integral on the horizontal acceleration. However, without further calibration, the problem of sensor drift [34] could introduce serious inaccuracies when estimating the step length. One way to calibrate the sensor drift error is called zero velocity update (ZUPT). When the swinging-foot touches the ground, the angular velocity of this foot will be close to zero, which can be used to reset the system to avoid sensor drift errors accumulating into the length estimation of the next step.

On the other hand, for a waist-mounted PDR system, ZUPT is not directly applicable, since one will not be able to find zero velocity in the horizontal direction. Some waist-mounted PDR systems use a constant step length [5154] while the others [30,55] use a trace-driven approach, by first collecting empirical data from multiple users and then using linear regression to find the relation between step length, walking frequency, and the variance of acceleration. The limitation of this approach is that one might need to collect new training data for a new user. Weinberg [31] observed that the upper body moves vertically when walking, and suggested that one can estimate the step length as follows Equation (1):

StepLength = 2 × heightchan ge / α
where α is the swinging angle of the leg from the body, and the heightchange can be estimated based on the vertical acceleration. However, he did not discuss how to measure α. Our idea is similar to Weinberg's, but we estimate the step length based on the height change and length of the leg using the Pythagorean Theorem. In addition, we use the concept of Simple Harmonic Motion (SHM) to find the zero velocity in the vertical direction and apply ZUPT to avoid the accumulation of sensor drift errors on the vertical-axis.

2.3. Map-Matching Algorithm

In some prior PDR systems, a map-matching mechanism is used to match the user trajectory onto the map [3537] in order to calibrate the sensor errors. There are two ways this is achieved. The first tries to match the user trajectory to the closest junction and road on the map [37], while the other utilizes the map information to filter out positions where the user is unlikely to walk (e.g., walls and obstacles) [35,36]. However, both techniques require the use of a detailed scaled map of the building (i.e., with detailed distance information for each route on the map) which, however, is usually not easy to obtain in reality. In our work, we utilize the more commonly-seen building floor plans instead of detailed scaled maps. Based on the geometric similarity between the trajectory data and the map, we propose a new map-matching method that uses the floor plan for locating the user. As shown later in Section 5, our approach has similar performance to those methods that rely on detailed scaled map information.

3. A Waist-mounted Pedestrian Dead Reckoning (PDR) System

In this section, we describe the design of our waist-mounted method for estimating the user's walking step length. We observe that, as shown in Figure 3, when the sensor is mounted on the waist, the height of the sensor (i.e., L-h, where h is the vertical displacement of the sensor during the walking), the length of the leg (i.e., L) and half of the step length (i.e., D/2) forms a right triangle in which leg length is the hypotenuse. Assuming that L is known, based on the Pythagorean Theorem, we can obtain step length D if we can obtain the vertical displacement of the sensor (i.e., h), and details of this are discussed later in this section.

We focus on the waist-mounted method because it is generally more feasible to be extended for a hand-held device, such as a smart-phone. When implementing a PDR system on a hand-held device, two cases can be considered. The first is when the user holds the device (e.g., talking on the phone), and the other is when the device is put in a pocket or a bag. Prior research [56] has shown the feasibility of using a waist-mounted method for the first case. Therefore, in this section we only discuss how to extend the results of a waist-mounted method for the second case. To implement a waist-mounted method on a hand-held device, two issues need to be considered: orientation and placement of the device. The orientation of the device may change from time to time when it is put in a pocket or bag during walking, and this change in orientation affects the influence of gravity on the three axes of the accelerometer, and results in different readings from the sensor. To resolve this problem, we can adopt a method similar to that in an earlier work, Su et al. [56] which used a gyroscope to record changes in orientation and to calibrate the system, as shown in Figure 4. Next, in a waist-mounted method, the vertical movement displacement of the device is the key parameter to estimate the step length. When the device is positioned above the waist-line, its vertical displacement during walking will be the same as when it is mounted on the waist. On the other hand, when the device is placed below the waist-line (e.g., in a pants pocket), we can estimate the vertical displacement of the waist as follows. Assuming the leg length (L) and the pocket position from the ground (L′) are already known, we can find two similar triangles ΔABC and ΔPQR, as shown in Figure 5. Based on the Pythagorean Theorem, we can obtain Q R ¯ by first measuring L′ and h′ (i.e., vertical displacement of the pocket during the walking). Then, using triangle similarity, we can find B C ¯ = D / 2 = ( L × Q R ¯ ) / L h can be measured in a similar way as to measure h, which is discussed next.

3.1. Using the Height Change of the Waist to Estimate Step Length

Previously Weinberg [31] proposed that one can estimate the step length using the height change of the waist and the swinging angle of the leg during walking [31]. However, he did not discuss how to measure this angle, and we found that, in practice, it is difficult to measure such a small angle during walking. In this paper we use the Pythagorean Theorem and propose a different way to estimate step length using the change in height. During walking, a person's body moves up and down. If we assume the length of the leg is L, the waist-line will move up-and-down between L and (L-h) from the ground, where h is the change in the height of the waist. Consider the triangle in Figure 3, formed by the two feet of a person and their step length D. Given that L is known, by using the Pythagorean Theorem, we can estimate D if we know the height of this triangle, i.e., (L-h). To obtain (L-h), we need to first calculate h which is the change in height of the waist during walking. Therefore, if we mount an accelerometer on the user's waist, the readings of the accelerometer can be used to estimate the height change h, which can then be used to calculate the step length D based on the Pythagorean Theorem.

To implement this method, we need to first resolve the following issues.

(1)

How to eliminate the influence of gravity?

(2)

How to remove signal noise, such as vibrations from the body?

(3)

How to recognize the beginning and end of a step based on readings from the sensor?

(4)

How to prevent sensor drift errors accumulating from one step to the next?

Our system architecture is shown in Figure 6. We first remove the influence of gravity from the original sensor data, and then the noise is filtered out by a low-pass filter. The filtered signal is fed into the step recognition module to identify each new step. We adopt the concept of simple harmonic motion (SHM) [57] to reset the vertical velocity at the beginning of each new step, which prevents sensor drift errors from being accumulated over to the next one. Finally, the step length is estimated based on the filtered sensor data and the foot length.

3.2. The elimination of the Influence of Gravity

The readings from an accelerometer are affected by gravity. An accelerometer is typically built on a silicon wafer, with a poly-silicon spring on the wafer surface to provide an opposing force against external force. Any external force, including gravity, can cause the deflection of the spring. According to the different levels and direction of this deflection, the values of resistance or capacitance on the circuit will change proportionally, and the device will then output the corresponding voltages, which are the readings measured from the sensor. Therefore, at any time, the sensor can detect the acceleration due to gravity (1 g on Earth), except when the axis (we use a three-axis accelerometer) from which we collect the data is perpendicular to gravity. Therefore, to obtain the vertical acceleration of the waist during walking for estimating changes in height, we first need to remove the effect of gravity from the sensor readings. Generally speaking, the angle between the direction of the measured axis and the direction of gravity can be from 0° to 180°. For the sake of discussion, we consider two cases: when the axis is parallel to the direction of gravity and when it is not.

3.2.1. If the Axis and the Direction of Gravity Are Parallel

In this case, the angle between the measured axis and the direction of gravity is 0° or 180°, as shown in Figure 7. The vertical acceleration caused by the up-and-down movement of the waist during walking can be estimated by Equation (2). Here, SensorReading is the reading we collect from the sensor for one particular axis, while α is the external force that causes the sensor to move up or down. The minus sign indicates that the direction of the force is downward.

α = Sensor Reading Gravity ( = 9.8 m / s 2 )

Note that since gravity is a downward force, the reading α from the sensor will be more than 1 g (−9.8 m/s2).

3.2.2. If the Axis and the Direction of Gravity Are Not Parallel

As shown in Figure 8(a), the measurements collected from the sensor will be affected by two different forces: gravity and the external force, as shown in Figure 8(b). Here we assume that the external force comes from the up-and-down movement of the waist during walking. To estimate the external force, we need to first compute the component of the external force on the measured axis, which can be achieved by subtracting the component of gravity from the sensor reading. Assuming the angle between the direction of measured axis and the direction of the gravity is θ, we can then obtain Equation (3.3):

M = gravity × cos θ
N = ExternalFo rce
Sensor Reading = N × cos θ + M

In the above equation, SensorReading is the value read from the measured axis of the sensor, which is the sum of the component of ExternalForce and the component of gravity. M is the gravity component which can be measured when the sensor is static (e.g., placed on a table):

N = ( Sensor Reading M ) ÷ cos θ
θ = cos 1 ( M / gravity )

The external force N can then be calculated using Equations (3.4) and (3.5).

N = ( SensorReading M ) × sec ( cos 1 ( M / 9.8 ) )

3.3. Noise Filtering

During walking, some unexpected and unpredictable body vibrations might cause some higher-frequency noise in the sensor readings. In this section, we discuss how we filter such noise.

Intuitively, one can use a low pass filter and preset a cut-off frequency to filter the noise. Some prior work has shown that the frequency of human muscle movement is lower than 16 Hz [58], and the human step frequency is never higher than 3 Hz [29]. Therefore, some step recognition systems use 3 Hz as their cut-off frequency to filter noise signal [53]. Initially, we also used 3 Hz, but then found our results became worse. This shows that, while it is good enough to detect a new step event, a 3 Hz threshold is too low for our purposes, and will remove data which is not noise. Some prior work [59] analyzed the acceleration of the waist during walking, and found the maximum acceleration is 8 Hz. Therefore, we set our cut-off frequency at 8Hz for filtering the noise, and this threshold worked well under repeated experiments with different subjects.

3.4. Step Recognition Module

In order to recognize a step, we first analyze the components of one step that can cause vertical changes to the body. There are three major events which may affect the height of the waist, as shown in Figure 9(a).

The first one is a heel-touching-ground event, which happens when the heel just hits the ground and the waist is in its lowest position during the entire step. The event that comes after this is the stance, which occurs when the foot is flat on the ground. Finally, the heel-off-ground event occurs right after the stance. Generally, as shown in Figure 9(b), the vertical acceleration of a heel-touching-ground event is the local minimum within a step. In addition, previous studies [51,61] showed that human walking frequency is never over 3 Hz. Therefore, the duration between two consecutive heel-touching-ground events must be over 0.33 second. Based on the above, we use a sliding window algorithm to detect every heel-touching-ground event, and define one step as from a heel-touching-ground event to the next heel-touching-ground event. Once a new step is identified, the sensor data between two consecutive heel-touching-ground events will be used to estimate the step length.

3.5. Step Length Estimator

3.5.1. Our Step Model

To measure the walking distance from point A to point B, we sum up the step length of all steps. We model a complete step as from one heel-touching-ground-event to the next. Therefore, we consider the first step [i.e., from the stance event to the heel-touching-ground event, as shown in Figure 9(a) ] and the last step (i.e., from the heel-touching-ground event to the stance event) as half a step. When the first or last step is detected, our system will divide the calculated step length by 2.

3.5.2. Simple Harmonic Motion

Once each step can be identified, we can then do the double integral to calculate the height change of the waist and then use this information to estimate the length of each stride based on the Pythagorean Theorem. However, as discussed previously, if the system only naively does the double integral on the accelerometer data, the sensor drifts errors could accumulate from one step to the next. To avoid this, we propose a zero velocity update method to calibrate the sensor data based on the concept of Simple Harmonic Motion (SHM) [57].

Simple harmonic motion has been widely used to model various physical phenomena, such as the oscillations of a spring. When an object is in simple harmonic motion, the displacement of the object is proportional to the external force placed on it, and the force always points to the position of equilibrium. In other words, when the object is displaced from its equilibrium position, it experiences a net restoring force toward this position. One characteristic of SHM is that when an object is at its largest displacement from the equilibrium position, the object's speed will become zero, and at the same time, the object will reach its maximum acceleration rate. For instance, if we put a pen on a spring and make the spring oscillate in vertical way, and then put a long ribbon of paper on a table and let the pen draw on it to show the trajectory of spring vibration at the same time as we pull the paper in a horizontal direction at a stable speed, then the trajectory shown on the paper will look like a sinusoidal wave, and the vertical velocity of the highest and the lowest points of the spring will be zero.

Inspired by this observation, if we mount a sensor on the waist of a pedestrian, then while they are walking the trajectory of the sensor can be approximated by a sinusoidal wave. In addition, given that the velocity of the highest and lowest points of the sensor will be zero, we can utilize these characteristics to detect when the user starts a new step. Finally, once the points with zero velocity in the vertical direction [where the vertical acceleration reaches its local maximum, i.e., points A and B in Figure 9(b) ] are identified, we can then reset the vertical velocity to zero before doing the double integral for calculating the height change of the user's waist. The height change can then be used to estimate the step distance based on the Pythagorean Theorem.

3.5.3. Step Length Estimation

Once a new step event is detected, we can compute the step length using the double integral: the integral of acceleration will give us the velocity and we can get the distance of the step from the integral of the velocity. To avoid sensor drift errors being accumulated, which would adversely affect the double integral results, we can utilize the abovementioned characteristics of SHM to calibrate the sensor data. As discussed previously, there are three major events during one step, namely heel-touching-ground, stance, and heel-off-ground. As shown in Figure 9(b), the lowest valley (point A) of the wave indicates when the heel is touching ground [corresponding to the 5th posture in Figure 9(a) ], the first peak occurs (point B) when the walker is in the stance state [corresponding to the first posture in Figure 9(a) ]. Following the stance event, the body starts to lean forward and the foot is now on its toes, which will give a force to push the body up, so that the vertical acceleration will change to the opposite direction and cause the second valley in Figure 9(b) [i.e., point C, corresponding to the second posture in Figure 9(b)], due to the law of inertia.

As shown in Figure 9(a), when the stance and heel-touching-ground events occur, the waist has the largest displacement from its equilibrium position. Therefore, we reset the vertical velocity to zero at these points. We performed an experiment to observe the effect of doing such a zero velocity update (ZUPT). As shown in Figure 10, implementing ZUPT can indeed avoid the accumulation of sensor drift errors:

V n = 1 n a n d t
h = ( 1 n | V n | d t ) / 2

Next, since we consider the change in height of the waist as a simple harmonic motion, if we simply do the double integral without considering the direction of the force (i.e., the height change is due to the ascent and descent of the body movement), the computed displacement will be zero. Therefore, we calculate the absolute value of current velocity and use it to do the next integral, as shown in Equations (4) and (5), where Vn is the vertical velocity, an is the vertical acceleration, and h is the change in height. We can get the velocity by doing the integral of vertical acceleration. The result of this is then divided by two to get an average of height change; here we assume that the height change in the directions of ascent and descent are approximately the same. Finally, we use the obtained height change of the waist to estimate stride length based on the Pythagorean Theorem. As shown in Figure 3, if we assume that the leg length is L and the height change is h, we can use Equation (6) to calculate the step distance D:

D = 2 × L 2 ( L h ) 2

The first and last steps are considered as special cases. If the system detects that the user is currently in his/her first or last step, the height change will not be divided by two. In addition, there is only one point where the velocity needs to be reset, because the initial velocity of the first step is already zero and the last step does not need to reset this when the body is in its stance state. Combining the distance and the orientation information (which can be obtained from a gyroscope), we can then calculate the user's coordinates and trajectory in a 2D space, as shown in Equation (7):

{ X n = X n 1 + D i s × C o s θ Y n = Y n 1 + D i s × S i n θ

3.5.4. Moving at High Speed or on the Same Spot

As shown in Figure 9(b), when one walks at a normal speed, the stance event generally occurs as the first peak after the heel-touching-event. However, during high speed movement, we observe that the counter-force from the ground could introduce an extra pulse between the heel-touching-ground and stance events, as shown in Figure 11 (points A′ and B′) and Figure 12. Without taking this into consideration, our system could reset the velocity at the wrong point. Previous studies in kinesiology [37,62] showed that the duration from the heel-touching-ground event to the stance event normally accounts for at least 16.7% of the time in a step. Based on this observation, we can identify the right point where the stance event occurs and reset the velocity when calculating the step length.

When one is moving on the same spot, the acceleration data should not be considered in the calculation of step length, and the height of the waist does not usually change significantly. Therefore, we use a simple threshold-based filter to detect this phenomenon by looking at whether the acceleration data in all three directions (vertical, horizontal, and lateral) are lower than a certain threshold ε, as shown below, and then reset the vertical acceleration accordingly (in our implementation, ε is set to 1 m/s2):

verticalAcc = 0 , i f | lateralAcc | < ɛ & | horizontalAcc | < ɛ & | verticalAcc | < ɛ

4. Gyroscope Error Calibration with Map Matching

The effectiveness of a PDR system lies in its success in accurately estimating the user's moving distance and direction. In the previous section, we discussed how to use zero velocity updating to reduce the sensor drift error when measuring the distance. In a PDR system, the direction in which the user is heading is most commonly obtained from a gyroscope sensor. However, a gyroscope can only produce the relative angular displacement (RAD) of a device with respect to a specific direction, and this is not necessarily the absolute direction. Therefore, while we could track the user's trajectory using the gyroscope, this trajectory might be biased by the error in its initial direction and appear as a rotated version of the true path, as in the example shown in Figure 13. When the error of the gyroscope is significant, and if left uncorrected, it can make the entire PDR system unusable. Map-matching is the process of comparing the pedestrian's trajectory data with a digital map of the environment to match the trajectory data to the route segment on which the pedestrian is walking, and it can be used to correct the heading error of a PDR system [63]. Unlike previous map-matching approaches that require the use of a detailed scaled map (i.e., with detailed distance information for each corridor on the map) [35,36], which is normally difficult to obtain in practice, in this study we propose a new map-matching method utilizing the more commonly-seen building floor plans to calibrate the gyroscope errors. We assume that a floor plan is an “approximate” scaled down version of the physical layout of the floor. Our basic idea is to utilize the geometric similarity between the trajectory data and the floor plan to infer the last-visited corner by the user. The flow chart of our algorithm is shown in Figure 14.

Before starting the map matching, we adopt an approach similar to that in a prior work [64] by first converting the floor plan into a link-node model in which information such as the turning angles of the corners and comparative ratios of the lengths between any two corridors can be estimated, as shown in Figure 15. The link-node model is used to approximate the layout of corridors and corners. We then compare the geometry of the user trajectory with the link-node model to find the possible routes that the user has travelled. Here we consider the map and the trajectory as two independent graphs, say M and T. We list out all the sub-graphs of M and compare T with all these sub-graphs to find the most similar one. We define the “similarity” between two graphs by comparing their shapes, vertex angles and relative edge lengths. Once a unique route is identified, this can then be used to calibrate the trajectory data and identify the most recently visited corner. Since the location of every corner is known within the floor plan, the system can then locate the user while they move between corners using the dead-reckoning data from the accelerometer, as previously discussed in Section 3. Note that, given that the link-node model is only an approximation of the physical layout of the building (e.g., the link length might not be an exact scaled down version of the corridor length), the results of this comparison between the map and trajectory could generate multiple candidate routes. Therefore, we also implement an RSSI-based filter by using the existing WiFi-signal-based landmarks (e.g., a corridor-corner may overhear a unique set of WiFi APs, but the set may change at short distances away from that spot; some dead spots inside a building may not overhear any WiFi signals, which by itself is a signature). When a WiFi AP signal is available, we can use this RSSI-filter to select the correct route from multiple candidates. For example, if we know the user has passed a certain landmark, we can remove those candidate routes that do not contain it. This approach is similar to the method used by another recent study [65]. The map-matching process is performed every time the system detects that the user makes a turn. Details of the turn detection process are described below.

4.1. Turn Detection

The flow chart of our map matching algorithm is shown in Figure 14. The first step is to find out all the turnings in the trajectory data and use these as a signature to match all possible routes on the map. However, given that the gyroscope can only provide the relative angular displacement (RAD), we use a sliding-window-based algorithm to infer the user's possible turnings from the data. To determine if a person is making a turn, we compare the standard deviation of the window with a threshold which is estimated during the period when the user is walking straight. A turning event is considered to have occurred when the standard deviation of the window exceeds the threshold. Note that, since it could take several steps to turn around a corridor corner, we record all the turning events that are related to a possible corner-turning event in order to compute the angle of making a complete turn of this corner. One example is shown in Figure 16. The turning angle (CT) of a possible corner can be estimated as:

C T = A T B T

Here AT is the heading angle after making a turn around a corner and BT is the heading angle before making a turn around a corner. For the example in Figure 16, CT = ATBT = 90 – 0 = 90. However, in reality, it is not necessary that one only makes a turn when encountering a corner. For example, one might walk back and forth along the same corridor/aisle. In addition, possible gyroscope drift errors can also produce false turning events. Therefore, detection of a turning event is not necessarily an indication that the user is indeed passing a corner. We consider these kinds of turning events, which are detected when the user is not passing a corner, as ‘fake’ turnings. Nevertheless, it is difficult to distinguish normal corner turnings from fake one based only on accelerometer and gyroscope data. In this study we utilize the floor map information to resolve the issue of fake turnings, as follows. We assume that, once the fake turnings are removed from the trajectory data, the trajectory data should be geometrically similar to a possible route on the map.

4.2. Filter Mechanism

We first try to find all the possible routes that a user might take (say, RT) based on all the possible combinations of detected turnings from the trajectory data. We then compare these RT with the all the possible routes on the map (say, RM). The objective here is to find an (RT, RM) pair which is “the most geometrically similar”. We use some methods from image processing theory to solve this problem. We first model the map as a graph, GM(VM,EM). VM is the corner of map, such as A in Figure 13(a), and the EM denotes the corridor between two adjacent corners. We also define the graph GMi(VMi, EMi) as the sub-graph of GM, which is used to model all possible routes on a map. In addition, we model the user trajectory as a graph GT(VT, ET), where VT stands for an ordered set of detected turnings (including fake ones), such as A′,B′,C′,D′ in Figure 13(b), and ET is the edge set whose element is the connection between two adjacent detected turnings. That is, assuming VT = {V1, V2, …, Vk}, Vk is the k-th detected turning, then E T = { V 1 V 2 ¯ , V 2 V 3 ¯ , ... , V k 1 V k ¯ }. We next define the graph GTj(VTj, ETj) as follows. There exists a set V′ which is the power set of VT (the set that contains all subsets of VT), i.e., V′= {ϕ, {V1}, {V2},… {V1, V2}, {V1, V3},… {V1, V2, V3,…Vk}}. Here we let VTj be an ordered set, VTjV′ and |VTj| > 1. The ETj is the edge set which contains the edge between any two adjacent elements in VTj. In other words, GTj is used to model all possible routes that could be generated based on the detected turnings (including fake ones), and GMi stands for the accessible route on the map.

Again, the idea here is to find a (GMi, GTj) pair which is the most geometrically similar. In other words, we want to eliminate those hypothetical routes generated in GTj that cannot be found on the real map. We consider two graphs are geometrically similar if they have similar shapes, angles (i.e., the angle between two connected edges) and edge lengths (after normalization). We design a two-phase filtering mechanism and employ three filters: shape filter, angle filter and edge filter. In phase one, we input all (GMi, GT) pairs into these three filters to remove those which are not geometrically similar. The purpose of phase two is to remove the non-existing routes caused by fake turnings, and we input all (GMi, GTj) pairs into these filters to remove the non-existing routes, as shown in Figure 14.

The aim of the above two-phase filtering mechanism is to produce a (GMi, GTj) pair which is geometrically similar. However, when we do not have “sufficient” trajectory data (e.g., when there is no detected turning in the trajectory data), it is possible that we can still have multiple candidate routes remaining after the geometry-similarity filtering. Therefore, when multiple candidate routes exist, we employ an RSSI-based filter by using the existing WiFi-signal-based landmarks, as described previously, to further select the correct one among multiple candidates. Next, we discuss the details of each filter.

4.2.1. The Shape Filter

We adopt the idea of a shape descriptor [66,67] to implement our shape filter by comparing the shapes of two graphs. Considering the nature of our input data and the computational overhead, we modify the original shape descriptor method to suit our scenario. To reduce the computational overhead, we calculate the centroid [68,69] of each graph and create a line every 10 degrees from 0° to 180° to pass through this. We then calculate how many crossing points on the edge can be made by each line and record them in a one-dimensional array, as shown in Figure 17. Finally, we compare the arrays generated based the two graphs to determine their similarity using Euclidean distance (L-2 norm) [70]. We set a threshold to judge the similarity between these two graphs. If the value is over the threshold, the system will consider these two graphs are different and remove them from the set of candidates.

4.2.2. The Angle Filter

We adopt the concept of chain code [71] to implement the angle filter. The chain code uses a sequence of numbers to represent a series of different moving directions and transform a graph to a one-dimensional expression. However, we cannot directly apply a full-fledged chain code to our system. First, it is difficult to decide the number of sampling points for the trajectory data, since each step distance could be different. Second, the computational overhead of using the chain code is proportional to the number of steps in the trajectory data and the number of candidate routes on the map. Given that what we consider here is a real-time localization system and the computation capability of a smart-phone is limited, we cannot just naively use the chain code. Therefore, we adopt the concept of the chain code and implement it separately via two different filters: the angle filter and the edge filter. Conceptually, the outputs of the chain code include the angle information (e.g., A and A′ in Figure 13), the direction of the edge and the normalized edge ratio (i.e., we divide the distance of each edge by the distance of the longest edge). In our angle filter, for example, we compare every matching angle and the direction of each edge between GT and GMi, and use a threshold to determine whether they are all close enough, as shown in Figure 18. After filtering out those GMi which do not have similar angles, we then implement the 2nd part of the chain code with an edge filter, as described below.

4.2.3. The Edge Filter

With this filter, we check whether the normalized edge ratios between two graphs, for example, GTj and GMi, are similar or not. The system calculates the displacement between any two adjacent vertices in the graph and stores these displacements as a vector.

Figure 19 shows an example input to the edge filter. We then use the Euclidean distance and set a threshold to determine whether the vectors produced for GTj and GMi are similar or not. If the value of the L-p norm of these two graphs is over the threshold, the system then removes the corresponding (GTj, GMi) pair from the candidates.

5. Evaluation

In this section, we discuss the performance of our PDR-based localization system. We first discuss the results of our step length estimation method, and then show the performance of the above-mentioned map matching algorithm.

5.1. Experiment Setup

We use two different sensor platforms in our experiments: Taroko motes and a variety of Android phones from HTC and Samsung. Taroko is a modified version of the TelosB mote [72], which was originally designed by UC Berkeley. Taroko is a programmable, low-power wireless sensor platform. The Taroko microcontroller unit is a TI MSP430 F1611 with 16 bit RISC [73]. MSP430 has 48 K bytes flash memory and 10 K bytes RAM that supports serial communication formats such as UART, I2C, SPI, and Digital I/O. Taroko is also equipped with a CC2420 RF transceiver [74], which is a low cost device for wireless communication in 2.4 GHz based on IEEE 802.15.4 [75]. The maximum radio distance is around 100 m. Taroko also supports the USB interface using an FTDI chip [76], and can use this to connect to a computer for powering, program upload and data collection. Taroko can accommodate different sensor components, e.g., temperature, humidity, infrared, accelerometer, gyroscope, and magnetic. In this work, we use an extended board to combine the accelerometer (ADXL330 [77]) and gyroscope (IDG500 [78]), as shown in Figure 20. The ADXL330 is a small, low power, 3-axis accelerometer with signal conditioned voltage outputs, all on a single monolithic IC. It can measure acceleration with a minimum full-scale range of ±3 g. The IDG-500 is an angular rate sensor and its full scale range is ±500°/s. It uses the MEMS technology with vertically driven, vibrating masses to make a functionally complete, low-cost angular rate sensor. All the required electronics are integrated onto a single chip with the sensor.

We first implement our algorithms on Taroko motes for ease of debugging. We then port the same algorithms into smart-phones after testing them on the motes. We perform two set of experiments on the smart-phones. One is placing the smart-phone in a shirt pocket and the other is putting it in a pants pocket. We use a laser distance meter to measure the actual travel distance of the user. In addition, to record the user's actual location, we pasted markers on the ground at precisely known locations. Each of these markers had a number on it, and the user recorded the numbers when they walked passed them. The results based on the Taroko motes are similar to those from the smart-phones.

5.2. Step Length Estimation

As discussed previously, we use a low-pass filter to filter out the noise in our step length estimation algorithm. We choose 8 Hz as our cut-off frequency for filtering the noise. To examine whether this chosen frequency produces the best results, we perform a set of experiments using different frequencies for the low-pass filter, ranging from 3 to 12 Hz, and compare their accuracies in estimating step length. As shown in Figure 21, the use of 8 Hz as the cut-off frequency produces the most accurate results.

Some of the state-of-the-art pedometers on the market can also output walking distance. We compare our method with these pedometers and find that they only achieve up to 90% accuracy, which is significantly lower than our results (98.25%). Furthermore, as discussed previously, our approach is based on the idea of using the change in height of the waist to estimate step length, which is similar to the method used in Weinberg [31]. He proposed an equation to estimate distance by observing the vertical acceleration during walking. We implement Weinberg's method and find that its accuracy is about 96.7%. However, one limitation of Weinberg's approach is that its system parameters need to be re-trained for every new user.

In our system, we use a low-pass filter (LPF) to filter out the noise and employ zero velocity update (ZUPT) to calibrate the drift error of the accelerometer. To understand the effect of these mechanisms, we enable only one of them at a time. As shown in Table 1, ZUPT has a higher impact on the system performance than the implementation of the low-pass filter. When ZUPT is disabled, the accuracy drops from 98.25% to 84.1%.

We also compare our results with those from existing foot-mounted PDR systems. As shown in Table 2, the performance of our method is close to that of the state-of-the-art foot-mounted approach [17]. Note that, although the foot-mounted method can be used to estimate step length, it is generally not good at estimating the direction a person is heading in, as noted previously.

Considering the variations that might exist in different individuals' walking gaits, we performed an experiment in which we asked the subject to walk with different speeds, as shown in Figure 22. Here the step frequency is defined as the number of walking steps in one second. The average accuracy is about 98.26% and standard deviation is about 1.09%. In addition, we asked two other persons (one male and one female) to perform the same experiment for the walking distance measurements. Each experiment, again, is repeated for 10 times. Their average accuracy and standard deviation are (98.42%, 0.98%) and (97.93%, 1.86%) respectively, which are similar to the results shown in Table 1.

5.3. Map Matching

As discussed above, we use a sliding-window-based algorithm to detect the user's possible turnings from the gyroscope data. To determine if a person is making a turn, we compare the standard deviation of the window with a threshold. Generally, when the chosen window size is too small, all the walking steps that happen during a turning might not be able to be included within a window. On the other hand, when the window size is too big, two different turnings can be included in the same window if the number of walking steps between two corners is less than the window size. As shown in Figures 23 and 24, the accuracy of detecting turnings becomes lower when the chosen window size is too small or too big. Therefore, in our experiments we choose the window size from the range defined below, in which D is the shortest distance between two adjacent corridors:

2 < WindowSize < D AvgStepLength

The threshold is obtained using the standard deviation of the window during the period when the user is walking straight. In the shape filter, we set a line, every 10 degrees from 0° to 180°, that passes through the centroid of the graph, and use the crossing points on the edges obtained in this way to determine if two graphs have similar shapes. To understand the effects of the gaps between two crossing lines (default 10°) on determining shape similarity, we vary this gap from 0° to 70°. Generally, the system will have less computational overhead when the gap is larger. However, a larger gap also suggests that poorer results will be obtained, since fewer crossing points will be generated for the comparison of shape similarity.

As shown in Figure 25, we start getting inaccurate results when the gap is larger than 15°.

Finally, to test the performance of our localization system, we choose a route which is about 40 meters long and includes four corridors and corners, as shown in Figure 26. We deliberately created some ‘fake turnings’ by having the user wander about around at certain spots, shown as the dashed circle in Figure 26. The solid circles in Figure 26 indicate when the user made a turn at the corner. We repeated this experiment 10 times and our results show that our location error is about 0.48 meter, and the standard deviation is about 0.43 meter.

6. Conclusions

In this paper we propose a novel method to accurately estimate the step length of the user based on the change in waist height. We use an accelerometer to measure the user's instant change in height, and utilize the characteristics of simple harmonic motion during walking to calibrate the drift errors of the accelerometer when calculating the change in height. Based on the Pythagorean Theorem, we can then estimate each step length using the user's change in height. Furthermore, we design a map matching algorithm to calibrate the direction errors from the gyroscope using building floor plans, which are readily available. Our map matching algorithm implements three filters, namely the shape filter, angle filter and edge filter, to infer the user's last-visited corner. Our results show that we can achieve about 98.26% accuracy in estimating the user's walking distance, while the overall location error in our experiment is about 0.48 meters.

Acknowledgments

This work was supported by the National Science Council of Taiwan under Grants NSC 101-2220-E-006-010 and NSC 101-2221-E-006-098-MY3.

References

  1. Liu, H.; Darabi, H.; Banerjee, P.; Liu, J. Survey of wireless indoor positioning techniques and systems. IEEE Trans. Syst. Man Cybern. C 2007, 37, 1067–1080. [Google Scholar]
  2. Carl, F.; Hans, G. Location and navigation support for emergency responders: A survey. IEEE Pervasive Comput. 2010, 9, 38–47. [Google Scholar]
  3. Bahl, P.; Padmanabhan, V.N. RADAR: An In-Building RF-based User Location and Tracking System. Proceedings of the Nineteenth Annual Joint Conference of the IEEE Computer and Communications Societies 2000 (INFOCOM 2000), Tel Aviv, Israel, 26–30 March 2000; Volume 2, pp. 775–784.
  4. Nissanka, B.P.; Anit, C.; Hari, B. The Cricket Location-support System. Proceedings of the 6th Annual International Conference on Mobile Computing and Networking, Boston, MA, USA, 6–11 August 2000.
  5. Seshadri, V.; Zaruba, G.V.; Huber, M. A Bayesian Sampling Approach to In-door Localization of Wireless Devices Using Received Signal Strength Indication. Proceedings of the Third IEEE International Conference on Pervasive Computing and Communications (PerCom 2005), Kauai Island, HI, USA, 8–12 March 2005; pp. 75–84.
  6. Cesare, A.; Giovanni, V. A RSSI-based and Calibrated Centralized Localization Technique for Wireless Sensor Networks. Proceedings of the 4th Annual IEEE International Conference on Pervasive Computing and Communications Workshops, Pisa, Italy, 13–17 March 2006.
  7. Sugano, M.; Kawazoe, T.; Ohta, Y.; Murata, M. Indoor Localization System Using RSSI Measurement of Wireless Sensor Network Based on Zigbee Standard. Proceedings of the IASTED International Conference on Wireless Sensor Networks (WSN 2006), Banff, Canada, 3–4 July 2006.
  8. Rolfe, B.F.; Ekanayake, S.W.; Pathirana, P.N.; Palaniswami, M. Localization with Orientation Using RSSI Measurements: RF Map Based Approach. Proceedings of 3rd International Conference on Intelligent Sensors, Sensor Networks and Information 2007 (ISSNIP 2007), Melbourne, Australia, 3–6 December 2007; pp. 311–316.
  9. Roy, W.; Andy, H.; Veronica, F.; Jonathan, G. The active badge location system. ACM Trans. Inf. Syst. 1992, 10, 91–102. [Google Scholar]
  10. Ho-lin, C.; Jr-ben, T.; Tsung-Te, L.; Hao-Hua, C.; Polly, H. Spinning Beacons for Precise Indoor Localization. Proceedings of the 6th ACM Conference on Embedded Network Sensor Systems, Raleigh, NC, USA, 4–7 November 2008.
  11. Tsung-Han, L.; Ng, I.-H.; Seng-Yong, L.; Chen, K.-M.; Polly, H. A Microscopic Examination of an RSSI-signature-Based Indoor Localization System. Fifth Workshop on Embedded Networked Sensors, Charlottesville, VI, USA, 2–3 June 2008; pp. 2–6.
  12. Cinotti, T.S.; Stefano, L.D.; Raffa, G.; Roffia, L.; Pettinari, M.; Mola, M. Dead Reckoning Supports Stereo Vision in Pedestrians Tracking. Proceedings of Fourth IEEE International Conference on Pervasive Computing and Communications Workshops (PERCOMW'06), Pisa, Italy, 13–17 March 2006.
  13. Carl, F.; Kavitha, M.; Mike, H.; Hans, G. Ultrasound-aided Pedestrian Dead Reckoning for Indoor Navigation. Proceedings of the First ACM International Workshop on Mobile Entity Localization and Tracking in GPS-less Environments, San Francisco, CA, USA, 14–19 September 2008.
  14. Renaudin, V.; Yalak, O.; Tomé, P.; Merminod, B. Indoor navigation of emergency agents. Eur. J. Navig. 2007, 5, 36–45. [Google Scholar]
  15. Ruiz, A.R.J.; Granja, F.S.; Honorato, J.C.P.; Rosas, J.I.G. Pedestrian Indoor Navigation by Aiding a Foot-mounted IMU with RFID Signal Strength Measurements. Proceedings of the 2010 International Conference on Indoor Positioning and Indoor Navigation (IPIN), Hoenggerberg, Switzerland, 15–17 September 2010; pp. 1–7.
  16. Burcu, C. HeadSLAM—Simultaneous Localization and Mapping with Head-mounted Inertial and Laser Range Sensors. Proceedings of the 12th IEEE INTERNATIONAL Symposium on Wearable Computers, Pittsburgh, PA, USA, 28 September–1 October 2008; pp. 3–10.
  17. Ojeda, L.; Borenstein, J. Personal Dead-reckoning System for GPS-denied Environments. Proceedings of the 2007 IEEE International Workshop on Safety, Security and Rescue Robotics (SSRR 2007), Rome, Italy, 27–29 September 2007; pp. 1–6.
  18. Alvarez, J.C.; Gonzalez, R.C.; Alvarez, D.; Lopez, A.M.; Rodriguez-Uria, J. Multisensor Approach to Walking Distance Estimation with Foot Inertial Sensing. Proceedings of the 29th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBS 2007), Lyon, France, 23–26 August 2007; pp. 5719–5722.
  19. Dippold, M. Personal Dead Reckoning with Accelerometers. Proceedings of the 3rd International Forum on Applied Wearable Computing 2006 (IFAWC2006), Bremen, Germany, 15–16 March 2006.
  20. Eric, F. Pedestrian tracking with shoe-mounted inertial sensors. IEEE Comput. Graph. Appl. 2005, 25, 38–46. [Google Scholar]
  21. Jimenez, A.R.; Seco, F.; Prieto, C.; Guevara, J. A Comparison of Pedestrian Dead-Reckoning Algorithms Using a low-cost MEMS IMU. Proceedings of the 2009 IEEE International Symposium on Intelligent Signal Processing (WISP 2009), Budapest, Hungary, 26–28 August 2009; pp. 37–42.
  22. Ojeda, L.; Borenstein, J. Non-GPS Navigation for Emergency Responders. Proceedings of the International Joint Topical Meeting: Sharing Solutions for Emergencies and Hazardous Environments, Salt Lake City, UT, USA, 12–15 February 2006.
  23. Ojeda, L.; Borenstein, J. Non-GPS navigation for security personnel and first responders. J. Navig. 2007, 60, 391–407. [Google Scholar]
  24. Oliver, W.; Robert, H. Pedestrian Localisation for Indoor Environments. Proceedings of the 10th International Conference on Ubiquitous Computing, Seoul, Korea, 21–24 September 2008.
  25. Feliz, R.; Zalama, E.; Gómez, J. Pedestrian tracking using inertial sensors. J. Phys. Agents 2009, 3, 35–42. [Google Scholar]
  26. Sagawa, K.; Inooka, H.; Satoh, Y. Non-restricted measurement of walking distance. IEEE Int. Conf. Syst. Man Cybern. 2000, 3, 1847–1852. [Google Scholar]
  27. Xiaoping, Y.; Bachmann, E.R.; Moore, H.; Calusdian, J. Self-contained Position Tracking of Human Movement Using Small Inertial/Magnetic Sensor Modules. Proceedings of the 2007 IEEE International Conference on Robotics and Automation, Roma, Italy, 10–14 April 2007; pp. 2526–2533.
  28. Alvarez, D.; Gonzalez, R.C.; Lopez, A.; Alvarez, J.C. Comparison of Step Length Estimators from Weareable Accelerometer Devices. Proceedings of the 28th Annual International Conference of the IEEE in Engineering Medicine and Biology Society (EMBS '06), New York, NY, USA, 30 August–3 September 2006; pp. 5964–5967.
  29. Lei, F.; Antsaklis, P.J.; Montestruque, L.A.; McMickell, M.B.; Lemmon, M.; Yashan, S.; Hui, F.; Koutroulis, I.; Haenggi, M.; Min, X.; Xiaojuan, X. Design of a wireless assisted pedestrian dead reckoning system—the NavMote experience. IEEE Trans. Instrum. Meas. 2005, 54, 2342–2358. [Google Scholar]
  30. Shin, S.H.; Park, C.G.; Hong, H.S.; Lee, J.M. MEMS-Based Personal Navigator Equipped on the User's Body. the ION GNSS 18th International Technical Meeting of the Satellite Division, Long Beach, CA, USA, 13–16 September 2005.
  31. Weinberg, H. Using the ADXL202 in Pedometer and Personal Navigation Applications. In Application Notes American Devices; Analog Devices, Inc.: Norwood, MA, USA, 2002. [Google Scholar]
  32. Shin, S.H.; Park, C.G.; Kim, J.W.; Hong, H.S.; Lee, J.M. Adaptive Step Length Estimation Algorithm Using Low-Cost MEMS Inertial Sensors. Proceedings of the IEEE Sensors Applications Symposium, SAS '07, San Diego, CA, USA, 6–8 February 2007; pp. 1–5.
  33. Stirling, R.; Collin, J.; Fyfe, K.; Lachapelle, F. An Innovative Shoe-mounted Pedestrian Navigation System. Proceedings of The European Navigation Conference GNSS 2003, Graz, Austria, 22–25 April 2003.
  34. Titterton, D.H.; Weston, J.L. Strapdown Inertial Navigation Technology; Peter Peregrinus Ltd: London, 1997. [Google Scholar]
  35. Ishikawa, T.; Kourogi, M.; Okuma, T.; Kurata, T. Economic and Synergistic Pedestrian Tracking System for Indoor Environments. Proceedings of the 2009 International Conference of Soft Computing and Pattern Recognition, Malacca, Malaysia, 4–7 December 2009; pp. 522–527.
  36. Ascher, C.; Kessler, C.; Wankerl, M.; Trommer, G.F. Dual IMU Indoor Navigation with Particle Filter based Map-matching on a Smartphone. Proceedings of the 2010 International Conference on Indoor Positioning and Indoor Navigation (IPIN), Hoenggerberg, Switzerland, 15–17 September 2010; pp. 1–5.
  37. Gusenbauer, D.; Isert, C.; krösche, J. Self-Contained Indoor Positioning on off-the-Shelf Mobile Devices. Proceedings of the 2010 International Conference on Indoor Positioning and Indoor Navigation (IPIN), Hoenggerberg, Switzerland, 15–17 September 2010; pp. 1–9.
  38. Google Latitude. Available online: http://www.google.com/intl/en/mobile/latitude/ (accessed on 14 June 2013).
  39. Savvides, A.; Han, C.C.; Srivastava, M.B. Danamic Fine-grained Localization in Ad-hoc Wireless Sensor Networks. Proceedings of the Proceedings of the 7th annual international conference on Mobile computing and networking, Rome, Italy, 16–21 July 2001; pp. 166–179.
  40. Li, X.; Pahlavan, K.; Latva-aho, M.; Ylianttila, M. Comparison of indoor geolocation methods in DSSS and OFDM wireless LAN. Proceedings of the IEEE Vehicular Technology Conference, Boston, MA, USA, 24–28 September 2000; Volume 6, pp. 3015–3020.
  41. KNN. Available online: http://en.wikipedia.org/wiki/K-nearest_neighbor_algorithm (accessed on 4 November 2012).
  42. Brunato, M.; Battiti, R. Statistical learning theory for location fingerprinting in wireless LANs. Comput. Netw. 2005, 47, 825–845. [Google Scholar]
  43. Chumkamon, S.; Tuvaphanthaphiphat, P.; Keeratiwintakorn, P. A Blind Navigation System Using RFID for Indoor Environments. Proceedings of 5th International Conference on Electrical Engineering/Electronics, Computer, Telecommunications and Information Technology, (ECTI-CON 2008), Krabi, Thailand, 14–17 May 2008; pp. 765–768.
  44. Hightower, J.; Want, R.; Borriello, G. SpotON: An Indoor 3D Location Sensing Technology Based on RF Signal Strength; Technical Report UW CSE 2000-02-02; University Washington: Seattle, WA, USA; February; 2000. [Google Scholar]
  45. Ni, L.M.; Liu, Y.; Lau, Y.C.; Patil, A.P. LANDMARC: Indoor Location sensing using active RFID. Wireless Netw. 2004, 10, 701–710. [Google Scholar]
  46. Godha, S.; Lachapelle, G.; Cannon, M.E. Integrated GPS/INS System for Pedestrian Navigation in a Signal Degraded Environment. Proceedings of the 19th International Technical Meeting of the Satellite Division of The Institute of Navigation (ION GNSS 2006), Fort Worth, TX, USA, 26–29 September 2006.
  47. Ladetto, Q.; Merminod, B. In step with INS—Navigation for the blind, tracking emergency crews. GPS World 2002, 13, 30–38. [Google Scholar]
  48. Krishna, C.; Padmanabha, I.A.; Venkata, N.P. Indoor Localization without the Pain. Proceedings of the Sixteenth Annual International Conference on Mobile Computing and Networking, Chicago, IL, USA, 20–24 September 2010.
  49. Ji, Y.; Biaz, S.; Pandey, S.; Agrawal, P. ARIADNE: A Dynamic Indoor Signal Map Construction and Localization System. Proceedings of the 4th International Conference on Mobile Systems, Applications, and Services, Uppsala Sweden, 19–22 June 2006.
  50. Lim, H.; Kung, C.L.; Hou, J.C.; Luo, H. Zero Configuration Robust Indoor Localization: Theory and Experimentation. Proceedings of the 25th Conference on Computer Communications (InfoCom 2006), Barcelona Catalunya, Spain, 23–29 April 2006.
  51. Shih, C.-C. The Implementation of a G-Sensdor-based Pedometer. M.Sc. Thesis, National Central University, Taoyuan County, Taiwan, 2010. [Google Scholar]
  52. Jorgensen, F. Accurate Step Counter. U.S. Patent 0 172 203 A1, 17 July 2008. [Google Scholar]
  53. Martin, M.; Michael, M. A Step Counter Service for Java-enabled Devices Using a Built-in Accelerometer. Proceedings of the 1st International Workshop on Context-Aware Middleware and Services: Affiliated with the 4th International Conference on Communication System Software and Middleware (COMSWARE 2009), Dublin, Ireland, 16 June 2009.
  54. Schneider, P.L.; Crouter, S.E.; Lukajic, O.; Bassett, D.R. Accuracy and reliability of ten pedometers for measuring steps over a 400-m walk. Med. Sci. Sports Exerc. 2003, 35, 1779–1784. [Google Scholar]
  55. Li, F.; Zhao, C.; Ding, G.; Gong, J.; Liu, C.; Zhao, F. A Reliable and Accurate Indoor Localization Method Using Phone Inertial Sensors. Proceedings of the 14th ACM International Conference on Ubiquitous Computing (Ubicomp'12), Pittsburgh, PA, USA, 5–8 September 2012.
  56. Su, C.-M.; Chou, J.-W.; Yi, C.-W.; Tseng, Y.-C.; Tsai, C.-H. Sensor-aided Personal Navigation Systems for Handheld Devices. Proceedings of the 2010 39th International Conference on Parallel Processing Workshops, (ICPPW'10), Washington DC, USA, 13–16 September 2010; pp. 533–541.
  57. Simple Harmonic Motion. Available online: http://en.wikipedia.org/wiki/Simple_harmonic_motion (accessed on 24 November 2011).
  58. Gerald, B.; Andreas, T. DiaTrace—Neuartiges Assistenz-System für die Gesundheitsprävention zur Nahrungsaufnahme und Bewegungserfassung. Proceeding of Ambient Assisted Living, Berlin, Germany, 30 January–1 February 2008.
  59. Chen, K.Y.; Bassett, D.R. The technology of accelerometry—Based activity: Monitors: Current & future. Med. Sci. Sports Exerc. 2005, 37, S490–S500. [Google Scholar]
  60. Walking diagram. Available online: http://www.juliecallahan.me/walkcycle.jpg (accessed on 14 July 2012).
  61. Gafurov, D.; Snekkenes, E.; Bours, P. Gait Authentication and Identification Using Wearable Accelerometer Sensor. Proceedings of the 2007 IEEE Workshop on Automatic Identification Advanced Technologies, Alghero, Italy, 7–8 June 2007; pp. 220–225.
  62. Chintalapudi, K.; Iyer, A.P.; Padmanabhan, N. Indoor Localization Without the Pain. Proceedings of the 16th ACM International Conference on Mobile Computing and Networking (MobiCom 2010), Chicago, IL, USA, 20–24 September 2010; pp. 173–184.
  63. Aggarwal, P.; Thomas, D.; Ojeda, L.; Borenstein, J. Map matching and heuristic elimination of gyro drift for personal navigation systems in GPSdenied conditions. Meas. Sci. Technol. 2011, 22, 025205. [Google Scholar]
  64. Gillieron, P.-Y.; Merminod, B. Personal Navigation System for Indoor Applications. Proceedings of the 11th IAIN World Congress on Smart Navigation, Systems and Services, Berlin, Germany, 21–24 October 2003.
  65. Wang, H.; Sen, S.; Elgohary, A.; Farid, M.; Youssef, M.; Choudhury, R.R. No Need to War-Drive: Unsupervised Indoor Localization. Proceedings of the 10th International Conference on Mobile Systems, Applications and Services, Low Wood Bay, Lake District, UK, 25–29 June 2012.
  66. Hao, Z.; Malik, J. Learning a Discriminative Classifier Using Shape Context Distances. Proceedings of the 2003 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Madison, WI, USA, 18–20 June 2003; Volume 1, pp. 242–247.
  67. Belongie, S.; Malik, J.; Puzicha, J. Shape matching and object recognition using shape contexts, Pattern Analysis and Machine Intelligence. IEEE Trans. Pattern Anal. Mach. Intell. 2002, 24, 509–522. [Google Scholar]
  68. Johnson, R.A. Modern Geometry: An Elementary Treatise on the Geometry of the Triangle and the Circle; Houghton Mifflin: Boston, MA, USA, 1929. [Google Scholar]
  69. Centroid. Available online: http://en.wikipedia.org/wiki/Centroid#cite_ref-0 (accessed on 22 May 2012).
  70. Dunford, N.; Schwartz, J.T. Linear Operators; Interscience Publishers: New York, NY, USA, 1958; Volumn I. [Google Scholar]
  71. Saghri, J.A.; Freeman, H. Analysis of the precision of generalized chain codes for the representation of planar curves, pattern analysis and machine intelligence. IEEE Trans. Pattern Anal. Mach. Intell. 1981, PAMI-3, 533–539. [Google Scholar]
  72. TelosB. Available online: http://www.willow.co.uk/html/telosb_mote_platform.html (accessed on 20 September 2010).
  73. MSP430 F1611. Available online: http://focus.ti.com/docs/prod/folders/print/msp430f1611.html (accessed on 20 September 2010).
  74. CC2420. Available online: http://focus.ti.com/docs/prod/folders/print/cc2420.html (accessed on 20 September 2010).
  75. 802.15.4. Available online: http://www.ieee802.org/15/pub/TG4.html (accessed on 20 September 2010).
  76. FTDI. Available online: http://www.ftdichip.com/Documents/DataSheets/ds232b18.pdf (accessed on 20 September 2010).
  77. ADXL 330. Available online: http://www.analog.com/en/mems-sensors/inertial-sensors/adxl330/products/product.html (accessed on 20 September 2010).
  78. IDG 500. Available online: http://invensense.com/mems/gyro/idg500.html (accessed on 20 September 2010).
Figure 1. (a) A detailed map and (b) a floor plan.
Figure 1. (a) A detailed map and (b) a floor plan.
Sensors 13 04781f1 1024
Figure 2. The entire system architecture.
Figure 2. The entire system architecture.
Sensors 13 04781f2 1024
Figure 3. Walking Diagram.
Figure 3. Walking Diagram.
Sensors 13 04781f3 1024
Figure 4. The orientation of the device.
Figure 4. The orientation of the device.
Sensors 13 04781f4 1024
Figure 5. The diagram of device in the lower body.
Figure 5. The diagram of device in the lower body.
Sensors 13 04781f5 1024
Figure 6. Flow Chart of PDR.
Figure 6. Flow Chart of PDR.
Sensors 13 04781f6 1024
Figure 7. Gravity Elimination.
Figure 7. Gravity Elimination.
Sensors 13 04781f7 1024
Figure 8. (a) Gravity elimination in different directions; (b) Force Diagram.
Figure 8. (a) Gravity elimination in different directions; (b) Force Diagram.
Sensors 13 04781f8 1024
Figure 9. (a) Walking diagram (modified from an earlier work [60]); (b) The vertical acceleration of walking.
Figure 9. (a) Walking diagram (modified from an earlier work [60]); (b) The vertical acceleration of walking.
Sensors 13 04781f9 1024
Figure 10. The velocity with ZUPT versus without ZUPT.
Figure 10. The velocity with ZUPT versus without ZUPT.
Sensors 13 04781f10 1024
Figure 11. The difference between normal and high speeds.
Figure 11. The difference between normal and high speeds.
Sensors 13 04781f11 1024
Figure 12. The acceleration at different high speeds.
Figure 12. The acceleration at different high speeds.
Sensors 13 04781f12 1024
Figure 13. (a) Link-model of the map and the ground truth (b) the estimated trajectory from the sensor data.
Figure 13. (a) Link-model of the map and the ground truth (b) the estimated trajectory from the sensor data.
Sensors 13 04781f13 1024
Figure 14. The flow chart of our map-matching algorithm.
Figure 14. The flow chart of our map-matching algorithm.
Sensors 13 04781f14 1024
Figure 15. The link-node model from a floor plan.
Figure 15. The link-node model from a floor plan.
Sensors 13 04781f15 1024
Figure 16. An example of turning detection.
Figure 16. An example of turning detection.
Sensors 13 04781f16 1024
Figure 17. Example of the shape filter.
Figure 17. Example of the shape filter.
Sensors 13 04781f17 1024
Figure 18. Features of the angle filter.
Figure 18. Features of the angle filter.
Sensors 13 04781f18 1024
Figure 19. The input to the edge filter.
Figure 19. The input to the edge filter.
Sensors 13 04781f19 1024
Figure 20. The Taroko with ADXL330 and IDG500.
Figure 20. The Taroko with ADXL330 and IDG500.
Sensors 13 04781f20 1024
Figure 21. The accuracy using different cut-off frequencies for the low pass filter.
Figure 21. The accuracy using different cut-off frequencies for the low pass filter.
Sensors 13 04781f21 1024
Figure 22. Accuracy of estimated distance under different walking speeds.
Figure 22. Accuracy of estimated distance under different walking speeds.
Sensors 13 04781f22 1024
Figure 23. The number of walking steps between two corners is more than the window size.
Figure 23. The number of walking steps between two corners is more than the window size.
Sensors 13 04781f23 1024
Figure 24. The number of walking steps between two corners is less than the window size.
Figure 24. The number of walking steps between two corners is less than the window size.
Sensors 13 04781f24 1024
Figure 25. The discrimination of different slopes.
Figure 25. The discrimination of different slopes.
Sensors 13 04781f25 1024
Figure 26. The testing environment and route.
Figure 26. The testing environment and route.
Sensors 13 04781f26 1024
Table 1. Comparison with other waist-mounted methods and the influence of the mechanism.
Table 1. Comparison with other waist-mounted methods and the influence of the mechanism.
PedometerWeinbergOurs (A:LPF, B:ZUPT)
MethodConstant step length × step number k × n × A c c max A c c min 4Original (A+B)ABNone
Accuracy (Estimation/Ground Truth)90.09%96.78%98.25%84.1%95.29%75.53%
Standard deviation1.83%3.18%1.29%7.41%2.22%16.3%
Table 2. Comparison with foot-mounted PDR.
Table 2. Comparison with foot-mounted PDR.
Our ApproachFoot-mounted method (L. Ojeda 2007)
Hardware3-axis accelerometer & gyroscope6-DoF inertial sensor
(accelerometer + Gyro + Magnetic)
Sample Rate25Hz200Hz
PlacementWaistFoot
ZUPTUse the concept of SHMCalculate the angular velocity to detect zero velocity
Accuracy98.25±1.29% (total distance)>98% (total distance)
>99% (coordinate)

Share and Cite

MDPI and ACS Style

Lan, K.-C.; Shih, W.-Y. On Calibrating the Sensor Errors of a PDR-Based Indoor Localization System. Sensors 2013, 13, 4781-4810. https://doi.org/10.3390/s130404781

AMA Style

Lan K-C, Shih W-Y. On Calibrating the Sensor Errors of a PDR-Based Indoor Localization System. Sensors. 2013; 13(4):4781-4810. https://doi.org/10.3390/s130404781

Chicago/Turabian Style

Lan, Kun-Chan, and Wen-Yuah Shih. 2013. "On Calibrating the Sensor Errors of a PDR-Based Indoor Localization System" Sensors 13, no. 4: 4781-4810. https://doi.org/10.3390/s130404781

APA Style

Lan, K. -C., & Shih, W. -Y. (2013). On Calibrating the Sensor Errors of a PDR-Based Indoor Localization System. Sensors, 13(4), 4781-4810. https://doi.org/10.3390/s130404781

Article Metrics

Back to TopTop