Next Article in Journal
Peptides, DNA and MIPs in Gas Sensing. From the Realization of the Sensors to Sample Analysis
Next Article in Special Issue
A Recursive Algorithm for Indoor Positioning Using Pulse-Echo Ultrasonic Signals
Previous Article in Journal
Focus on the Visible Regions: Semantic-Guided Alignment Model for Occluded Person Re-Identification
Previous Article in Special Issue
Simulating Signal Aberration and Ranging Error for Ultrasonic Indoor Positioning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Adaptive Residual Weighted K-Nearest Neighbor Fingerprint Positioning Algorithm Based on Visible Light Communication

1
Key Laboratory of OptoElectronic Science and Technology for Medicine of Ministry of Education, College of Photonic and Electronic Engineering, Fujian Normal University, Fuzhou 350007, China
2
Concord University College, Fujian Normal University, Fuzhou 350117, China
3
Department of Aeronautical Engineering, Chaoyang University of Technology, Taichung 413310, Taiwan
4
Information and Engineering College, Jimei University, Fujian 361021, China
*
Author to whom correspondence should be addressed.
Sensors 2020, 20(16), 4432; https://doi.org/10.3390/s20164432
Submission received: 6 July 2020 / Revised: 22 July 2020 / Accepted: 6 August 2020 / Published: 8 August 2020
(This article belongs to the Special Issue Sensors and Systems for Indoor Positioning)

Abstract

:
The weighted K-nearest neighbor (WKNN) algorithm is a commonly used fingerprint positioning, the difficulty of which lies in how to optimize the value of K to obtain the minimum positioning error. In this paper, we propose an adaptive residual weighted K-nearest neighbor (ARWKNN) fingerprint positioning algorithm based on visible light communication. Firstly, the target matches the fingerprints according to the received signal strength indication (RSSI) vector. Secondly, K is a dynamic value according to the matched RSSI residual. Simulation results show the ARWKNN algorithm presents a reduced average positioning error when compared with random forest (81.82%), extreme learning machine (83.93%), artificial neural network (86.06%), grid-independent least square (60.15%), self-adaptive WKNN (43.84%), WKNN (47.81%), and KNN (73.36%). These results were obtained when the signal-to-noise ratio was set to 20 dB, and Manhattan distance was used in a two-dimensional (2-D) space. The ARWKNN algorithm based on Clark distance and minimum maximum distance metrics produces the minimum average positioning error in 2-D and 3-D, respectively. Compared with self-adaptive WKNN (SAWKNN), WKNN and KNN algorithms, the ARWKNN algorithm achieves a significant reduction in the average positioning error while maintaining similar algorithm complexity.

1. Introduction

Positioning systems can be divided into outdoor positioning system (OPS) and indoor positioning system (IPS). The OPS usually uses global positioning system (GPS) to obtain the coordinates of the target. Since the GPS signal is not able to penetrate the wall and other obstacles, GPS cannot be applied in the indoor positioning scene [1]. As a supplement to OPS, IPS has attracted increasing attention among researchers. At present, there are two main research areas on IPS. One is based on radio frequency communication technology, such as radio frequency identification (RFID) [2], wireless sensor network (WSN) [3], ultra-wideband (UWB) [4], wireless fidelity (WiFi) [5], Bluetooth [6], etc. The other is based on visible light communication (VLC) [7]. IPS can be divided into range-based IPS and range-free IPS. The methods of range-based include time of arrival (TOA), angle of arrival (AOA), and received signal strength indication (RSSI), etc. [8,9]. The range-free IPS usually uses fingerprint matching to achieve positioning [10]. Compared with radio frequency communication technology, using a light-emitting diode (LED) to achieve indoor positioning has the following advantages: (1) LED communication uses the visible light spectrum, which can be applied to some areas where electromagnetic radiation is prohibited, such as operating rooms and gas stations; (2) generally, LED is uniformly distributed on the ceiling, there is mainly line-of-sight (LoS) communication between transceivers and receivers; (3) existing LED lighting devices can be used directly, and the receiver can use integrated photodiode (PD) devices [11,12]; (4) the signal-to-noise ratio (SNR) is usually very high due to lighting requirements.
Typical fingerprint-based localization algorithms usually use machine-learning algorithms [13], for example, random forest (RF) [14], K-nearest neighbor (KNN) [15], extreme learning machine (ELM) [16], artificial neural network (ANN) [17], etc. In [11], for indoor positioning based on VLC, three classical machine-learning algorithms, RF, ELM and KNN are adopted to train multiple classifiers based on received signal strength indication (RSSI) fingerprints, and a grid-independent least square (GI-LS) algorithm was proposed to combine the outputs of these classifiers. Experimental results show that compared with RF, KNN and ELM algorithms, the positioning error based on the GI-LS algorithm is lower. In machine learning algorithms, K-nearest neighbor (KNN) is one of the most widely used fingerprint positioning. The KNN fingerprint positioning algorithm [15] works in two stages. The first one is run offline, and it consists of generating a set of fingerprint points in the application area. In the second step, the target measures an RSSI vector of M LEDs, which is then matched with the K nearest fingerprints obtained previously offline. When K fingerprints have different weights, this method is called the weighted K-nearest neighbor (WKNN) fingerprint positioning algorithm. The WKNN localization algorithm is based on the shortest physical distance between fingerprints and the target position [12,18,19], which usually adopts two ranging methods: Euclidean distance [18] and Manhattan distance [19]. In Hu et al. [19], for indoor positioning based on WiFi, a self-adaptive WKNN (SAWKNN) algorithm with a dynamic K was proposed. Experimental results show that the positioning error based on the SAWKNN algorithm is lower than that of the WKNN algorithm. In most cases, M LEDs are laid out on the ceiling of the same horizontal plane. The traditional trilateration method and least linear multiplication method can only solve the two-dimensional (2-D) coordinates of targets [20], and the height of the target from the floor needs to be known in advance, which is not feasible in many applications. A Newton–Raphson method was proposed in Şahin et al. [21] and Mathias et al. [22] to estimate the PD location. For a non-convex optimization problem of 3-D positioning, it is easy for the least linear multiplication method and Newton–Raphson method to fall into the local optimal solution, resulting in large positioning error. Particle swarm optimization [23] and differential evolution algorithm [24] are adopted to perform 3-D visible light positioning, which will increase the complexity of the algorithm. In Van et al. [25], compared with trilateration method in the case of ambient light interference and without ambient light interference, simulation results show that the positioning accuracy of the WKNN algorithm is improved by 36% and 50%, respectively. In Alam et al. [12], experiment results show that the average positioning error of the fingerprints established by Lambertian regeneration model is close to that of the actual RSSI measurement fingerprints, which are 2.7 cm and 2.2 cm, respectively. Therefore, the WKNN positioning algorithm based on VLC does not need a large number of human resources to acquire the fingerprints. In Gligorić et al. [26], a visible light localization algorithm based on compressed sensing (CS) was proposed. The orthogonal matching pursuit (OMP) reconstruction algorithm [27] is used to determine the overlapping region, and the KNN algorithm is used to determine the coordinates of the target. In Zhang et al. [28], an visible light inversion positioning system based on CS and a 4-sparse 2-D fingerprint matching algorithm was proposed. When CS is used to realize fingerprint positioning, the measurement matrix needs to satisfy the restricted isometry property (RIP) attribute, and the orthogonal decomposition of the measurement matrix is needed [29,30], which will increase the complexity of the algorithm. The fingerprint positioning algorithm based on CS must satisfy O (K log (N)), where this is the value of the number of measurements M (i.e., the number of LEDs) [29]. When the neighboring fingerprints K and the number of fingerprints N become larger, a high-density LED layout is required to satisfy the compression sensing reconstruction condition. However, an excessively dense LED layout not only wastes resources but also increases interference between LEDs. Although the WKNN algorithm is a commonly used fingerprint positioning, the difficulty lies in how to optimize the value of K to obtain the minimum positioning error. Compared with the traditional WKNN positioning algorithm, this paper makes the following contributions:
  • A novel adaptive residual weighted K-nearest neighbor (ARWKNN) fingerprint positioning algorithm is proposed, and K is a dynamic value according to the matched RSSI residual. As far as the authors know, most fingerprint positioning algorithms based on VLC only consider a fixed neighbor value, e.g., [12,25,28].
  • The impact of modulation bandwidth, transmit power, the signal-to-noise ratio, the maximum number of neighboring fingerprints, the sampling interval, the number of LEDs, the sampling ratio, and distance metric on positioning errors are analyzed in detail. The distribution of optimal K and the complexity of the algorithm are also analyzed in detail. The results can provide a useful reference for the design of the actual VLP system.
  • Simulation results show that the ARWKNN algorithm based on Clark distance (CLD) and minimum maximum distance (MMD) metrics produces the smallest average positioning error in 2-D and 3-D, respectively, as far as the authors know, this is the first work to report the impact of CLD and MMD metrics on the positioning error of the fingerprint positioning algorithm.
  • Simulation results show that when the SNR is 20 dB, in 2-D, compared with the fingerprint positioning algorithm based on RF [14], ELM [16], ANN [17], GI-LS [11], SAWKNN [19], WKNN [12] or KNN [15,25] algorithms, the average positioning error of the ARWKNN algorithm based on Manhattan distance can be reduced by 81.82%, 83.93%, 86.06%, 60.15%, 43.84%, 47.81%, and 73.36%, respectively. Compared with SAWKNN, WKNN and KNN algorithms, the ARWKNN algorithm can significantly reduce the average positioning error while maintaining similar algorithm complexity.
The rest of this paper is organized as follows: the ARWKNN algorithm is proposed in Section 2. Simulation results are shown and discussed in Section 3. Finally, Section 4 concludes this paper.
Notation: Matrices and vectors are in boldface. The field of real numbers is denoted by . ‖.‖2 is the 2 norm of the vector. | · | is the absolute value, and   denotes the rounding up operator. The transpose operation is denoted by [.]T.

2. Design of the Adaptive Residual Weighted K-Nearest Neighbor (ARWKNN) Algorithm

2.1. System Model

The positioning model is shown in Figure 1a. If there is Mtotal LEDs in the room, the target checks and selects M LEDs that has the highest RSSI for positioning. For simplicity, we assume that the target appears in a 3-D space with M LEDs. The coordinates of M LEDs are βi = [xLED-i, yLED-i, zLED-i]T, for i = 1, 2, …, M. It is assumed that M LEDs are evenly distributed on the same horizontal plane, i.e., zLED-i = zLED, zLED is the height from the floor to the LED. α i 3 × 1 represents the angle of the ith LED. θj 3 × 1 and γj 3 × 1 represent the coordinate and angle of the jth fingerprint point, respectively, for j = 1, 2, …, N, N represents the number of fingerprint points. Suppose the target moves in an interval from hL to hH at the z-axis, hL and hH are the minimum and maximum vertical distance from the floor to the target, respectively.
We use S to denote the spacing of the fingerprints, as shown in Figure 1a. m, n and l are used to represent the collection directions of fingerprints in x-axis, y-axis and z-axis, respectively, the meanings of m, n, and l are shown in Table 1. To make it easier to understand, an example is given, as shown in Figure 1b. In Figure 1b, columns are arranged from left to right (in the positive direction of the x-axis), rows are arranged from bottom to top (in the positive direction of the y-axis), and dimensions are arranged from low to high (in the positive direction of the z-axis). The starting point of fingerprint collection is θinit = [xinit, yinit, zinit]T, xinit, yinit, and zinit are given by:
{ x init = min ( x LED - i ) y init = min ( y LED - i ) z init = h L
Then, in the positioning space, the coordinates corresponding to the fingerprint points in the l dimension, i.e., the m column and the n row are
{ x fin - m = x init + S ( m 1 ) ,   m = 1 ,   2 ,     ,   L 1 S + 1 y fin - n = y init + S ( n 1 ) ,         n = 1 ,   2 ,     ,   L 2 S + 1 z fin - l = z init + S ( l 1 ) ,             l = 1 ,   2 ,     ,   L 3 S + 1
where L1 = max (xLED-i) − min (xLED-i), L2 = max (yLED-i) − min (yLED-i) and L3 = hHhL. Then the distance dl,m,ni between each fingerprint point and the ith LED can be obtained as:
d l , m , n i = ( x fin - m x LED - i ) 2 + ( y fin - n y LED - i ) 2 + ( z fin - l z LED - i ) 2

2.2. Fingerprint Matrix Construction

We use Φ M × N to denote the measurement matrix of the fingerprints, which is given by:
Φ = ( ϕ 1 , 1 , 1 - 1   ϕ 1 , 1 , 2 - 1     ϕ l , m , n - 1 ϕ 1 , 1 , 1 - 2   ϕ 1 , 1 , 2 - 2     ϕ l , m , n - 2                                                                 ϕ 1 , 1 , 1 - M   ϕ 1 , 1 , 2 - M     ϕ l , m , n - M )
where N is given by:
N = L 1 S + 1 L 2 S + 1 L 3 S + 1
And ϕ l , m , n i represents the RSSI, which is given by:
ϕ l , m , n i = 10 log 10 ( P l , m , n i )
where Pl,m,ni represents the optical power value from the ith LED received by the fingerprint point in the l dimension, m column and n row within the positioning area.

2.3. Measurement Vector

Suppose the coordinates of targets in 3-D are Ψk = [xtarget-k, ytarget-k, ztarget-k]T, for k = 1, 2, …, C, and C represents the number of targets. Thus, the receiving signal intensity vector Yk of M LEDs collected by the kth target is given by:
Y k = [ Y k , 1 , Y k , 2 ,     ,   Y k , M ] T
where Yk,i is given by
Y k , i = 10 log 10 ( P k , i )
where Pk,i represents the optical power value of the ith LED received by the kth target.

2.4. Measurement Model

In this paper, the measurement matrix Φ and measurement vector Yk are generated by the Lambertian radiation model. Because the LED is distributed on the ceiling, there is mainly LoS communication between the fingerprint point and the LED. Without loss of generality, this paper only considers the Lambertian radiation model of the LoS, which are widely adopted in papers such as [12,28,30,31,32], the received light power value of the fingerprint point is:
P Re = P Tr A P D ( b + 1 ) T s g 2 π d 2 ( cos ( λ i ) ) b cos ( ω i )
where PRe represents the received light power value; PTr represents the transmit power of the LED; d is the distance between the transmitter and the receiver; Ts and g are the optical filter gain and optical concentrator gain, respectively; b is the Lambertian order; λ1/2 is the half-power angles of the LED; APD is the effective area of the PD detection; The field of view (FOV) of PD is defined as ωFOV, and 0 < ωi < ωFOV. λi and ωi are the radiation and incident angles, i.e., the transmitter’s normal and receiver’s normal, respectively, as shown in Figure 1a.

2.5. Channel Access Method

As LEDs transmit a unique identification (ID) code independently, however, signals sent from different LEDs will interfere with each other at the receiver. In order to receive the power from different LEDs, we also use time division multiplexing to achieve this goal [20,31,32], and in a real scenario, we can also use different modulation frequencies, such as Guo et al. [11] and Alam et al. [12]. M LEDs have synchronous frames [20,31], and different LEDs use different time slots to transmit signals within each frame cycle, when one LED transmits the ID code, other LEDs emit a constant light intensity (CLI) for illumination purposes only. The frame structure is shown in Figure 2. After photoelectric conversion, a high-pass filter can be used to filter out the power from other LEDs [20].

2.6. Setting of K

According to the principle of fingerprint positioning, the purpose of positioning is to find K fingerprint points that are close to the target. When in a different experimental environment, K generally takes different values, such as in Xue et al. [15], the optimal positioning accuracy is obtained when K = 5; in Alam et al. [12] and Zhang et al. [28], the optimal positioning accuracy is obtained when K = 4; in Van et al. [25], the optimal positioning accuracy is obtained when K = 3. One thing they all have in common is that K is a fixed value. In this paper, N fingerprint points are evenly distributed in the 2-D or 3-D space. In a specific time, there are K fingerprint points close to the same target, which is called the KNN fingerprint positioning algorithm. For example, when the target exactly matches the fingerprint point, as shown in Figure 3a, obviously, the optimal positioning accuracy is obtained when K = 1. When the target falls on a straight line formed by two fingerprint points, as shown in Figure 3b, i.e., K = 2. When the target is in a triangular area composed of three fingerprint points, as shown in Figure 3c, i.e., K = 3. If the 3-D fingerprints map is adopted, and the target is obviously located in a minimum cube composed of 8 fingerprint points with a high probability, i.e., K = 8, as shown in Figure 3d.
Figure 4 is the positioning error of five targets at different 3-D positions using WKNN algorithm, for K increases from 1 to 8. When K = 4, the positioning error of target 1 is minimal. When K = 3, the positioning error of target 2 is minimal. When K = 8, the positioning error of target 3 is minimal. When K = 1, the positioning error of target 4 is minimal. When K = 6, the positioning error of target 5 is minimal. It can also be seen from Figure 4 that the positioning error varies with the K value fluctuation, and there is no monotonous increasing or decreasing relationship. In a 2-D visible light localization, the average positioning error based on the WKNN algorithm can be minimized when K = 3 or K = 4, e.g., [12,25,28]. In the 3-D visible light localization, the average positioning error based on the WKNN algorithm can be minimized when K = 8, which will be discussed in Section 3 . The minimum mean positioning error does not mean that the positioning error of each target is the smallest, so the dynamic K value can effectively reduce the positioning error of different targets. To address this issue, this paper proposes an adaptive residual weighted K-nearest neighbor fingerprint positioning algorithm, which is called ARWKNN fingerprint positioning algorithm.

2.7. ARWKNN Algorithm

The WKNN fingerprint positioning algorithm is based on the shortest RSSI physical distance between the fingerprint and the target position. The positioning error for the WKNN algorithm is affected by the weight of the fingerprint point and this weight is affected by the K value. If the optimal K value can be obtained, the positioning error can be reduced, so a novel ARWKNN algorithm is proposed in this paper. The pseudo-code of the ARWKNN algorithm is shown in Algorithm 1. In Algorithm 1, if we only consider Steps 1, 2 and 5, then it is the WKNN algorithm, and in Step 5, if the location of the target is estimated by averaging the coordinates of K fingerprints, then it is the KNN algorithm. By contrast with the KNN and WKNN algorithms, the ARWKNN algorithm also performs Step 3 and 4 in Algorithm 1. Because there is no prior information about the location of the target, that is, the value of Ψk is unknown, but we known the fingerprint matrix Φ and the target RSSI measurement vector Yk, we can adaptively select the K value by matching the residual between the measured and calculated RSSI values. Therefore, the purpose of Steps 3 and 4 in algorithm 1 is to obtain the optimal K value, i.e., the K value corresponding to the smallest RSSI matching residual. In Algorithm 1, because the maximum number of neighboring fingerprint points Kmax is much smaller than the total number of fingerprint points N, the ARWKNN algorithm has a large reduction in the average positioning error while maintaining similar algorithm complexity, which will be discussed in Section 3.4.
Algorithm 1. ARWKNN algorithm
Input: the maximum number of nearest neighbor fingerprints Kmax, fingerprint matrix Φ, and the kth target measurement vector Yk.
Output: The coordinates of the kth target, i.e., Ψk.
Step 1: Calculate the distance from the kth target to N fingerprint points.
d i s l , m , n k = ( i = 1 M | ϕ l , m , n i Y k , i | r ) 1 / r
where r = 1 represents the Manhattan distance, r = 2 represents the Euclidean distance.
Step 2: Sort the distance values in ascending order, i.e.,
[X, I] = sort (dis).
where dis = [dis1,1,1−k, dis1,1,2−k, …, disl,m,nk]T N × 1 , X N × 1 represents the distance vector after sorting, and I N × 1 represents the corresponding index set.
Step 3: Calculate the matched RSSI residuals.
K = 1,
while K K max do
  for ii = 1: K
   A ( : , i i ) = Φ ( : , I ( i i ) ) ;
  end for
where A M × K represents finding the K column values corresponding to the fingerprint matrix Φ according to the index set I.
Calculate the kth target RSSI vector via K nearest neighbor fingerprints,
Y ˜ k = A B ,
where B = [B1, B2, …, BK]T K × 1 and
B t = 1 X ( t ) / t t = 1 K 1 X ( t t ) , for t = 1, 2, …, K,
Calculate the matched RSSI residual between the measured and calculated RSSI values,
E residual = Y k Y ˜ k ,
and calculate the sum of the absolute values of the residuals,
E sum ( K ) = i = 1 M | E residual ( i ) | ,
K = K + 1.
end while
Step 4: Output the K value, i.e.,
K = arg min ( E sum ) ;   s . t .   1 K K max
Step 5: Calculate the coordinates of the kth target,
Ψ k = t = 1 K 1 X ( t ) θ I ( t ) t = 1 K 1 X ( t )
where θI(t) represents the coordinates of the corresponding fingerprint point found according to the index set I.

3. Simulation Analysis

In this Section, the ARWKNN algorithm is compared with RF [14], ELM [16], ANN [17], GI-LS [11], SAWKNN [19], WKNN [12] or KNN [15,25] algorithms. The basic principle of the fingerprint positioning algorithm based on RF, ELM, ANN, and GI-LS machine learning is as follows [11,13]: Firstly, the positioning area is divided into several equal grid points according to the sampling interval S, RSSI measurements are obtained by placing the receiver at different grid points, and each grid point represents a category. Secondly, machine-learning algorithms are used to train the category to which each grid point belongs. Thirdly, the RSSI measurements obtained in the online phase are compared with the derived model to predict the location of the target.

3.1. Error Definition

Suppose the actual coordinates of targets are Ψ ˜ k 3 × 1 , then the positioning error Ek is defined as:
E k = Ψ k Ψ ˜ k 2
and the average positioning error EAPE is defined as:
E APE = 1 C k = 1 C E k

3.2. Noise Model of Visible Light Communication (VLC)

In indoor VLC, the noise σnoise includes shot noise σshot and thermal noise σthermal [33], which are given by:
σ noise 2 = σ shot 2 + σ thermal 2
σ shot 2 = 2 q R PD P r B + 2 q I b g I 2 B
σ thermal 2 = 8 π k B T K G 0 η A P D I 2 B 2 + 16 π 2 k B T K Γ g m η 2 A P D 2 I 3 B 3
where q is elementary charge, RPD is the responsivity of the PD, B is the equivalent noise bandwidth, Pr indicates the received power from M LEDs, kB is the Boltzmann’s constant, TK is the absolute temperature, G0 is the open loop gain, η is the fixed capacitance of PD, Ibg is the background light current, Γ is the channel noise factor, gm is the field effect transistor (FET) transconductance, I2 and I3 are the noise bandwidth factors.
According to the noise model, the signal-to-noise ratio (SNR) is given by [32]
SNR ( dB ) = 10 log 10 ( R PD P r A PD ) 2 σ noise 2

3.3. Simulation Parameters

Without loss of generality, we suppose αi = [0, 0, −1]T and γj = [0, 0, 1]T, i.e., cos(λi) = cos(ωi) = hl,m,n-i/dl,m,n-i, hl,m,n-i is the z-axis distance from the fingerprint point to the ith LED in the l dimension, the m column and the n row, which are widely adopted in papers such as [12,20,28]. The parameter setting of the Lambertian radiation model is as follows: Ts = g = 1, λ1/2 = π/3, ωFOV = π/2, APD = 1 cm2, b = 1, which follow from a typical LED setup. M LEDs are evenly distributed in a 3-D space with an area of 200 cm × 200 cm × 300 cm, min (xLED-i) = min (yLED-i) = 0 cm, max (xLED-i) = max (yLED-i) = 200 cm, and zLED = 300 cm. C = 200, i.e., 200 targets randomly appear in the 3-D or 2-D positioning area. In the KNN and WKNN algorithms, K is a fixed value, that is, K = Kmax. The parameter setting of the noise model is as follows [33]: TK = 295 K, G0 = 10, gm = 30 mS, Γ = 1.5, I2 = 0.562, I3 = 0.0868, RPD = 0.54 A/W, η = 112 pF/cm2, Ibg = 5100 µA. Unless otherwise specified, PTr = 6 W, M = 4, r = 1 (i.e., Manhattan distance). In 3-D, hL = 20 cm and hH = 100 cm. In 2-D, hL = hH = 100 cm. The simulation tool is MATLAB R2017a.
For simplicity, unless otherwise specified, we only consider the 2-D case, and S = 20 cm. In order to obtain the optimal classification accuracy of ANN, ELM, and RF algorithms, and the optimal positioning accuracy of KNN, WKNN, and SAWKNN algorithms. The optimal parameters obtained through offline training and learning are as follows: In KNN, WKNN, ARWKNN and SAWKNN algorithms, Kmax = 4. In the Section 3.4, we will also discuss the impact of different Kmax values on the average positioning error. For the optimal number of hidden nodes and trees, the classification method is the same as that in Guo et al. [11], i.e., each grid point represents a category, and the cross-validation method is adopted based on experience adjustment. For the optimal number of hidden nodes, the cross-validation method has a range of 100 to 700 and a step size of 50. For the optimal number of trees, the cross-validation method has a range of 10 to 50 and a step size of 5. After comprehensive evaluation of the positioning accuracy and classification accuracy, the optimal number of hidden nodes and trees are selected to be 600 and 40, respectively. The impact of γth on the average positioning error is shown in Figure 5, it can be seen from the Figure 5 that minimum average positioning error is achieved when γth is within the range of [30%, 50%], so, the value of γth is selected to be 40%, which denotes the threshold of two RSSI difference values that can be considered similar [19].

3.4. Result Analysis

We only consider positioning in this paper, so B = 640 KHz will be able to label 3.4 × 1038 LEDs [34], which is far exceeds the actual needs. The SNR experimental results are shown in Table 2. If B = 640 KHz, typical SNR for indoor visible light communication ranges from 42.97 to 60.92 dB, and the average value reaches 52.45 dB. In addition to indoor positioning, LEDs can also provide high-speed data rate, If B = 100 MHz, the average SNR can also reach 28.86 dB.
When Ptr = 6 W, the average positioning errors of eight algorithms are analyzed when B is within 50 MHz to 400 MHz, the results are shown in Figure 6. As the value of modulation bandwidth increases, the average positioning errors of eight algorithms increase. The higher the modulation bandwidth, the lower the SNR and the higher the average positioning errors. As only positioning is considered in this paper, a very high modulation bandwidth is not necessary. With a high-modulation bandwidth, it may be more suitable to modulate the transmission signal of the LED by modified orthogonal frequency division multiplexing (OFDM) to achieve indoor positioning [22,35,36], but this is beyond the scope of this paper. It can also be seen from Figure 6 that when B is within 50 MHz to 400 MHz, the average positioning error based on the ARWKNN algorithm is the smallest.
When B = 100 MHz, the average positioning errors of eight algorithms are analyzed when Ptr is within 1 W to 6 W, the results are shown in Figure 7. As the Ptr increases, the average positioning errors of eight algorithms decrease. When Ptr = 3 W, the average positioning errors of eight algorithms are close to convergence. The higher the transmitting power, the higher the SNR and the smaller the average positioning errors. It can also be seen from Figure 7 that when Ptr is within 1 W to 6 W, the average positioning error based on the ARWKNN algorithm is the smallest.
The average positioning errors of eight algorithms under different SNR are compared, simulation results are shown in Figure 8. As shown in Figure 8, when SNR = 10 dB, the average positioning errors of eight algorithms are large due to severe noise interference. As the SNR increases, the average positioning errors of eight algorithms decrease. When SNR = 20 dB, the average positioning errors of eight algorithms are close to convergence. Since fingerprint positioning based on RF, ELM and ANN algorithms can only determine the category of the target, compared with WKNN algorithm, the positioning error is larger. When the SNR is higher than 15, the average positioning error based on the ARWKNN algorithm is the smallest. Due to lighting requirements and LoS communication, within the typical SNR range of indoor visible light communication, the average positioning error based on the ARWKNN algorithm is significantly lower than that of RF, ELM, ANN, GI-LS, SAWKNN, WKNN and KNN algorithms. The average positioning error based on the SAWKNN algorithm is lower than that of the WKNN algorithm. The GI-LS algorithm uses the complementary advantages of KNN, RF, and ELM classifiers to weight the estimation results, the average positioning error based on the GI-LS algorithm is lower then that of KNN, RF, ELM and ANN algorithms, but higher then WKNN, ARWKNN and SAWKNN algorithms.
When SNR = 20 dB, the average positioning errors based on ARWKNN, RF, ELM, ANN, GI-LS, SAWKNN, WKNN and KNN algorithms are shown in Table 3. It can be seen from Table 3 that compared with RF, ELM, ANN, GI-LS, SAWKNN, WKNN and KNN algorithms, the average positioning error based on the ARWKNN algorithm can be reduced by 81.82%, 83.93%, 86.06%, 60.15%, 43.84%, 47.81%, and 73.36%, respectively.
When SNR = 20 dB, the simulation results of cumulative distribution function (CDF) are shown in Figure 9. It can be seen from Figure 9 that the CDF of positioning errors based on the ARWKNN algorithm is significantly better than that of the RF, ELM, ANN, GI-LS, SAWKNN, WKNN and KNN algorithms. The KNN algorithm is one of the simplest of all machine learning algorithms. Compared with the RF, ELM, ANN and GI-LS algorithms, fingerprint positioning based on the ARWKNN algorithm, not only has lower complexity, but also has lower positioning error. Fingerprint positioning is based on machine-learning algorithms, which require a large amount of data for training and learning. If there are not enough training data, the positioning error will be large, and a large amount of training data will increase the complexity of the algorithm. Compared with the SAWKNN, WKNN, and KNN algorithms, the ARWKNN algorithm can significantly reduce the average positioning error while maintaining similar algorithm complexity, which will be discussed in the section of algorithm complexity analysis.
When SNR = 20 dB, the average positioning errors of ARWKNN, SAWKNN, WKNN, and KNN algorithms are analyzed, in WKNN and KNN algorithms, K is a fixed value, that is, K = Kmax. The simulation results of 2-D and 3-D are shown in Figure 10 and Figure 11, respectively. As can be seen from Figure 10, when Kmax is within 1 to 8, similar to the experimental results in most papers, in 2-D, the optimal K based on the WKNN algorithm is 3 or 4, which exactly conforms with the fact that the target will be located in a minimum triangle or square composed of 3 or 4 fingerprint points with a high probability. It can also be analyzed from Figure 10 that when Kmax is greater than 3, the average positioning error based on the ARWKNN algorithm is significantly lower than that of the KNN, WKNN, and SAWKNN algorithms. From Figure 11, It can be seen that as the Kmax increases from 1 to 12, the average positioning error based on the ARWKNN algorithm decreases. When Kmax = 8, the average positioning error is not significantly reduced if the value of Kmax continues to increase. Therefore, a reasonable value of Kmax is taken as 8. From Figure 11, we can also see that when Kmax = 8, the average positioning error based on the KNN and WKNN algorithms is the smallest, which exactly conforms that the target will be located in a minimum cube composed of 8 fingerprint points with a high probability. It can also be analyzed from Figure 11 that when Kmax is greater than 6, the average positioning error based on the ARWKNN algorithm is significantly lower than that of the KNN, WKNN, and SAWKNN algorithms, and the advantages of the ARWKNN algorithm are more obvious as Kmax increases.
When SNR = 20 dB, the average positioning errors of ARWKNN, SAWKNN, WKNN, and KNN algorithms are analyzed with the variation of the fingerprint sampling point spacing S, the results of 2-D and 3-D are shown in Figure 12 and Figure 13, respectively. It can be seen that as S decreases from 40 cm to 20 cm, whether in 2-D or 3-D, the average positioning error based on the ARWKNN algorithm is significantly lower than that of the KNN, WKNN, and SAWKNN algorithms, and the larger the S, the more obvious the advantage. As S decreases to 5 cm, the average positioning errors of four algorithms tend to be the same. The lower the value of S, the larger the number of fingerprint points N to be acquired, and the more complicated the algorithm becomes.
When SNR = 20 dB, the average positioning errors of ARWKNN, SAWKNN, WKNN, and KNN algorithms are analyzed when M is within 3 to 8, the results of 2-D and 3-D are shown in Figure 14 and Figure 15, respectively. It can be seen that as M increases from 3 to 8, the average positioning errors based on the KNN, WKNN, SAWKNN and ARWKNN algorithms do not change much. Thus, only 4 LEDs are needed to achieve very low positioning error in this paper.
When SNR = 20 dB, in order to analyze the robustness of the algorithm, fingerprints adopt non-uniform distribution structure, i.e., the RSSI values in the fingerprint map are chosen randomly at different sampling ratios SR. The average positioning errors of the ARWKNN, SAWKNN, WKNN, and KNN algorithms are analyzed with the variation of the fingerprint sampling ratio SR, the results of 2-D and 3-D are shown in Figure 16 and Figure 17, respectively. It can be seen that as SR increases from 50% to 100%, whether in 2-D or 3-D, the average positioning error based on the ARWKNN algorithm is significantly lower than that of the KNN, WKNN, and SAWKNN algorithms, and the larger the SR, the smaller the average positioning errors of the four algorithms. When SR = 50%, the average positioning errors of the ARWKNN, SAWKNN, WKNN, and KNN algorithms are analyzed with the variation of the time, and the results of 2-D and 3-D are shown in Figure 18 and Figure 19, respectively. It can be seen that as t increases from 1 to 50, whether in 2-D or 3-D, the average positioning error based on the ARWKNN algorithm is significantly lower than that of the KNN, WKNN, and SAWKNN algorithms. As can be seen from Figure 16, Figure 17, Figure 18 and Figure 19, the ARWKNN algorithm has good robustness. When the fingerprint sampling rate is only 50%, lower positioning errors can still be achieved.
The WKNN fingerprint positioning algorithm is based on the shortest RSSI physical distance between the fingerprint and the target position. It can be seen from Step 5 of the ARWKNN algorithm that the positioning error is affected by the weight of the fingerprint point and this weight is affected by the distance metric; therefore, it is necessary to analyze the impact of different distance metrics on the positioning error. In addition to Euclidean distance (ED) and Manhattan distance (MD), there are other distance metrics [12,37], such as:
Minimum maximum distance (MMD), which is defined as:
d i s l , m , n k = 1 i = 1 M ( min ( | ϕ l , m , n i | , | Y k , i | ) ) i = 1 M ( max ( | ϕ l , m , n i | , | Y k , i | ) )
Squared Euclidean distance (SED), which is defined as:
d i s l , m , n k = i = 1 M ( ϕ l , m , n i Y k , i ) 2
Chebyshev distance (CHD), which is defined as:
d i s l , m , n k = max i | ϕ l , m , n i Y k , i |
Squared-chord distance (SCD), which is defined as:
d i s l , m , n k = i = 1 M ( | ϕ l , m , n i | | Y k , i | ) 2
Wave hedges distance (WHD), which is defined as:
d i s l , m , n k = i = 1 M ( 1 min ( | ϕ l , m , n i | , | Y k , i | ) max ( | ϕ l , m , n i | , | Y k , i | ) )
Lorentzian distance (LD), which is defined as:
d i s l , m , n k = i = 1 M ln ( 1 + | ϕ l , m , n i Y k , i | )
Matusita distance (MTD), which is defined as:
d i s l , m , n k = i = 1 M ( | ϕ l , m , n i | | Y k , i | ) 2
Squared chi-squared distance (SCSD), which is defined as:
d i s l , m , n k = i = 1 M ( ϕ l , m , n i Y k , i ) 2 | ϕ l , m , n i + Y k , i |
Canberra distance (CAD), which is defined as:
d i s l , m , n k = i = 1 M | ϕ l , m , n i Y k , i | | ϕ l , m , n i + Y k , i |
Clark distance (CLD), which is defined as:
d i s l , m , n k = i = 1 M ( | ϕ l , m , n i Y k , i | | ϕ l , m , n i + Y k , i | ) 2
For different distance metrics, if the same γth value is used, the positioning error based on the SAWKNN algorithm will be greatly affected, so this section does not consider the SAWKNN algorithm. When SNR = 20 dB, we investigated 30 distance metrics and selected 12 distance metrics with the best performances, the results of which are shown in Table 4 and Table 5. It can be seen from Table 4 and Table 5 that when the KNN algorithm is used for positioning, ED and SED metrics produce the minimum average positioning error in 2-D and 3-D. In 2-D, the average positioning error based on the WKNN algorithm is similar to the experimental results in Alam et al. [12], we also get SCD and SCSD metrics produce the minimum average positioning error, but in 3-D, SED metric produces the minimum average positioning error. When the ARWKNN algorithm is used for positioning, the CLD metric produces the minimum average positioning error in 2-D and MMD metric produces the minimum average positioning error in 3-D. As far as the authors know, this is the first work to report the impact of CLD and MMD metrics on the positioning error of the fingerprint positioning algorithm. It can also be seen from Table 4 that the best values of the KNN, WKNN and ARWKNN algorithms are 4.84 cm, 2.03 cm and 1.45 cm, respectively. Compared with the KNN and WKNN algorithms, in 2-D, the minimum average positioning error of the ARWKNN algorithm can be reduced by 70.04%, and 28.57%, respectively. It can also be seen from Table 5 that the best values of the KNN, WKNN and ARWKNN algorithms are 4.46 cm, 3.05 cm and 2.18 cm, respectively. Compared with the KNN and WKNN algorithms, in 3-D, the minimum average positioning error of the ARWKNN algorithm can be reduced by 51.12%, and 28.52%, respectively. In 2-D or 3-D, the average positioning errors of the ARWKNN algorithm proposed in this paper are all smaller than that of the KNN and WKNN algorithms under 12 distance metrics.
Figure 20 shows the cumulative distributions of positioning errors for the ED and CLD metrics with various S values. As can be seen from Figure 20, in 2-D, compared with the ED metric, the CLD metric produces smaller positioning error. In addition, compared with the CLD metric, the positioning error of the ED metric increases faster when S becomes larger. Figure 21 shows the cumulative distributions of positioning errors for the ED and MMD metrics with various S values. As can be seen from Figure 21, in 3-D, compared with the ED metric, the MMD metric produces smaller positioning error. In addition, compared with the MMD metric, the positioning error of ED metric increases faster when S becomes larger. ED is a commonly used distance metric, however, as can be seen from Table 4 and Table 5, in fact, the ED is not the most accurate metric for calculating weights when the WKNN and ARWKNN algorithms are used for positioning.
SNR = 20 dB. To make the graph have a certain degree of discrimination, we only choose the ED, MMD, SED, SCD, and CLD metrics to analyze the cumulative distributions of optimal K. The results of 2-D and 3-D are shown in Figure 22 and Figure 23, respectively. As can be seen from Figure 22 and Figure 23, there are differences in the optimal K values for 200 targets, and there are also differences in cumulative distributions of the optimal K for five distance metrics. The optimal K cumulative distributions for ED, MMD and CLD are very close, and the optimal K cumulative distributions for SED and SCD are also very close.
The complexity of the KNN and WKNN algorithms mainly depends on the size of N and the sorting operation of Step 2 in Algorithm 1. Compared with the KNN and WKNN algorithms, the ARWKNN algorithm also performs Step 3 loop function and Step 4 min function in Algorithm 1. The time complexity of Step 3 plus Step 4 depends on the size of Kmax. Since Kmax is much smaller than N, that is, the number of neighboring fingerprint points are much smaller than the total number of fingerprint points, the complexity of the ARWKNN algorithm is similar to the KNN and WKNN algorithms. In 3-D, when Kmax = 8, the average computing time of 200 targets varying with S is analyzed, the result of which is shown in Table 6. It can be seen that when S is the same, the average calculation time of the KNN, WKNN, SAWKNN, and ARWKNN algorithms is almost the same. It can also be seen from Figure 13 that when S decreases, the average positioning errors of four algorithms decrease, but the complexity of the algorithm also increases. Therefore, according to the actual situation, the power consumption and positioning error of the algorithm can be compromised by selecting an appropriate S.

4. Conclusions

At present, the classical KNN and WKNN algorithms are mainly aimed at 2-D positioning, assuming that the height of the target from the floor is known, and it is not feasible to know the height of the target from the floor in advance. The least linear multiplication method and Newton–Raphson method are suitable for solving 2-D coordinates. Solving the 3-D coordinate is a non-convex optimization problem, which is easy to fall into a local optimal solution. In this paper, the shortcomings of the fingerprint positioning algorithm and the trilateration method are discussed, and an adaptive residual weighted K-nearest neighbor fingerprint positioning algorithm is proposed. Compared with the fingerprint positioning algorithm based on compressed sensing, the range-based WKNN algorithm can achieve high-precision positioning under the low-density LED layout. Compared with RF [14], ELM [16], ANN [17], and GI-LS [11] machine-learning algorithms, fingerprint positioning based on the ARWKNN algorithm not only has lower complexity, but also has lower positioning error. The impact of LEDs modulation bandwidth, LEDs transmit power, the signal-to-noise ratio, the maximum number of neighboring fingerprints, the sampling interval, the number of LEDs, the sampling ratio and distance metric on positioning errors are analyzed in detail. The distribution of optimal K and the complexity of the algorithm are also analyzed in detail. Simulation results show that the ARWKNN algorithm based on CLD and MMD metrics produces the smallest average positioning error in 2-D and 3-D, respectively. Compared with the SAWKNN [19], WKNN [12] and KNN [15,25] algorithms, the ARWKNN algorithm can significantly reduce the average positioning error while maintaining similar algorithm complexity.
Due to lighting requirements and LoS communication, the typical SNR of indoor visible light communication is relatively high, however, the RF, ELM, ANN, GI-LS, SAWKNN, WKNN, KNN, and ARWKNN algorithms have higher positioning error under low SNR conditions. Our next step is to design an efficient noise filtering algorithm to achieve higher positioning accuracy under low SNR conditions. LED communication can not only achieve high-precision positioning, but also achieve high rates. We will consider using modified OFDM to achieve high-precision positioning with high modulation bandwidth and provide a real scenario in the future.

Author Contributions

Conceptualization, S.X. and Y.W.; methodology, S.X. and Y.W.; software, S.X., X.W., and F.W.; validation, S.X., C.-C.C., and Y.W.; formal analysis, S.X. and Y.W.; investigation, S.X., C.-C.C., and Y.W.; resources, Y.W.; data curation, Y.W.; writing—original draft preparation, S.X.; writing—review and editing, Y.W., X.W., and F.W.; visualization, S.X., C.-C.C., Y.W., X.W., and F.W.; supervision, Y.W.; project administration, Y.W.; funding acquisition, Y.W. and X.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China, grant number U1805262, 61871131, 61701118 and 61901117, this research also was funded by Research Project of Fujian Provincial, grant number 2018H6007 and 2019J01267, this research also was funded by Special Fund for Marine Economic Development of Fujian Province, grant number ZHHY-2020-3, and this research also was funded by Fujian Provincial Engineering Technology Research Center of Photoelectric Sensing Application, grant number 2018003.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Jang, B.; Kim, H. Indoor positioning technologies without offline fingerprinting map: A survey. IEEE Commun. Surv. Tutor. 2019, 21, 508–525. [Google Scholar] [CrossRef]
  2. Yao, C.Y.; Hsia, W.C. An indoor positioning system based on the dual-channel passive RFID technology. IEEE Sens. J. 2018, 18, 4654–4663. [Google Scholar] [CrossRef]
  3. Gharghan, S.K.; Nordin, R.; Jawad, A.M.; Jawad, H.M.; Ismail, M. Adaptive neural fuzzy inference system for accurate localization of wireless sensor network in outdoor and indoor cycling applications. IEEE Access 2018, 6, 38475–38489. [Google Scholar] [CrossRef]
  4. Abdulrahman, A.; Abdulmalik, A.S.; Mansour, A.; Ahmad, A.; Suheer, A.H.; Mai, A.A.A.; Hend, S.A.K. Ultra wideband indoor positioning technologies: Analysis and recent advances. Sensors 2016, 16, 707. [Google Scholar]
  5. Zou, H.; Jin, M.; Jiang, H.; Xie, L.; Spanos, C.J. WinIPS: WiFi-based non-intrusive indoor positioning system with online radio map construction and adaptation. IEEE Trans. Wirel. Commun. 2017, 16, 8118–8130. [Google Scholar] [CrossRef]
  6. Yadav, R.K.; Bhattarai, B.; Gang, H.S.; Pyun, J.Y. Trusted K nearest bayesian estimation for indoor positioning system. IEEE Access 2019, 7, 51484–51498. [Google Scholar] [CrossRef]
  7. Trong-Hop, D.; Myungsik, Y. An in-depth survey of visible light communication based positioning systems. Sensors 2016, 16, 678. [Google Scholar]
  8. Yuan, P.; Zhang, T.; Yang, N.; Xu, H.; Zhang, Q. Energy efficient network localisation using hybrid TOA/AOA measurements. IET Commun. 2019, 13, 963–971. [Google Scholar] [CrossRef]
  9. Wang, Y.; Ho, K.C. Unified near-field and far-field localization for AOA and hybrid AOA-TDOA positionings. IEEE Trans. Wirel. Commun. 2018, 17, 1242–1254. [Google Scholar] [CrossRef]
  10. Khalajmehrabadi, A.; Gatsis, N.; Akopian, D. Modern WLAN fingerprinting indoor positioning methods and deployment challenges. IEEE Commun. Surv. Tutor. 2017, 19, 1974–2002. [Google Scholar] [CrossRef] [Green Version]
  11. Guo, X.; Shao, S.; Ansari, N.; Khreishah, A. Indoor localization using visible light via fusion of multiple classifiers. IEEE Photon. J. 2017, 9, 7803716. [Google Scholar] [CrossRef]
  12. Alam, F.; Chew, M.T.; Wenge, T.; Gupta, G.S. An accurate visible light positioning system using regenerated fingerprint database based on calibrated propagation model. IEEE Trans. Instrum. Meas. 2019, 68, 2714–2723. [Google Scholar] [CrossRef]
  13. Lovón-Melgarejo, J.; Castillo-Cara, M.; Huarcaya-Canal, O.; Orozco-Barbosa, L.; García-Varea, I. Comparative study of supervised learning and metaheuristic algorithms for the development of bluetooth-based indoor localization mechanisms. IEEE Access 2019, 7, 26123–26135. [Google Scholar] [CrossRef]
  14. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef] [Green Version]
  15. Xue, W.; Yu, K.; Hua, X.; Li, Q.; Qiu, W.; Zhou, B. APs’ virtual positions-based reference point clustering and physical distance-based weighting for indoor Wi-Fi positioning. IEEE Internet Things J. 2018, 5, 3031–3042. [Google Scholar] [CrossRef]
  16. Huang, G.B.; Zhou, H.; Ding, X.; Zhang, R. Extreme learning machine for regression and multiclass classification. IEEE Trans. Syst. Man Cybern. B. 2012, 42, 513–529. [Google Scholar] [CrossRef] [Green Version]
  17. Zhang, S.; Du, P.; Chen, C.; Zhong, W.D.; Alphones, A. Robust 3D indoor VLP system based on ANN using hybrid RSS/PDOA. IEEE Access 2019, 7, 47769–47780. [Google Scholar] [CrossRef]
  18. Fang, X.; Jiang, Z.; Nan, L.; Chen, L. Optimal weighted K-nearest neighbour algorithm for wireless sensor network fingerprint localisation in noisy environment. IET Commun. 2018, 12, 1171–1177. [Google Scholar] [CrossRef]
  19. Hu, J.; Liu, D.; Yan, Z.; Liu, H. Experimental analysis on weight K-nearest neighbor indoor fingerprint positioning. IEEE Internet Things J. 2019, 6, 891–897. [Google Scholar] [CrossRef]
  20. Gu, W.; Aminikashani, M.; Deng, P.; Kavehrad, M. Impact of multipath reflections on the performance of indoor visible light positioning systems. IEEE/OSA J. Lightwave Technol. 2016, 34, 2578–2587. [Google Scholar] [CrossRef] [Green Version]
  21. Şahin, A.; Eroğlu, Y.S.; Güvenç, I.; Pala, N.; Yüksel, M. Hybrid 3D localization for visible light communication systems. IEEE/OSA J. Lightwave Technol. 2015, 33, 4589–4599. [Google Scholar] [CrossRef]
  22. Mathias, L.C.; Melo, L.F.D.; Abrao, T. 3-D localization with multiple LEDs lamps in OFDM-VLC system. IEEE Access 2018, 7, 6249–6261. [Google Scholar] [CrossRef]
  23. Zhou, B.; Lau, V.; Chen, Q.; Cao, Y. Simultaneous positioning and orientating for visible light communications: Algorithm design and performance analysis. IEEE Trans. Veh. Technol. 2018, 67, 11790–11804. [Google Scholar] [CrossRef] [Green Version]
  24. Wu, Y.; Liu, X.; Guan, W.; Chen, B.; Chen, X.; Xie, C. High-speed 3D indoor localization system based on visible light communication using differential evolution algorithm. Opt. Commun. 2018, 424, 177–189. [Google Scholar] [CrossRef]
  25. Van, M.T.; Tuan, N.V.; Son, T.T.; Le-Minh, H.; Burton, A. Weighted k-nearest neighbour model for indoor VLC positioning. IET Commun. 2017, 11, 864–871. [Google Scholar] [CrossRef]
  26. Gligorić, K.; Ajmani, M.; Vukobratović, D.; Sinanović, S. Visible light communications-based indoor positioning via compressed sensing. IEEE Commun. Lett. 2018, 22, 1410–1413. [Google Scholar] [CrossRef] [Green Version]
  27. Tropp, J.A.; Gilbert, A.C. Signal recovery from random measurements via orthogonal matching pursuit. IEEE Trans. Inf. Theory 2007, 53, 4655–4666. [Google Scholar] [CrossRef] [Green Version]
  28. Zhang, R.; Zhong, W.D.; Qian, K.; Zhang, S.; Du, P. A reversed visible light multitarget localization system via sparse matrix reconstruction. IEEE Internet Things J. 2018, 5, 4223–4230. [Google Scholar] [CrossRef]
  29. Feng, C.; Au, W.S.A.; Valaee, S.; Tan, Z. Received-signal-strength-based indoor positioning using compressive sensing. IEEE Trans. Mob. Comput. 2012, 11, 1983–1993. [Google Scholar] [CrossRef]
  30. Keskin, M.F.; Sezer, A.D.; Gezici, S. Localization via visible light systems. Proc. IEEE 2018, 106, 1063–1088. [Google Scholar] [CrossRef]
  31. Zhou, Z.; Kavehrad, M.; Deng, P. Indoor positioning algorithm using light-emitting diode visible light communications. Opt. Eng. 2012, 51, 085009. [Google Scholar] [CrossRef] [Green Version]
  32. Yasir, M.; Ho, S.W.; Vellambi, B.N. Indoor positioning system using visible light and accelerometer. IEEE/OSA J. Lightwave Technol. 2014, 32, 3306–3316. [Google Scholar] [CrossRef]
  33. Komine, T.; Nakagawa, M. Fundamental analysis for visible-light communication system using LED lights. IEEE Trans. Consum. Electron. 2004, 50, 100–107. [Google Scholar] [CrossRef]
  34. Zhang, W.; Chowdhury, M.I.S.; Kavehrad, M. Asynchronous indoor positioning system based on visible light communications. Opt. Eng. 2014, 53, 045105. [Google Scholar] [CrossRef]
  35. Li, H.; Wang, J.; Zhang, X.; Wu, R. Indoor visible light positioning combined with ellipse-based ACO-OFDM. IET Commun. 2018, 12, 2181–2187. [Google Scholar] [CrossRef]
  36. Mossaad, M.S.A.; Hranilovic, S.; Lampe, L. Visible light communications using OFDM and multiple LEDs. IEEE Trans. Commun. 2015, 63, 4304–4313. [Google Scholar] [CrossRef]
  37. Cha, S.H. Comprehensive survey on distance/similarity measures between probability density functions. Int. J. Math. Models Methods Appl. Sci. 2007, 1, 300–307. [Google Scholar]
Figure 1. Fingerprint positioning based on visible light communication (VLC): (a) The positioning model; (b) The collection directions of fingeprints in x-axis, y-axis and z-axis.
Figure 1. Fingerprint positioning based on visible light communication (VLC): (a) The positioning model; (b) The collection directions of fingeprints in x-axis, y-axis and z-axis.
Sensors 20 04432 g001
Figure 2. Frame structure of the positioning system for one period.
Figure 2. Frame structure of the positioning system for one period.
Sensors 20 04432 g002
Figure 3. The relationship between target and fingerprints: (a) K = 1; (b) K = 2; (c) K = 3; (d) K = 8.
Figure 3. The relationship between target and fingerprints: (a) K = 1; (b) K = 2; (c) K = 3; (d) K = 8.
Sensors 20 04432 g003
Figure 4. Positioning errors of 5 targets using weighted K-nearest neighbor (WKNN) algorithm.
Figure 4. Positioning errors of 5 targets using weighted K-nearest neighbor (WKNN) algorithm.
Sensors 20 04432 g004
Figure 5. The impact of γth on the average positioning error.
Figure 5. The impact of γth on the average positioning error.
Sensors 20 04432 g005
Figure 6. The impact of B on the average positioning error with Kmax = 4.
Figure 6. The impact of B on the average positioning error with Kmax = 4.
Sensors 20 04432 g006
Figure 7. The impact of Ptr on the average positioning error with Kmax = 4.
Figure 7. The impact of Ptr on the average positioning error with Kmax = 4.
Sensors 20 04432 g007
Figure 8. The impact of SNR on the average positioning error with Kmax = 4.
Figure 8. The impact of SNR on the average positioning error with Kmax = 4.
Sensors 20 04432 g008
Figure 9. The cumulative distributions of positioning errors with Kmax = 4.
Figure 9. The cumulative distributions of positioning errors with Kmax = 4.
Sensors 20 04432 g009
Figure 10. In 2-D, the impact of Kmax on the average positioning error.
Figure 10. In 2-D, the impact of Kmax on the average positioning error.
Sensors 20 04432 g010
Figure 11. In 3-D, the impact of Kmax on the average positioning error.
Figure 11. In 3-D, the impact of Kmax on the average positioning error.
Sensors 20 04432 g011
Figure 12. In 2-D, the impact of S on the average positioning error with Kmax = 4.
Figure 12. In 2-D, the impact of S on the average positioning error with Kmax = 4.
Sensors 20 04432 g012
Figure 13. In 3-D, the impact of S on the average positioning error with Kmax = 8.
Figure 13. In 3-D, the impact of S on the average positioning error with Kmax = 8.
Sensors 20 04432 g013
Figure 14. In 2-D, the impact of M on the average positioning error with Kmax = 4.
Figure 14. In 2-D, the impact of M on the average positioning error with Kmax = 4.
Sensors 20 04432 g014
Figure 15. In 3-D, the impact of M on the average positioning error with Kmax = 8.
Figure 15. In 3-D, the impact of M on the average positioning error with Kmax = 8.
Sensors 20 04432 g015
Figure 16. In 2-D, the impact of sampling ratio (SR) on the average positioning error with Kmax = 4.
Figure 16. In 2-D, the impact of sampling ratio (SR) on the average positioning error with Kmax = 4.
Sensors 20 04432 g016
Figure 17. In 3-D, the impact of SR on the average positioning error with Kmax = 8.
Figure 17. In 3-D, the impact of SR on the average positioning error with Kmax = 8.
Sensors 20 04432 g017
Figure 18. In 2-D, the impact of t on the average positioning error with Kmax = 4.
Figure 18. In 2-D, the impact of t on the average positioning error with Kmax = 4.
Sensors 20 04432 g018
Figure 19. In 3-D, the impact of t on the average positioning error with Kmax = 8.
Figure 19. In 3-D, the impact of t on the average positioning error with Kmax = 8.
Sensors 20 04432 g019
Figure 20. In 2-D, the cumulative distributions of positioning errors for the ED and CLD metrics with various S values. Kmax = 4.
Figure 20. In 2-D, the cumulative distributions of positioning errors for the ED and CLD metrics with various S values. Kmax = 4.
Sensors 20 04432 g020
Figure 21. In 3-D, the cumulative distributions of positioning errors for the ED and CLD metrics with various S values. Kmax = 8.
Figure 21. In 3-D, the cumulative distributions of positioning errors for the ED and CLD metrics with various S values. Kmax = 8.
Sensors 20 04432 g021
Figure 22. In 2-D, the cumulative distributions of the optimal K with Kmax = 4.
Figure 22. In 2-D, the cumulative distributions of the optimal K with Kmax = 4.
Sensors 20 04432 g022
Figure 23. In 3-D, the cumulative distributions of the optimal K with Kmax = 8.
Figure 23. In 3-D, the cumulative distributions of the optimal K with Kmax = 8.
Sensors 20 04432 g023
Table 1. The meaning of the indices.
Table 1. The meaning of the indices.
IndicesMeaning
mThe number of columns corresponding to the fingerprint point
nThe number of rows corresponding to the fingerprint point
lThe number of dimensions corresponding to the fingerprint point
Table 2. Typical signal-to-noise ratio (SNR) in indoor environment.
Table 2. Typical signal-to-noise ratio (SNR) in indoor environment.
MinimumMaximumAverage
SNR (B = 640 KHz)42.97 dB60.92 dB52.45 dB
SNR (B = 100 MHz)19.72 dB37.35 dB28.86 dB
Table 3. Average positioning error for each algorithm with SNR = 20 dB.
Table 3. Average positioning error for each algorithm with SNR = 20 dB.
AlgorithmAverage Positioning Error
ARWKNN1.55 cm
RF8.53 cm
ELM9.65 cm
ANN11.12 cm
GI-LS3.89 cm
SAWKNN2.76 cm
WKNN2.97 cm
KNN5.82 cm
Table 4. In 2-D, the impact of distance metrics on the average positioning error with Kmax = 4, best values for KNN, WKNN, and ARWKNN algorithms are highlighted in bold.
Table 4. In 2-D, the impact of distance metrics on the average positioning error with Kmax = 4, best values for KNN, WKNN, and ARWKNN algorithms are highlighted in bold.
Distance MetricsKNNWKNNARWKNN
ED4.84 cm2.61 cm1.51 cm
MD5.82 cm2.97 cm1.55 cm
MMD5.84 cm2.97 cm1.54 cm
SED4.84 cm2.16 cm1.85 cm
CHD5.22 cm2.97 cm1.61 cm
SCD4.99 cm2.03 cm1.82 cm
WHD5.79 cm2.94 cm1.54 cm
LD6.17 cm3.45 cm1.81 cm
MTD4.99 cm2.61 cm1.50 cm
SCSD4.99 cm2.03 cm1.82 cm
CAD5.97 cm2.93 cm1.53 cm
CLD5.05 cm2.62 cm1.45 cm
Table 5. In 3-D, the impact of distance metrics on the average positioning error with Kmax = 8, best values for KNN, WKNN, and ARWKNN algorithms are highlighted in bold.
Table 5. In 3-D, the impact of distance metrics on the average positioning error with Kmax = 8, best values for KNN, WKNN, and ARWKNN algorithms are highlighted in bold.
Distance MetricsKNNWKNNARWKNN
ED4.46 cm3.30 cm2.31 cm
MD4.63 cm3.29 cm2.28 cm
MMD4.69 cm3.33 cm2.18 cm
SED4.46 cm3.05 cm2.42 cm
CHD5.18 cm4.12 cm2.86 cm
SCD4.53 cm3.13 cm2.58 cm
WHD4.91 cm3.53 cm2.43 cm
LD5.30 cm4.16 cm2.61 cm
MTD4.53 cm3.41 cm2.45 cm
SCSD4.53 cm3.13 cm2.58 cm
CAD4.88 cm3.51 cm2.41 cm
CLD4.64 cm3.58 cm2.60 cm
Table 6. Computational complexity analysis.
Table 6. Computational complexity analysis.
AlgorithmThe Value of SAverage Computing Time
KNNS = 10 cm15.07 ms
WKNNS = 10 cm15.18 ms
SAWKNNS = 10 cm15.51 ms
ARWKNNS = 10 cm15.28 ms
KNNS = 20 cm8.62 ms
WKNNS = 20 cm8.68 ms
SAWKNNS = 20 cm8.95 ms
ARWKNNS = 20 cm8.91 ms

Share and Cite

MDPI and ACS Style

Xu, S.; Chen, C.-C.; Wu, Y.; Wang, X.; Wei, F. Adaptive Residual Weighted K-Nearest Neighbor Fingerprint Positioning Algorithm Based on Visible Light Communication. Sensors 2020, 20, 4432. https://doi.org/10.3390/s20164432

AMA Style

Xu S, Chen C-C, Wu Y, Wang X, Wei F. Adaptive Residual Weighted K-Nearest Neighbor Fingerprint Positioning Algorithm Based on Visible Light Communication. Sensors. 2020; 20(16):4432. https://doi.org/10.3390/s20164432

Chicago/Turabian Style

Xu, Shiwu, Chih-Cheng Chen, Yi Wu, Xufang Wang, and Fen Wei. 2020. "Adaptive Residual Weighted K-Nearest Neighbor Fingerprint Positioning Algorithm Based on Visible Light Communication" Sensors 20, no. 16: 4432. https://doi.org/10.3390/s20164432

APA Style

Xu, S., Chen, C. -C., Wu, Y., Wang, X., & Wei, F. (2020). Adaptive Residual Weighted K-Nearest Neighbor Fingerprint Positioning Algorithm Based on Visible Light Communication. Sensors, 20(16), 4432. https://doi.org/10.3390/s20164432

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop