1. Introduction
In this work we address the problem of positioning and tracking using beacons that transmit data with Bluetooth low-energy (BLE) messages. In the most general setup, we distribute beacons whose exact positions are not necessarily known in a closed indoor area, and we need to estimate and track the position of a mobile sensor using BLE messages emitted by the beacons. As the underlying technology uses short wavelength radio signals, data transfer is affected by a high number of factors, such as scattering and reflections, all of which can not be exactly known. We attempt to model and solve this issue using probabilistic methods.
BLE is a distinctive feature that was added to the Bluetooth technology standard at v4.0. It has been introduced to facilitate short-range communication, which requires large amounts of data transfer. However, this new feature in Bluetooth technology made it a good candidate as part of positioning systems as it provides an efficient way of transferring messages with low energy, with the help battery-powered mobile beacons to emit messages for durations in the order of years [
1]. Moreover, BLE technology provides the variables that can be used to predict the position information. One of those variables is the received signal strength indicator (RSSI), which may give an estimate of the flight distance of the signal from the beacon to the receiver [
2]. Neburka et al. [
3], after extensively analyzing the performance of BLE technology in indoor environments, show that it is a promising technique for indoor positioning, even though the RSSI values are inaccurate and highly depend on the BLE module used. Its new abilities, like durability, mobility, and high reaction time have led to the Bluetooth BLE technology replacing Wi-Fi for positioning purposes.
An accurate, lightweight, and easily deployable tracking system for indoor environments has many practical applications. For example, in marketing and retail, a reliable indoor positioning and tracking system (IPS) deployed in a closed area such as a mall or a shopping center, is used for targeted advertising, campaign management and customer behavior monitoring [
4,
5,
6]. In healthcare, elderly people or patients with dementia can be tracked in indoor environments without needing any other caretaking personnel or complex sensors [
7]. In a factory or production environment, assets or hardware can be continuously logged for security reasons or inventory analysis [
8]. In mobile robotics, localization is the basis for trajectory planning and smooth and secure navigation. Researchers attempt to solve this problem with dead-reckoning using inertial sensors accompanied by visual/range-related environmental sensors like cameras, range scanners, and sonars, or use a combination of these methods [
9,
10]. BLE fits into this domain as another practical, simple, and economical sensory system.
There are mainly two basic techniques used in IPS:
trilateration and
fingerprinting: In
trilateration, estimated distances are used to calculate the most probable coordinates using the geometry of triangles [
11]. With BLE, the distances to the beacons are estimated via the RSSI readings which are expected to map to the actual distances according to the
radio frequency (RF) propagation model [
12]. Chen et al. [
13,
14] combine the pedesterian dead reckoning (PDR) with a weighted path loss (WPL) algorithm that bases on the log-distance path loss model between a router and a client, under an extended Kalman filter. While the technique is based on a fast training phase, problems arise due to the RSSI-to-distance mapping model, which is regarded unstable due to the unreliable distance estimation in BLE [
15,
16]. On the other hand,
fingerprinting relies on a prior scene analysis (or radio map construction [
17]), in which measurements are collected from the environment and calibrated into specific features, called fingerprints, which describe a part of the environment [
18]. The system estimates the location of a sensor by comparing the online measurements with the fingerprints, and the most similar fingerprint location is regarded as an estimate for the expected position. To make the technique robust and precise, the fingerprint locations must be as dense as possible. Thus, there arises a trade-off between the installation overhead versus the precision. Besides, the overhead may also include the recalibration after the system is installed [
19,
20]. Previous work on IPS shows that fingerprinting-based techniques outperform the trilateration technique with respect to reliability and precision [
16,
21,
22].
Different algorithms are employed over the above mentioned IPS techniques. The authors of [
23] study three typical algorithms in IPS domain: the k-nearest neighbors (kNN), neural networks and support vector machines (SVM). They show that kNN methods yield better results for localization than the other two. The authors of [
24] use fingerprint technique and evaluate BLE localization performances based on extensive indoor measurements, mapped by propagation modes. The authors of [
25] combine the propagation model (PM) with the extended Kalman filter (EKF), and reach an error rate of 2.56 m, improving the localization accuracy with sparse beacon deployment. The authors of [
26] compare the fingerprinting technique via Wi-Fi with the current fingerprinting via BLE. They use a Bayesian estimator as the main method for positioning with a grid size of 1 m, and achieve a tracking accuracy of 2.6 m using a dense beacon distribution and 4.8 m using a sparse distribution, which, as they claim, is a significant improvement over Wi-Fi even using sparse beacon deployment [
27]. As another technique, Chen et al. [
28] study the Bayesian fusion methods using Bluetooth fingerprints and achieve a horizontal positioning accuracy of 4.7 m on average with respect to Bayesian static estimation and Kalman filter.
Related literature guides us to use the fingerprinting technique. Nevertheless, we attempt to see if the trilateration technique may be applied with the beacon set in hand. As previous research [
16] suggests, we ask if we can fit a logarithm-based attenuation model to the RSSI data so that we can map the data to a distance value easily, but without even attempting to perform a fitting, we can see the inconsistency in the histograms changing with distance.
Figure 1 shows the points that move away from two different beacons and their corresponding histograms of the RSSI data on each position. While the expected behavior of the modes would be to move from lower values to higher values as the point gets farther away from the beacon, we cannot visualize this behavior consistently. The direction of the blue arrow is selected to be aligned with almost linearly positioned fingerprint locations, so that we expect a consistent attenuation of strength values of the corresponding beacon data. The histograms of the blue beacon on
Figure 1 do not display the expected behavior. Instead, we see more than one peaked histograms as we get farther away. The same inconsistency can be visualized with the similar setting for the red beacon.
As we cannot map the data to some distance metric, the trilateration method does not seem to be an option in our case. Even when the beacon is in the line-of-sight (LoS) of the sensor, the data values show interesting fluctuations. Moreover, a single reading, which may be a reflection, would lead to incorrect distance mappings, but with fingerprinting the same reading can be explained by another mode of a histogram.
Accordingly in this work, we rely on the fingerprinting technique and we compute the histograms of RSSI readings on certain positions as the fingerprints (see
Figure 2a), but we cannot use the conventional tracking algorithms like the Kalman filter or its variations as the fingerprint histograms do not display any Gaussian-like structure, possibly because of the signal reflections from various objects and surfaces (see
Figure 2b). Moreover, we could have discarded the data of the lower strengths by using sliding windows to arrive at Gaussian-like histograms, but we have the tendency to keep the multimodality in the histograms, believing that the multimodality has the location information within.
However, for fine localization, we require many fingerprints, the collection of which costs time and thus is impractical. Vector interpolation techniques are considered to estimate the histograms associated with arbitrary map positions. Such a technique is applied in the works of [
29,
30] with Wi-Fi strength indicators. They approximate the likelihood of an observation via Nadaraya–Watson kernel regression, which is a computationally complex model, but they achieve accuracies under one meter with Wi-Fi.
We introduce a histogram interpolation technique, affine Wasserstein histogram interpolation, into the IPS domain to approximate a radio map using lower number of fingerprints, which is obviously less time consuming and more practical. With an intelligent interpolation, sparsely distributed fingerprints are used to estimate the densities at any location. Our radio map model differs from the state-of-the-art by its tendency to define the interpolated histograms by a combination of the surrounding histograms, which makes the method totally data driven, and thus the output is expected to keep the original modes of the histograms (because of scattering and reflections).
The Wasserstein distance is originally a distance metric in optimal transport, defining the distance between two distributions [
31]. While optimally transporting measures between distributions, a transport map is calculated and the measures are gradually moved according to this map [
32,
33], which corresponds to an interpolation between two distributions. The authors of [
34] use a similar regression intuition in computer graphics under the name `Wasserstein Barycenter coordinates’. One of their applications is recolorizing the images with a transition between the original and the objective color histograms. Likewise, the authors of [
35] introduce a class of algorithms for tractable optimization problems to solve optimal transportation over geometric domains and apply the idea for shape interpolation, Bidirectional Reflectance Distribution Function (BRDF) design, color histogram manipulation, skeleton layout, and soft maps.
With a radio map approximation model in hand, we require a tracking method to estimate the positions. For this purpose, the authors of [
36] employ the article filter for indoor localization fusing Wi-Fi- received signal strengths with accelerometer, gyroscope, map, and barometer data. They show that map information provides a significant improvement on the positioning accuracy. The authors of [
37] use the particle filter over the iBeacon data. They achieve an error rate as low as 0.27 m using the trilateration technique.
For the localization purposes, the SMC method can be employed if we can model the motion and have a location related evaluation [
38]. As we have both, we build an SMC method to tackle the problem of indoor positioning and tracking. However, the main contribution of this work is the application of Wasserstein interpolation to estimate the observation densities that map the positions to the beacon RSSI densities. As side contributions, we also add the other regression methods, like the k-nearest neighbor convex combination and neural networks to estimate the observation densities on arbitrary map coordinates. The estimated observation densities are then used to in the update step of the SMC algorithm.
In
Section 2, we give the details on the tracking algorithms and the accessories used in these algorithms.
Section 3 describes the test setup and the experiments. We give the experimental results in
Section 4 and finalize our report with a summary and a plan of the future of the research in
Section 5.
2. Methodology
We are dealing with an indoor positioning and tracking problem. This section shows the model of this problem, a hidden Markov model (HMM), followed by a simple transition density model. We focus primarily on the estimation of the observation densities on arbitrary positions, so that the likelihood of a single position can be computed. While we contribute to the domain with the application of Wasserstein interpolation to estimate observation density, many other regression or interpolation methods are also discussed. With transition and observation densities in hand, we show how to construct an SMC filter for BLE localization and tracking at the end of the section.
In the most general case, bold uppercase letters () denote matrices, and bold lowercase letters () are the vectors. Concordantly, will always symbolize the histogram vectors. An element of a vector will be denoted with the unbold version of the original vector and with an index in the subscript. Index numbering is classically in lower case. Particularly, t always denotes the time points. The superscript on the histogram () symbolizes the position it is generated on. If the histogram is estimated by a specific method, it will be marked with a tilde (∼), the method will be written in the subscript, and possible hyperparameters will be written as parameters to the method function (). The positions are of two types: P, the arbitrary ones, and F, the fingerprint positions. Fingerprint positions will be indexed in the subscript as they belong to a finite set. denote the conditional probability distributions, and are the problem specific hyperparameters that, in this work, depend on the estimation algorithms.
2.1. Tracking Problem
We form a hidden Markov model (HMM) for the tracking problem. Let
be a subset of
and
be a position in this set
, forming the latent variables. We observe the RSSI values belonging to a discrete set of
, the size of which is adjusted according to the maximal values of the captured RSSI values. We denote each data instance
, with
t being the time stamp of such an instance and
b being the index of the beacon that transmits data. According to our model
depends on the position
and the beacon
. The generative model follows and the graphical model is given in
Figure 3:
For the transition density,
, we use the motion model given in
Section 2.2. In
Section 2.3, we construct the estimators for the observation density via multiple methods that use the fingerprints. With the estimated observation densities we compute the likelihood of a new observation sample,
. The two models are then combined in a tracking filter described in
Section 2.4.
2.2. Transition Model (Diffusion Motion Model)
We restrict our agent to reside on a plane, in which the position is composed of two Cartesian coordinates,
(see
Figure 4).
We assume that the robot is not fed by any internal sensory data that enables dead reckoning. The diffusion motion model relies on the assumption that the robot will be in a close position to a previous one, which is thus modeled as a Gaussian distribution with the previous position as the mean value. The covariance of the distribution determines how far the robot can move in a time unit [
10].
The next state,
, can be modeled by:
where
. We assume that distributions on individual dimensions are independent and their variances are equal (
R).
2.3. Observation Models
In this section, we list our contributions for the observation models to find an estimate on the likelihood of a measurement given a position of the map.
2.3.1. Nearest Fingerprint (NF)
The most naïve method to estimate the likelihood density on a position
P is to use directly the information on the nearest fingerprint position, corresponding to the
k-nearest neighbor method with
. For a given position,
P, we find the nearest fingerprint position to
P with respect to the Euclidean distance. With
and
:
We use the corresponding histogram on the closest fingerprint location,
, as the likelihood:
This method assumes that the area in which the position estimation will be handled is densely sampled, so that for every position P, there happens to be a close fingerprint position.
2.3.2. k-Nearest Fingerprint Combination (kNF)
Alternatively, we can take the convex combination of the surrounding fingerprints. We compute a linear combination of the histograms with respect to their distances to the estimation position which reside in the convex hull of the input histograms (see
Figure 5). Furthermore, we generalize the convex combination onto the plane spanned by the fingerprint positions, relaxing the convex hull to an affine hull, making it an affine vector combination.
Let
denote the fingerprint positions on which the histograms are previously computed,
, and let the position
P reside in the affine hull of a subset of fingerprint positions. Then the histogram on
P,
, is defined as:
where
are multipliers that are tuned with respect to the distances between the estimation position
P and the fingerprint positions
, and necessarily
. In particular, we use the softmax function to boost the effect of the close fingerprint positions:
. The observation density of a specific position on the map can then be estimated as
where
k is a parameter that decides how many fingerprint positions closest to the estimation position will be used. Note that the method
NF is the special case of
kNF with
.
2.3.3. The Artificial Neural Network (ANN)
Considering the problem as a regression, the intuition leads to tackle it using the neural network approach. A multilayer neural network takes the map coordinates as the input, and returns the histogram measures at that coordinates as the output. We expect the synaptic links to learn the histograms with respect to the positions (see
Figure 6).
Layers are fully connected via nonlinear activation functions (hyperbolic tangent and sigmoid) of the linear perceptron equations. We also add a bias for each node to make the structure learn an offset value if possible. The forward neural equations in matrix notation are given in (
7).
where
is of size
,
of
,
of
,
of
, and
of
, and the sizes
and
L denote the sample size, number of inputs (coordinates), hidden layer size, and number of outputs (histogram indices), respectively.
We employ the Keras Deep Learning Libary [
39] for training the neural network model and find the interlayer weights. The algorithm starts with uniformly randomized weights with values in
. The first layer is passed through the hyperbolic tangent and the second layer through the sigmoid function as the relations are considered to be nonlinear. We use the RMSprop, an adaptive gradient descent-based learning rate method as the optimizer [
40], and the mean squared error as the loss function.
With the weights in hand after training, we run the network in the forward direction to get a histogram estimation for the position of interpolation,
P:
where
is the size of the hidden layer.
2.3.4. Affine Wasserstein Combination (AWC)
The Wasserstein distance is originally a distance metric used to compare densities [
32,
33]. In the process of calculating the cost to transport the measures between densities, we also obtain a transport function that maps the measures of one density to the measures of the other one. We employ the transport function to produce an interpolation by transporting the mesaures gradually and to develop an affine combination between two or more histogram positions.
In a discrete state space, we replace the transport function,
with a transport matrix,
, which represents the measure to be transported from a one dimensional vector,
, to another one dimensional vector,
. In our case, these vectors are actually normalized histograms, whose bins are identical and each bin index corresponds to an integer between the given extremities.
In the discrete case [
41], the Wasserstein distance is the minimum cost of transporting one histogram onto another:
where
is the measure that would be transported from the location
i to the location
j and
is the cost multiplier that is usually related to the distance between the locations
.
If the cost function, c, is a linear function, the Wasserstein transport matrix is efficiently computed by scanning the array and keeping track of the transported weights between bins.
Wasserstein Interpolation
We can perform the interpolation operation in multiple ways (
Figure 7). In fact, we define two parameters,
and
, both in the interval of
[
33]. The actual interpolation between densities is controlled by the coefficient
, which decides how similar would the interpolation be to the original densities. For the values of
near zero, the interpolation will be similar to the source histogram, whereas for the higher values will make the interpolation resemble the destination histogram.
controls the evolution of the interpolaton between a linear interpolation (
) and a displacement interpolation (
) [
33].
Formally, given two histograms, namely
and
, we first compute the Wasserstein mapping,
, which is merely a matrix that demonstrates the transport plan of the measures. According to this transport plan, we calculate two intermediate plans,
and
, which represent the destination indices. The measures,
will be distributed to:
where
is the ceil operator as the indices should be integers. Indices of the final interpolation are computed as follows:
where
is the Kronecker delta operator.
Note that a two-position convex combination (kNF) of histograms is merely a special case of the Wasserstein combination, in which is set to 0.
Affine Wasserstein Combination
The interpolation of histograms can be generalized to their linear combinations by varying
on the real line, relaxing the limitation of the interval
to acquire values out of the interval. With the intermediate transport plans (
9), we redefine the interpolation operation at (
10) as:
The algorithm is given in Algorithm 1 and the combinations for different values of
and
can be seen in
Figure 8.
Algorithm 1 Affine Wasserstein Histogram Combination |
- 1:
Initial two histograms and . - 2:
Calculate the discrete Wasserstein mapping . - 3:
Initialize the interpolation with zeros, . - 4:
for all do - 5:
Calculate new histogram indices of the measure - 6:
Distribute the weight on these indices - 7:
end for - 8:
return
|
Pair Selection Strategy (Loose Alignment)
For a given position,
P, we find the nearest fingerprint positions
from the set of all fingerprint locations,
. If an arbitrary point in a closed area (
) is
loosely aligned with two fingerprint locations on which some histograms are defined, we can compute an affine combination of histograms on this point
P using the two fingerprints. A point
P is aligned loosely with two other points
if
where
defines the orthogonal distance of a point to a line,
is the set of affine combinations or simply the line passing through the points
and
, and
is the
looseness parameter (See
Figure 9).
Finally, if there exist predefined histograms at positions and and if P is aligned loosely with these positions with respect to a previously defined , we can apply an affine Wasserstein combination of histograms between and at P.
Two-Position Affine Wasserstein Combination (AWC-B)
As an attenuation model of wireless signals does not necessarily behave linearly, we propose multiple mappings, , that map the distance between map positions to the Wasserstein similarity parameter, . These mappings are linear, quadratic, cubic, logarithmic, and exponential mappings, . This mapping is another parameter to be decided upon.
With
being the Wasserstein mapping from the histogram at
to the histogram at
, the affine combination of histograms between
and
at
P is computed. An estimator for the probability density at the position
P would be:
The original Wasserstein transport problem is defined to be an interpolation [
33], where
is restricted to be in
, however without this limitation in
, the interpolation generalizes to an affine combination, so that the position
P does not necessarily have to reside in the convex hull of the fingerprints. The method is expected to give a likelihood density estimation out of this hull, as the candidate position
P may exist out of the hull. The definition does not depend on the density of the fingerprints, but the accuracy would depend on the scatter of the fingerprints.
Multiple-Position Combination (AWC-E)
The Wasserstein combination method for a single position can be employed using only two fingerprints, but the method is prone to losing any other surrounding information. In order to take advantage of the whole map and be able to compare this method with other estimation algorithms that use more than two fingerprints, we develop an extension on the affine Wasserstein combination method.
According to this extension, for the position of interest,
P, we select all of the possible fingerprint pairs that can generate a combination according to the selection model. For each pair we perform the pairwise combination and obtain a histogram estimation on
P with each pair. The final combination,
, is found by taking the affine combination of the available two-position affine Wasserstein combinations,
, and averaging them with the weights,
, inverse exponentially proportional to the pairwise average distances to
P. We will denote the final histogram with
:
2.4. Inference with the SMC Filter
Having a transition density between positions and an estimate for the observation density to evaluate the measurements with respect to the positions, we have all the ingredients to build up a sequential Monte Carlo method (particle filter) [
42,
43] to track an agent that reads data from the surrounding beacons. The algorithm of the SMC filter is given in Algorithm 2.
In our setting, we use the generative model defined in (
1) with the graphical model in
Figure 3. For the transition density,
, we employ the diffusion motion model (
2), assuming that no rotation or translation information is supplied. For the observation density,
, we have five different estimators, listed in
Section 2.3 as (
4), (
5), (
8), (
12) and (
13).
Algorithm 2 SMC Filter for BLE Localization |
- 1:
Instantiate For , sample - 2:
Importance Sampling Update time index For , sample For , evaluate to obtain the importance weights Normalize the importance weights - 3:
Select Resample N particles as from the particles according to the importance weights. - 4:
Recurse with repeating 2.
|
4. Simulations and Results
We defined five different estimator methods to estimate the observation densities:
,
,
,
, and
. The estimations are performed using 50, 32, 21, 15, and 8 measurement locations. The estimators estimate the observation densities for particle evaluation in the update step of the SMC algorithm. For the prediction step we employ the motion model defined in
Section 2.2, and the recursive loop of the SMC algorithm is closed. The simulations are designed with Python-3.4 and are run with a particle size of 10 K on an Intel Xeon 2630. The particle evaluation step is parallelized by 32 processes. Each iteration takes about 0.8 s. A snapshot of the running SMC algorithm is given in
Figure 14.
The SMC filter was run for the same trajectory of 232 points, with 32 of them discarded as the burn-in period, leaving us with 200 points of error for each run. We log the distances of the predicted positions to the original positions as errors. Each combination (estimator and fingerprint set size) is repeated 30 times, which gives 6000 position estimations or error measurements per combination. The box plot of the statistics is given in
Figure 15. We report both mean and median errors.
It can be seen from the results that high-resolution fingerprint information with 50 fingerprint locations (FL) defines the map the best for the tracking purposes with any of the applied methods. Without employing any complicated method, that is, using only in estimating the observation density, gives the best result: an error rate of 0.66 m, with the lowest variance. This is an expected result but with an impractical setting. With more than 50 fingerprint locations, this error rate will surely get lower, but making the scene analysis much more of a burden at the same time. Collecting data from many locations is obviously not practical. A reduced number of fingerprints would facilitate the installation and calibration procedure in BLE positioning and tracking infrastructure, so that the results for such scenarios are substantial and realistic for us.
As the configurations lose a number of fingerprint locations, we see that is unable to keep its success with the highest resolution, and Wasserstein interpolation-based techniques stand out. Even though the error rates are over 1 m, looks especially promising with an error rate of 1.29 m with 15 fingerprint locations. The same algorithm gives an error rate of 1.9 m with only 8 fingerprint locations.
The neural network approach () is both inconvenient as we do not have many samples for training and impractical as the training iterations take too long. Moreover, the results show that yields consistently higher error rates compared to the other methods. These results are expected whilst the neural networks require many samples beforehand, and we also try to reduce the required fingerprint size, which makes naturally inappropriate for our purposes.
Table 1 summarizes the mean localization errors for different fingerprint confugurations with respect to the applied estimator algorithms. Concentrating on the results with lower number of fingerprints (15 and 8), we see that Wasserstein-based methods race head to head with the
method. We also run two
t-tests for significance analysis on the last two configurations. In the significance test tables we report the
p-values for the one-sided hypothesis if the value on the row is less than the value on the column. According to the
p-values, for 15 fingerprint locations (see
Table 2),
has the best localization performance, but for the configuration of 8 fingerprint locations (see
Table 3), there is no best estimator, because
and
cannot outperform each other.
Even though the box plots in
Figure 15 give a hint about error distributions, we also supply the explicit distributions and corresponding cumulative error distributions. In
Figure 16, we encode the methods with the same color codes with
Figure 15.
The figures clearly show that the error distributions are not normally distributed, which is probably due to sudden incorrect far position estimations in the SMC filter. We believe that such results are better shown with box plots and medians, as the errors have skewed distributions, where means would be misleading. We also report the error medians in
Table 4.
For the significance test, because the error results are not necessarily normally distributed, we apply one-sided Wilcoxon signed-rank tests on the error rates of the configuration pairs. We report the two low sized fingerprint sets’ values with the confidence value of .
Table 5 shows that for the configuration of 15 fingerprint locations, the proposed methods
and
, perform significantly better than any other methods we compared. For a smaller configuration of eight locations, we see almost the same results.
Table 6 shows that the proposed methods perform significantly better than the applied techniques, except against
, in which we can say that our methods perform slightly better than
. This leads to the claim that, for the small-sized fingerprint sets (8 and 15), Wasserstein-based interpolation techniques reduce the errors significantly, (except for only one case) compared to the applied techniques so far. Moreover, as the complexity of the methods is similar, Wasserstein interpolation techniques are preferable for small-sized fingerprint sets. Amongst the proposed methods,
is seen to perform slightly better than its counterpart
.
Thus, we conclude that that two methods and are the two candidates to be used in observation density estimation in the SMC filter for tracking purposes with BLE beacon information for lower number of fingerprints if the positions of these fingerprints are scattered evenly in order to perform better histogram interpolations.
5. Conclusions
In this work, we have developed a method to render indoor localization and tracking practical using only BLE sensors. Our model-based approach relies critically on the accurate estimation of a probabilistic radio map—a distribution of RSSI values for every position inside the region of interest—from a few RSSI fingerprint measurements, obtained only at a few locations. The estimated radio map is subsequently used as the observation model of a dynamical system where we do target tracking by a sequential Monte Carlo algorithm.
Not surprisingly, there is a direct relationship between the accuracy of the radiomap and accuracy of localization and tracking. We show first that when RSSI fingerprints are collected on a very dense grid, a radio map can be accurately estimated and a satisfactory tracking performance can be obtained despite the high variability of the actual measurements. However, dense fingerprint sampling is often not practical or at least not desired in real applications as this requires careful measurement and a long data collection process. In this paper, we develop methods to obtain accurate radio maps from far fewer and sparsely sampled fingerprints and show empirically that tracking accuracy is still acceptable.
We cast the radio map estimation as a fingerprint interpolation method, that we cast as a regression problem. We have proposed two variations of the Wasserstein interpolation method, which is also originally derived from the Wasserstein distance computation method. The first one uses two points as the pivot points to find an affine combination on an unknown coordinate aligned with the pivot points. The second one fuses multiple affine combinations to estimate a histogram on an unknown coordinate. The results of these Wasserstein variations are compared with the results of the well known regression methods, namely the nearest neighbor approach, the affine combination and the neural networks
We can generalize from the results that we can estimate the unknown radio information on the arbitrary positions by an interpolation using the radio strength densities on previously measured positions. Moreover, Wasserstein combination variations are good candidates for histogram estimation purposes, as the error rates of these metods increase consistently while the number of fingerprints are reduced, but they perform better compared to the other applied techniques.
An expected result is that with such small data, neural networks are unable to learn the nonlinear histogram on a planar surface without overlearning. The domain can benefit from the autoregressive models to train the histograms against the positions.
In fingerprinting techniques, initial scene analysis is a must, but it is not practical to collect data from high number of locations. For future work, an immediate study could to perform an analysis on the appropriate fingerprint locations. Having the practicality in mind, the researcher can also find the number of fingerprint locations with respect to the area size and the required fineness. Additionally, if a method can give confidence for a position in the area, it may urge the researcher to gather data from locations with low confidence values. A proper Gaussian process regression for multiple outputs would be a good future study using this data, as the algorithm by nature provides the confidence values.