Model Free Localization with Deep Neural Architectures by Means of an Underwater WSN
<p>Example of feedforward neural network. Each circle represents a neuron, which combines in a nonlinear way its inputs following Equation (<a href="#FD3-sensors-19-03530" class="html-disp-formula">3</a>). The inputs are <math display="inline"><semantics> <msub> <mi>x</mi> <mn>1</mn> </msub> </semantics></math> and <math display="inline"><semantics> <msub> <mi>x</mi> <mn>2</mn> </msub> </semantics></math>, and the outputs are <math display="inline"><semantics> <msub> <mi>z</mi> <mn>1</mn> </msub> </semantics></math> and <math display="inline"><semantics> <msub> <mi>z</mi> <mn>2</mn> </msub> </semantics></math>. There is a single hidden layer, which has three neurons. Note how each of the outputs <span class="html-italic">z</span> is a nonlinear combination of the inputs <math display="inline"><semantics> <msub> <mi>x</mi> <mn>1</mn> </msub> </semantics></math> and <math display="inline"><semantics> <msub> <mi>x</mi> <mn>2</mn> </msub> </semantics></math>.</p> "> Figure 2
<p>Illustration of the procedure of an LSTM for three timesteps. The output <math display="inline"><semantics> <msub> <mi>y</mi> <mi>t</mi> </msub> </semantics></math> is updated in each timestep using Equation (<a href="#FD5-sensors-19-03530" class="html-disp-formula">5</a>) and the cell state <math display="inline"><semantics> <msub> <mi>c</mi> <mi>t</mi> </msub> </semantics></math> is updated using Equation (<a href="#FD4-sensors-19-03530" class="html-disp-formula">4</a>). The LSTM block is composed of four neural networks, which are the same for all timesteps. Note that, in the first timestep, it is necessary to provide an initial <math display="inline"><semantics> <msub> <mi>c</mi> <mn>0</mn> </msub> </semantics></math> and <math display="inline"><semantics> <msub> <mi>y</mi> <mn>0</mn> </msub> </semantics></math> in order to obtain <math display="inline"><semantics> <msub> <mi>c</mi> <mn>1</mn> </msub> </semantics></math> and <math display="inline"><semantics> <msub> <mi>y</mi> <mn>1</mn> </msub> </semantics></math>.</p> "> Figure 3
<p>Block diagram illustrating the steps needed to obtain the covariance feature vector.</p> "> Figure 4
<p>The localization DNN architecture used, where Tanh stands for hyperbolic tangent function. The number of neurons of the first layer depends on the dimensionality of the feature vector used: <span class="html-italic">S</span> neurons if power features are used, and <math display="inline"><semantics> <mrow> <mi>S</mi> <mo>(</mo> <mi>S</mi> <mo>+</mo> <mn>1</mn> <mo>)</mo> </mrow> </semantics></math> if covariance features are used.</p> "> Figure 5
<p>The localization procedure overview. From a set of trajectories, we obtain the signal received in each sensor <span class="html-italic">s</span>, <math display="inline"><semantics> <mrow> <msub> <mi>y</mi> <mi>s</mi> </msub> <mrow> <mo>(</mo> <mi>t</mi> <mo>)</mo> </mrow> </mrow> </semantics></math>, which is then used to extract power or covariance features, and these features are grouped in sequences to form the input data to our localization DNN. The output data are the trajectory Cartesian coordinates <span class="html-italic">z</span>. By training the DNN in this way, we minimize the error between the predicted localization by the localization DNN, <math display="inline"><semantics> <mover accent="true"> <mi>z</mi> <mo>˜</mo> </mover> </semantics></math>, and the actual localization <span class="html-italic">z</span>.</p> "> Figure 6
<p>Example trajectories, for <math display="inline"><semantics> <mrow> <mi>S</mi> <mi>N</mi> <mi>R</mi> <mo>=</mo> <mn>0</mn> </mrow> </semantics></math> dB and <math display="inline"><semantics> <mrow> <mi>N</mi> <mo>=</mo> <mn>64</mn> </mrow> </semantics></math>. The black points represent the sensor positions. Blue is the actual value of <span class="html-italic">z</span>, and, in red, we observe the localization DNN prediction, <math display="inline"><semantics> <mover accent="true"> <mi>z</mi> <mo>˜</mo> </mover> </semantics></math>, using power features, and, in green, using covariance features. The axis are in <span class="html-italic">m</span>.</p> "> Figure 7
<p>Mean Squared Error (MSE) during training, when <math display="inline"><semantics> <mrow> <mi>N</mi> <mo>=</mo> <mn>64</mn> </mrow> </semantics></math>. We can observe that, for low <math display="inline"><semantics> <mi>θ</mi> </semantics></math> and high SNR, the localization DNN converges to a low localization error. However, as <math display="inline"><semantics> <mi>θ</mi> </semantics></math> increases, i.e., the channel becomes more variable, and SNR diminishes, i.e., the received noise signal increases, the localization precision significantly worsens. In addition, as <math display="inline"><semantics> <mi>θ</mi> </semantics></math> increases and SNR diminishes, there appears overfitting: the validation values remain constant while the training values decrease, which means that there is a certain localization precision bound that our localization DNN is not able to improve.</p> "> Figure 8
<p>Validation Mean Absolute Error (MAE), in <span class="html-italic">m</span>, for different <math display="inline"><semantics> <mi>θ</mi> </semantics></math>, SNR and <span class="html-italic">N</span>. We use the MAE because its meaning is more intuitive than MSE as the mean error in the distance between <span class="html-italic">z</span> and <math display="inline"><semantics> <mover accent="true"> <mi>z</mi> <mo>˜</mo> </mover> </semantics></math>. Note that, as <math display="inline"><semantics> <mi>θ</mi> </semantics></math> increases, the MAE increases as well, specially from <math display="inline"><semantics> <mrow> <mi>θ</mi> <mo>=</mo> <mn>0.4</mn> </mrow> </semantics></math> onwards. We can also observe that, for <math display="inline"><semantics> <mrow> <mi>N</mi> <mo>=</mo> <mo>{</mo> <mn>8</mn> <mo>,</mo> <mn>16</mn> <mo>}</mo> </mrow> </semantics></math>, the localization DNN is not learning to locate accurately, which does for higher <span class="html-italic">N</span>, that is, for more neurons in the hidden layer. However, note that the increase in accuracy between <math display="inline"><semantics> <mrow> <mi>N</mi> <mo>=</mo> <mn>64</mn> </mrow> </semantics></math> and <math display="inline"><semantics> <mrow> <mi>N</mi> <mo>=</mo> <mn>128</mn> </mrow> </semantics></math> is minimal. Finally, observe that the results are similar for all SNR values and that, for low <math display="inline"><semantics> <mi>θ</mi> </semantics></math>, using power or covariance features does not matter, but it does for higher <math display="inline"><semantics> <mi>θ</mi> </semantics></math> values.</p> "> Figure 8 Cont.
<p>Validation Mean Absolute Error (MAE), in <span class="html-italic">m</span>, for different <math display="inline"><semantics> <mi>θ</mi> </semantics></math>, SNR and <span class="html-italic">N</span>. We use the MAE because its meaning is more intuitive than MSE as the mean error in the distance between <span class="html-italic">z</span> and <math display="inline"><semantics> <mover accent="true"> <mi>z</mi> <mo>˜</mo> </mover> </semantics></math>. Note that, as <math display="inline"><semantics> <mi>θ</mi> </semantics></math> increases, the MAE increases as well, specially from <math display="inline"><semantics> <mrow> <mi>θ</mi> <mo>=</mo> <mn>0.4</mn> </mrow> </semantics></math> onwards. We can also observe that, for <math display="inline"><semantics> <mrow> <mi>N</mi> <mo>=</mo> <mo>{</mo> <mn>8</mn> <mo>,</mo> <mn>16</mn> <mo>}</mo> </mrow> </semantics></math>, the localization DNN is not learning to locate accurately, which does for higher <span class="html-italic">N</span>, that is, for more neurons in the hidden layer. However, note that the increase in accuracy between <math display="inline"><semantics> <mrow> <mi>N</mi> <mo>=</mo> <mn>64</mn> </mrow> </semantics></math> and <math display="inline"><semantics> <mrow> <mi>N</mi> <mo>=</mo> <mn>128</mn> </mrow> </semantics></math> is minimal. Finally, observe that the results are similar for all SNR values and that, for low <math display="inline"><semantics> <mi>θ</mi> </semantics></math>, using power or covariance features does not matter, but it does for higher <math display="inline"><semantics> <mi>θ</mi> </semantics></math> values.</p> ">
Abstract
:1. Introduction
- We study the localization accuracy of a simple DNN using the underwater acoustic channel characterization proposed in [14], which models a time-varying shallow water acoustic channel, which takes into account several propagation effects such as multipath or frequency-dependent path loss. It is a realistic channel model, which allows us to test the accuracy of the DNN approach under different propagation conditions.
- We make use of a recurrent neural networks architecture to locate, studying the results obtained for different sizes of the neural network in order to evaluate how this influences the localization precision. As recurrent neural networks have memory, they not only learn to locate, but also to track the target, as [15] shows.
- We compare the localization accuracy using both power measurements and the covariance matrix, for several channel parameters, DNN architectures and Signal to Noise Ratio (SNR) values. To the best of our knowledge, ours is the first work in which the DNN approach is put to the test under all these conditions, which allows us to draw conclusions on both the precision and the convenience of using DNN for underwater localization. Currently, covariance measurements are used for underwater location, but we show that, under certain conditions, power measurements can be enough for underwater localization. This opens the door to new localization methods that use received signal power as input.
2. Background
2.1. Channel Model
- The paths are indexed by p, where denotes the reference path transfer function. The effect of each propagation path is considered as a low-pass filter, which takes into account both the attenuation expressed in Equation (1) and the reflections encountered along the propagation path.
- and are, respectively, the channel gain and delay for each path. They are obtained taking into account variations in the path length—for instance, due to tides.
- The coefficient includes other propagation effects, namely, scattering, correlation and Doppler shifting.
2.2. Neural Networks
- First, is updated with the following expression:The second term in Equation (4) intuitively controls what new information we are adding to the cell state. Note that the hyperbolic tangent term could be considered the new information that the LSTM wants to add to the cell state, whereas the sigmoid term controls again how much of that information will be added to the cell state. Thus, the cell state update consists of two main terms: the first controls how much information from the previous state cell is remembered, and the second how much information from the current input and previous output we are adding to the state cell to remember in the next timesteps.
- Second, we obtain the output to the LSTM using the following expression:
3. Problem Setup
3.1. Target Signal Model
3.2. Received Signal Model
3.3. Covariance Matrix Feature Extraction
3.4. Power Feature Extraction
3.5. Localization Neural Network
4. Results
4.1. Simulations Setup
4.2. Results Obtained
- Regarding N, the number of neurons of the hidden layer, we note that having more neurons does benefit our localization precision, as it decreases the MAE. Actually, for and , the localization DNN does not learn to locate, but it does for . We remark that, while adding more neurons does increase the localization precision, the marginal increase in the MAE is low when we pass from 64 to 128 neurons. In addition, note that there is a lower bound in the localization precision for all values, which is around 50 m in our setup, due to the uncertainties in the localization that are derived from the channel propagation effects.
- Regarding , we reach the same conclusion as in Figure 7: as increases, the precision localization decreases, as the MAE increases. Note that this affects , all SNRs and both power and covariance features. This is no surprise at all: as we showed in the Introduction, the main obstacle to using power measures for locating purposes is precisely the complexities of the underwater channel.
- Regarding the SNR, we note that it does not play a significant role in determining the localization precision in our setup, as the plots are similar for all SNR cases.
- Regarding the features used, note that, for low values, covariance and power features reach a similar localization precision, which is to be noted as the power feature is a single scalar per sensor, while covariance is a vector. However, as increases, power features’ localization precision has a faster increase in the MAE, which means that covariance locates better, although the localization results are not good for any of these features. These results are to be expected: covariance features include more information than power features, and hence their performance should not be worse than power. It is remarkable, however, that power features include enough information for locating accurately when the channel variability is low.
5. Conclusions
Author Contributions
Funding
Acknowledgments
Conflicts of Interest
Abbreviations
WSN | Wireless Sensor Network |
AUWSN | Acoustic Underwater Wireless Sensor Network |
DNN | Deep Neural Network |
SNR | Signal-to-Noise Ratio |
RNN | Recurrent Neural Network |
LSTM | Long-Short Term Memory |
MSE | Mean Squared Error |
MAE | Mean Absolute Error |
References
- Felemban, E.; Shaikh, F.K.; Qureshi, U.M.; Sheikh, A.A.; Qaisar, S.B. Underwater sensor network applications: A comprehensive survey. Int. J. Distrib. Sens. Netw. 2015, 11, 896832. [Google Scholar] [CrossRef]
- Erol-Kantarci, M.; Mouftah, H.T.; Oktug, S. A survey of architectures and localization techniques for underwater acoustic sensor networks. IEEE Commun. Surv. Tutor. 2011, 13, 487–502. [Google Scholar] [CrossRef]
- Tan, H.P.; Diamant, R.; Seah, W.K.; Waldmeyer, M. A survey of techniques and challenges in underwater localization. Ocean Eng. 2011, 38, 1663–1676. [Google Scholar] [CrossRef] [Green Version]
- Han, G.; Jiang, J.; Shu, L.; Xu, Y.; Wang, F. Localization algorithms of underwater wireless sensor networks: A survey. Sensors 2012, 12, 2026–2061. [Google Scholar] [CrossRef] [PubMed]
- Chandrasekhar, V.; Seah, W.K.; Choo, Y.S.; Ee, H.V. Localization in underwater sensor networks: Survey and challenges. In Proceedings of the 1st ACM International Workshop on Underwater Networks, Los Angeles, CA, USA, 25 September 2006; pp. 33–40. [Google Scholar]
- Diamant, R.; Tan, H.P.; Lampe, L. NLOS identification using a hybrid ToA-signal strength algorithm for underwater acoustic localization. In Proceedings of the OCEANS 2010 MTS/IEEE SEATTLE, Seattle, WA, USA, 20–23 September 2010; pp. 1–7. [Google Scholar]
- Zhou, Z.; Peng, Z.; Cui, J.H.; Shi, Z.; Bagtzoglou, A. Scalable localization with mobility prediction for underwater sensor networks. IEEE Trans. Mob. Comput. 2011, 10, 335–348. [Google Scholar] [CrossRef]
- Zazo, J.; Macua, S.V.; Zazo, S.; Pérez, M.; Pérez-Álvarez, I.; Jiménez, E.; Cardona, L.; Brito, J.H.; Quevedo, E. Underwater electromagnetic sensor networks, part II: Localization and network simulations. Sensors 2016, 16, 2176. [Google Scholar] [CrossRef] [PubMed]
- Dubrovinskaya, E.; Diamant, R.; Casari, P. Anchorless underwater acoustic localization. In Proceedings of the 2017 14th Workshop on Positioning, Navigation and Communications (WPNC), Bremen, Germany, 25–26 October 2017; pp. 1–6. [Google Scholar]
- Dai, H.; Ying, W.H.; Xu, J. Multi-layer neural network for received signal strength-based indoor localisation. IET Commun. 2016, 10, 717–723. [Google Scholar] [CrossRef]
- Payal, A.; Rai, C.S.; Reddy, B.R. Analysis of some feedforward artificial neural network training algorithms for developing localization framework in wireless sensor networks. Wirel. Pers. Commun. 2015, 82, 2519–2536. [Google Scholar] [CrossRef]
- Niu, H.; Reeves, E.; Gerstoft, P. Source localization in an ocean waveguide using supervised machine learning. J. Acoust. Soc. Am. 2017, 142, 1176–1188. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Huang, Z.; Xu, J.; Gong, Z.; Wang, H.; Yan, Y. A Deep Neural Network Based Method of Source Localization in a Shallow Water Environment. In Proceedings of the 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Calgary, AB, Canada, 15–20 April 2018; pp. 3499–3503. [Google Scholar]
- Qarabaqi, P.; Stojanovic, M. Statistical characterization and computationally efficient modeling of a class of underwater acoustic communication channels. IEEE J. Ocean. Eng. 2013, 38, 701–717. [Google Scholar] [CrossRef]
- Ondruska, P.; Posner, I. Deep tracking: Seeing beyond seeing using recurrent neural networks. In Proceedings of the Thirtieth AAAI Conference on Artificial Intelligence, Phoenix, AR, USA, 2 February 2016. [Google Scholar]
- Qarabaqi, P.; Stojanovic, M. Statistical modeling of a shallow water acoustic communication channel. In Proceedings of the 2010—MILCOM 2010 Military Communications Conference, San Jose, CA, USA, 31 October–3 November 2009; pp. 1341–1350. [Google Scholar]
- Socheleau, F.X.; Passerieux, J.M.; Laot, C. Characterisation of time-varying underwater acoustic communication channel with application to channel capacity. In Proceedings of the Underwater Acoustic Measurements, Nafplion, Greece, 21–26 June 2009. [Google Scholar]
- Tomasi, B.; Casari, P.; Badia, L.; Zorzi, M. A study of incremental redundancy hybrid ARQ over Markov channel models derived from experimental data. In Proceedings of the Fifth ACM International Workshop on UnderWater Networks, Woods Hole, MA, USA, 30 September–1 October 2010; p. 4. [Google Scholar]
- Yang, W.B.; Yang, T. High-frequency channel characterization for M-ary frequency-shift-keying underwater acoustic communications. J. Acoust. Soc. Am. 2006, 120, 2615–2626. [Google Scholar] [CrossRef]
- Brekhovskikh, L.; Lysanov, J.P.; Lysanov, Y.P. Fundamentals of Ocean Acoustics; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2003. [Google Scholar]
- Etter, P.C. Underwater Acoustic Modeling and Simulation; CRC Press: Boca Raton, FL, USA, 2018. [Google Scholar]
- Hornik, K.; Stinchcombe, M.; White, H. Multilayer feedforward networks are universal approximators. Neural Netw. 1989, 2, 359–366. [Google Scholar] [CrossRef]
- LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
- Zhao, B.; Feng, J.; Wu, X.; Yan, S. A survey on deep learning-based fine-grained object classification and semantic segmentation. Int. J. Autom. Comput. 2017, 14, 119–135. [Google Scholar] [CrossRef]
- Mnih, V.; Kavukcuoglu, K.; Silver, D.; Rusu, A.A.; Veness, J.; Bellemare, M.G.; Graves, A.; Riedmiller, M.; Fidjeland, A.K.; Ostrovski, G.; et al. Human-level control through deep reinforcement learning. Nature 2015, 518, 529–533. [Google Scholar] [CrossRef] [PubMed]
- Wang, X.; Gao, L.; Mao, S.; Pandey, S. DeepFi: Deep learning for indoor fingerprinting using channel state information. In Proceedings of the Wireless Communications and Networking Conference (WCNC), New Orleans, LA, USA, 9–12 March 2015; pp. 1666–1671. [Google Scholar]
- Hecht-Nielsen, R. Theory of the backpropagation neural network. In Neural Networks for Perception; Elsevier: Amsterdam, The Netherlands, 1992; pp. 65–93. [Google Scholar]
- Kingma, D.P.; Ba, J. Adam: A Method for Stochastic Optimization. In Proceedings of the 3rd International Conference on Learning Representations, San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
- Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef] [PubMed]
- Werbos, P.J. Backpropagation through time: What it does and how to do it. Proc. IEEE 1990, 78, 1550–1560. [Google Scholar] [CrossRef]
- Ostrowski, Z.; Marszal, J.; Salamon, R. Underwater Navigation System Based on Doppler Shifts of a Continuous Wave. In Proceedings of the 2018 Joint Conference-Acoustics, Ustka, Poland, 11–14 September 2018; pp. 1–6. [Google Scholar]
- Aktas, B.; Atlar, M.; Turkmen, S.; Shi, W.; Sampson, R.; Korkut, E.; Fitzsimmons, P. Propeller cavitation noise investigations of a research vessel using medium size cavitation tunnel tests and full-scale trials. Ocean Eng. 2016, 120, 122–135. [Google Scholar] [CrossRef] [Green Version]
- Nielsen, R.O. Sonar Signal Processing; Artech House, Inc.: Norwood, MA, USA, 1991. [Google Scholar]
- Luo, H.; Wu, K.; Ruby, R.; Hong, F.; Guo, Z.; Ni, L.M. Simulation and experimentation platforms for underwater acoustic sensor networks: Advancements and challenges. ACM Comput. Surv. (CSUR) 2017, 50, 28. [Google Scholar] [CrossRef]
- Acoustic Channel Modeling and Simulation. Available online: http://millitsa.coe.neu.edu/?q=projects (accessed on 10 July 2019).
- Ioffe, S.; Szegedy, C. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift. In Proceedings of the 32nd International Conference on Machine Learning, Lille, France, 6–11 July 2015; Volume 37, pp. 448–456. [Google Scholar]
- Antoni, J.; Hanson, D. Detection of surface ships from interception of cyclostationary signature with the cyclic modulation coherence. IEEE J. Ocean. Eng. 2012, 37, 478–493. [Google Scholar] [CrossRef]
Parameter | Value |
---|---|
Spreading factor k | |
Speed of sound in water (m/s) | 1500 |
Speed of sound in bottom (m/s) | 1200 |
Minimum relative path strength | 50 |
Frequency band (kHz) | |
Frequency resolution (Hz) | 25 |
Coherence time small scale (s) | 6 |
Variance of small scale surface variations (m) | |
Variance of small scale bottom variations (m) | |
3-dB width of psd of intrapath, (Hz) | |
Number of intrapaths | |
Mean of intrapaths amplitudes | |
Variance of intrapaths amplitudes | |
Range of surface height (m) | |
Range of target height (m) | |
Range of sensor height (m) | |
Range of channel distance height (m) | |
Large scale standard deviation of surface height | |
Large scale standard deviation of target height | |
Large scale standard deviation of sensor height | |
Large scale standard deviation of channel distance | |
Large scale auto regressive process parameter |
© 2019 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
Share and Cite
Parras, J.; Zazo, S.; Pérez-Álvarez, I.A.; Sanz González, J.L. Model Free Localization with Deep Neural Architectures by Means of an Underwater WSN. Sensors 2019, 19, 3530. https://doi.org/10.3390/s19163530
Parras J, Zazo S, Pérez-Álvarez IA, Sanz González JL. Model Free Localization with Deep Neural Architectures by Means of an Underwater WSN. Sensors. 2019; 19(16):3530. https://doi.org/10.3390/s19163530
Chicago/Turabian StyleParras, Juan, Santiago Zazo, Iván A. Pérez-Álvarez, and José Luis Sanz González. 2019. "Model Free Localization with Deep Neural Architectures by Means of an Underwater WSN" Sensors 19, no. 16: 3530. https://doi.org/10.3390/s19163530
APA StyleParras, J., Zazo, S., Pérez-Álvarez, I. A., & Sanz González, J. L. (2019). Model Free Localization with Deep Neural Architectures by Means of an Underwater WSN. Sensors, 19(16), 3530. https://doi.org/10.3390/s19163530