[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
sensors-logo

Journal Browser

Journal Browser

Intelligent Sensor Signal in Machine Learning

A special issue of Sensors (ISSN 1424-8220). This special issue belongs to the section "Intelligent Sensors".

Deadline for manuscript submissions: closed (31 August 2019) | Viewed by 90369

Special Issue Editors


E-Mail Website
Guest Editor
Department of Computer Engineering, Keimyung University, Shindang-Dong, Dalseo-Gu, Daegu 704-701,Republic of Korea
Interests: computer vision; pattern recognition; object detection tracking; deep learning
Special Issues, Collections and Topics in MDPI journals

E-Mail Website
Guest Editor
Department of Computer Engineering, Keimyung University, Daegu 704-701, Republic of Korea
Interests: camera calibration; computer vision; image processing; signal processing
Special Issues, Collections and Topics in MDPI journals

Special Issue Information

Dear Colleagues.

With the advancement of sensor technology, research has been actively carried out to fuse sensor signals and to extract useful information for various recognition problems based on machine learning. Recently, we have been obtaining signals from various sensors, such as wearable sensors, mobile sensors, cameras, heart rate monitoring devices, EEG head-caps and headbands, ECG sensors, breathing monitors, EMG sensors, and temperature sensors. However, as the sensor signal itself has no meaning, the machine learning algorithm must be combined in order to process the signals and make various decisions. Therefore, the use of machine learning, including deep learning, is appropriate for these challenging tasks.

The purpose of this Special Issue is to take the opportunity to introduce the current developments of intelligent sensor applications and innovative sensor fusion techniques combined with machine learning, including computer vision, pattern recognition, expert systems, deep learning, and so on. In this Special Issue, you are invited to submit contributions of original research, advancement, developments, and experiments pertaining to machine learning combined with sensors. Therefore, this Special Issue welcomes the newly developed methods and ideas combining the data obtained from various sensors in the following fields (but not limited to these fields):

  • Sensor fusion techniques based on machine learning
  • Sensors and big data analysis with machine learning
  • Autonomous vehicle technologies combining sensors and machine learning
  • Wireless sensor networks and communication based on machine learning
  • Deep network structure/learning algorithm for intelligent sensing
  • Autonomous robotics with intelligent sensors and machine learning 
  • Multi-modal/task learning for decision-making and control
  • Decision algorithms for autonomous driving
  • Machine learning and artificial intelligence for traffic/quality of experience management in IoT
  • Fuzzy fusion of sensors, data, and information
  • Machine learning for IoT and sensor research challenges
  • Advanced driver assistant systems (ADAS) based on machine learning
  • State-of-practice, research overview, experience reports, industrial experiments, and case studies in the intelligent sensors or IoT

Prof. Dr. ByoungChul Ko
Dr. Deokwoo Lee
Guest Editors

Manuscript Submission Information

Manuscripts should be submitted online at www.mdpi.com by registering and logging in to this website. Once you are registered, click here to go to the submission form. Manuscripts can be submitted until the deadline. All submissions that pass pre-check are peer-reviewed. Accepted papers will be published continuously in the journal (as soon as accepted) and will be listed together on the special issue website. Research articles, review articles as well as short communications are invited. For planned papers, a title and short abstract (about 100 words) can be sent to the Editorial Office for announcement on this website.

Submitted manuscripts should not have been published previously, nor be under consideration for publication elsewhere (except conference proceedings papers). All manuscripts are thoroughly refereed through a single-blind peer-review process. A guide for authors and other relevant information for submission of manuscripts is available on the Instructions for Authors page. Sensors is an international peer-reviewed open access semimonthly journal published by MDPI.

Please visit the Instructions for Authors page before submitting a manuscript. The Article Processing Charge (APC) for publication in this open access journal is 2600 CHF (Swiss Francs). Submitted papers should be well formatted and use good English. Authors may use MDPI's English editing service prior to publication or during author revisions.

Benefits of Publishing in a Special Issue

  • Ease of navigation: Grouping papers by topic helps scholars navigate broad scope journals more efficiently.
  • Greater discoverability: Special Issues support the reach and impact of scientific research. Articles in Special Issues are more discoverable and cited more frequently.
  • Expansion of research network: Special Issues facilitate connections among authors, fostering scientific collaborations.
  • External promotion: Articles in Special Issues are often promoted through the journal's social media, increasing their visibility.
  • e-Book format: Special Issues with more than 10 articles can be published as dedicated e-books, ensuring wide and rapid dissemination.

Further information on MDPI's Special Issue polices can be found here.

Published Papers (18 papers)

Order results
Result details
Select all
Export citation of selected articles as:

Research

Jump to: Review

15 pages, 36176 KiB  
Article
Panchromatic Image Super-Resolution Via Self Attention-Augmented Wasserstein Generative Adversarial Network
by Juan Du, Kuanhong Cheng, Yue Yu, Dabao Wang and Huixin Zhou
Sensors 2021, 21(6), 2158; https://doi.org/10.3390/s21062158 - 19 Mar 2021
Cited by 7 | Viewed by 2899
Abstract
Panchromatic (PAN) images contain abundant spatial information that is useful for earth observation, but always suffer from low-resolution ( LR) due to the sensor limitation and large-scale view field. The current super-resolution (SR) methods based on traditional attention mechanism have shown remarkable advantages [...] Read more.
Panchromatic (PAN) images contain abundant spatial information that is useful for earth observation, but always suffer from low-resolution ( LR) due to the sensor limitation and large-scale view field. The current super-resolution (SR) methods based on traditional attention mechanism have shown remarkable advantages but remain imperfect to reconstruct the edge details of SR images. To address this problem, an improved SR model which involves the self-attention augmented Wasserstein generative adversarial network ( SAA-WGAN) is designed to dig out the reference information among multiple features for detail enhancement. We use an encoder-decoder network followed by a fully convolutional network (FCN) as the backbone to extract multi-scale features and reconstruct the High-resolution (HR) results. To exploit the relevance between multi-layer feature maps, we first integrate a convolutional block attention module (CBAM) into each skip-connection of the encoder-decoder subnet, generating weighted maps to enhance both channel-wise and spatial-wise feature representation automatically. Besides, considering that the HR results and LR inputs are highly similar in structure, yet cannot be fully reflected in traditional attention mechanism, we, therefore, designed a self augmented attention (SAA) module, where the attention weights are produced dynamically via a similarity function between hidden features; this design allows the network to flexibly adjust the fraction relevance among multi-layer features and keep the long-range inter information, which is helpful to preserve details. In addition, the pixel-wise loss is combined with perceptual and gradient loss to achieve comprehensive supervision. Experiments on benchmark datasets demonstrate that the proposed method outperforms other SR methods in terms of both objective evaluation and visual effect. Full article
(This article belongs to the Special Issue Intelligent Sensor Signal in Machine Learning)
Show Figures

Figure 1

Figure 1
<p>Structure of standard WGAN.</p>
Full article ">Figure 2
<p>Architecture of self augmented attention (SAA)-WGAN.</p>
Full article ">Figure 3
<p>Attention-augmented results.</p>
Full article ">Figure 4
<p>Channel and spatial attention block (CBAM).</p>
Full article ">Figure 5
<p>Attention-augmented convolution results.</p>
Full article ">Figure 6
<p>Panchromatic (PAN) image super-resolution (SR) from GEO.</p>
Full article ">Figure 7
<p>DOTA image for x4 scale SR.</p>
Full article ">Figure 8
<p>PAN image SR from GEO (scale = 4).</p>
Full article ">Figure 9
<p>SR results from DOTA datasets (scale = 4).</p>
Full article ">Figure 10
<p>The metrics of GEO image (The proposed SAA-WGAN is the purple curve.).</p>
Full article ">
23 pages, 1106 KiB  
Article
Machine Learning Methodology in a System Applying the Adaptive Strategy for Teaching Human Motions
by Krzysztof Wójcik and Marcin Piekarczyk
Sensors 2020, 20(1), 314; https://doi.org/10.3390/s20010314 - 6 Jan 2020
Cited by 13 | Viewed by 3874
Abstract
The teaching of motion activities in rehabilitation, sports, and professional work has great social significance. However, the automatic teaching of these activities, particularly those involving fast motions, requires the use of an adaptive system that can adequately react to the changing stages and [...] Read more.
The teaching of motion activities in rehabilitation, sports, and professional work has great social significance. However, the automatic teaching of these activities, particularly those involving fast motions, requires the use of an adaptive system that can adequately react to the changing stages and conditions of the teaching process. This paper describes a prototype of an automatic system that utilizes the online classification of motion signals to select the proper teaching algorithm. The knowledge necessary to perform the classification process is acquired from experts by the use of the machine learning methodology. The system utilizes multidimensional motion signals that are captured using MEMS (Micro-Electro-Mechanical Systems) sensors. Moreover, an array of vibrotactile actuators is used to provide feedback to the learner. The main goal of the presented article is to prove that the effectiveness of the described teaching system is higher than the system that controls the learning process without the use of signal classification. Statistical tests carried out by the use of a prototype system confirmed that thesis. This is the main outcome of the presented study. An important contribution is also a proposal to standardize the system structure. The standardization facilitates the system configuration and implementation of individual, specialized teaching algorithms. Full article
(This article belongs to the Special Issue Intelligent Sensor Signal in Machine Learning)
Show Figures

Figure 1

Figure 1
<p>The scheme of the learning process with teacher participation. The motion activity of the learner’s body is evaluated by the teacher or automatic controller, which uses an actuator to send feedback to the learner. The object of this process (i.e., the learner) uses the local feedback loop to control his or her movements.</p>
Full article ">Figure 2
<p>General flowchart of the real-time teaching system prototype. The signal that selects the teaching algorithm is symbolized by the thick gray arrow.</p>
Full article ">Figure 3
<p>Simplified flowchart of the kNNModel method and data structures utilized in the implementation of this method.</p>
Full article ">Figure 4
<p>Signals and patterns created from the signals. For example, the red signal in the upper window represents the <span class="html-italic">x</span> component of acceleration. The unit of the vertical axis is 1 m/s<math display="inline"><semantics> <msup> <mrow/> <mn>2</mn> </msup> </semantics></math>, and the time axis is scaled in seconds. The bottom window depicts the time pattern (black) created from the <span class="html-italic">x</span> component of acceleration on the base of several periods depicted in the top. Two shape patterns created from the <span class="html-italic">x</span> and <span class="html-italic">z</span> components of position are also depicted.</p>
Full article ">Figure 5
<p>The multi-dimensional signals and patterns applied in the motion learning system: the current multi-dimensional signal and the collection of multi-dimensional class patterns for the classification process (left side), current multi-dimensional signal and multi-dimensional time pattern, and current multi-dimensional signal and multi-dimensional shape pattern.</p>
Full article ">Figure 6
<p>Peripheral elements of the teaching system. (<b>a</b>) Actuator: the band built on the base of an elastic hook-and-loop strip. (<b>b</b>) Actuator’s units (behind a protective film). (<b>c</b>) VN-100 inertial sensor.</p>
Full article ">Figure 7
<p>A simplified diagram of the signal flow of the class <math display="inline"><semantics> <msub> <mi>C</mi> <mi>α</mi> </msub> </semantics></math> algorithm. The elements of the general system (<a href="#sensors-20-00314-f002" class="html-fig">Figure 2</a>) are depicted in gray.</p>
Full article ">Figure 8
<p>Schematic view of the system elements during the test.</p>
Full article ">Figure 9
<p>Position of the left and right wrists projected on the selected plane (it is parallel to the plane defined by the shoulder blades and tailbone). Continuously changing colors of trajectories are related to time flow; the brightest colors correspond to the latest signal probes. The units on the axes refer to 0.2 m.</p>
Full article ">
26 pages, 11889 KiB  
Article
Robust Stride Detector from Ankle-Mounted Inertial Sensors for Pedestrian Navigation and Activity Recognition with Machine Learning Approaches
by Bertrand Beaufils, Frédéric Chazal, Marc Grelet and Bertrand Michel
Sensors 2019, 19(20), 4491; https://doi.org/10.3390/s19204491 - 16 Oct 2019
Cited by 8 | Viewed by 3844
Abstract
In this paper, a stride detector algorithm combined with a technique inspired by zero velocity update (ZUPT) is proposed to reconstruct the trajectory of a pedestrian from an ankle-mounted inertial device. This innovative approach is based on sensor alignment and machine learning. It [...] Read more.
In this paper, a stride detector algorithm combined with a technique inspired by zero velocity update (ZUPT) is proposed to reconstruct the trajectory of a pedestrian from an ankle-mounted inertial device. This innovative approach is based on sensor alignment and machine learning. It is able to detect 100% of both normal walking strides and more than 97% of atypical strides such as small steps, side steps, and backward walking that existing methods can hardly detect. This approach is also more robust in critical situations, when for example the wearer is sitting and moving the ankle or when the wearer is bicycling (less than two false detected strides per hour on average). As a consequence, the algorithm proposed for trajectory reconstruction achieves much better performances than existing methods for daily life contexts, in particular in narrow areas such as in a house. The computed stride trajectory contains essential information for recognizing the activity (atypical stride, walking, running, and stairs). For this task, we adopt a machine learning approach based on descriptors of these trajectories, which is shown to be robust to a large of variety of gaits. We tested our algorithm on recordings of healthy adults and children, achieving more than 99% success. The algorithm also achieved more than 97% success in challenging situations recorded by children suffering from movement disorders. Compared to most algorithms in the literature, this original method does not use a fixed-size sliding window but infers this last in an adaptive way. Full article
(This article belongs to the Special Issue Intelligent Sensor Signal in Machine Learning)
Show Figures

Figure 1

Figure 1
<p>Our approach for trajectory reconstruction and human activity recognition (HAR) is based on stride detection.</p>
Full article ">Figure 2
<p>(<b>a</b>) WATA (Wearable Ankle Trajectory Analyzer) device worn at the ankle. (<b>b</b>) WATA devices connected to their case</p>
Full article ">Figure 3
<p>Default device placement.</p>
Full article ">Figure 4
<p>Example of a computed pseudo-trajectory in <math display="inline"><semantics> <msup> <mi>B</mi> <mrow> <mi>t</mi> <mi>e</mi> <mi>r</mi> <mi>r</mi> </mrow> </msup> </semantics></math> and <math display="inline"><semantics> <msubsup> <mi>B</mi> <mi>j</mi> <mrow> <mi>t</mi> <mi>e</mi> <mi>r</mi> <mi>r</mi> </mrow> </msubsup> </semantics></math>.</p>
Full article ">Figure 5
<p>Stride length as a function of stride duration during MOCAP sessions.</p>
Full article ">Figure 6
<p>The three foot rockers during stance phase: (<b>a</b>) heel rocker, (<b>b</b>) ankle rocker, (<b>c</b>) forefoot rocker.</p>
Full article ">Figure 7
<p>Stride detection combined with dead reckoning in an extended Kalman filter. INS: inertial navigation system.</p>
Full article ">Figure 8
<p>Computed altitude during (<b>a</b>) the first walking period and (<b>b</b>) the third walking period.</p>
Full article ">Figure 9
<p>Computed trajectory during the first walking period on (<b>a</b>) the ground floor and (<b>b</b>) the first floor.</p>
Full article ">Figure 10
<p>Computed trajectory during the second walking period on the ground floor.</p>
Full article ">Figure 11
<p>Computed trajectory during the third walking period on (<b>a</b>) the ground floor and (<b>b</b>) the first floor.</p>
Full article ">Figure 12
<p>Example of one stairs step progression every half second of patient 5.</p>
Full article ">Figure 13
<p>Activity recognition (AR) results associated with the distribution of (<b>a</b>) the strides’ length/duration and (<b>b</b>) the strides’ speed.</p>
Full article ">Figure 14
<p>The four main algorithm stages with machine learning uses (red).</p>
Full article ">Figure A1
<p>Pseudo-speed norm during walking.</p>
Full article ">Figure A2
<p>Pseudo-speed norm during fast side stepping.</p>
Full article ">
17 pages, 1812 KiB  
Article
Fast Depth Estimation in a Single Image Using Lightweight Efficient Neural Network
by Sangwon Kim, Jaeyeal Nam and Byoungchul Ko
Sensors 2019, 19(20), 4434; https://doi.org/10.3390/s19204434 - 13 Oct 2019
Cited by 4 | Viewed by 7027
Abstract
Depth estimation is a crucial and fundamental problem in the computer vision field. Conventional methods re-construct scenes using feature points extracted from multiple images; however, these approaches require multiple images and thus are not easily implemented in various real-time applications. Moreover, the special [...] Read more.
Depth estimation is a crucial and fundamental problem in the computer vision field. Conventional methods re-construct scenes using feature points extracted from multiple images; however, these approaches require multiple images and thus are not easily implemented in various real-time applications. Moreover, the special equipment required by hardware-based approaches using 3D sensors is expensive. Therefore, software-based methods for estimating depth from a single image using machine learning or deep learning are emerging as new alternatives. In this paper, we propose an algorithm that generates a depth map in real time using a single image and an optimized lightweight efficient neural network (L-ENet) algorithm instead of physical equipment, such as an infrared sensor or multi-view camera. Because depth values have a continuous nature and can produce locally ambiguous results, pixel-wise prediction with ordinal depth range classification was applied in this study. In addition, in our method various convolution techniques are applied to extract a dense feature map, and the number of parameters is greatly reduced by reducing the network layer. By using the proposed L-ENet algorithm, an accurate depth map can be generated from a single image quickly and, in a comparison with the ground truth, we can produce depth values closer to those of the ground truth with small errors. Experiments confirmed that the proposed L-ENet can achieve a significantly improved estimation performance over the state-of-the-art algorithms in depth estimation based on a single image. Full article
(This article belongs to the Special Issue Intelligent Sensor Signal in Machine Learning)
Show Figures

Figure 1

Figure 1
<p>Lightweight efficient neural network architecture. The output size consists of three tuples: number of feature maps, width of a feature map and height of a feature map (142×513×385). The bold black line indicates the newly proposed network structure.</p>
Full article ">Figure 2
<p>Operation structure of lightweight efficient neural network initial block and bottleneck module. (<b>a</b>) Initial module, (<b>b</b>) bottleneck module of downsampling operation, and (<b>c</b>) bottleneck module of dilated and asymmetric convolution.</p>
Full article ">Figure 3
<p>Comparison of a conventional and an asymmetric convolution. (<b>a</b>) General conventional convolution operation and (<b>b</b>) asymmetric convolution operation.</p>
Full article ">Figure 4
<p>Several convolution methods. (<b>a</b>) general convolution, (<b>b</b>) depth-wise convolution and (<b>c</b>) point-wise convolution.</p>
Full article ">Figure 5
<p>Discretize depth interval [<math display="inline"><semantics> <mrow> <mi>a</mi> <mo>,</mo> <mo> </mo> <mi>b</mi> </mrow> </semantics></math>] into <math display="inline"><semantics> <mrow> <mi>D</mi> <mi>p</mi> <mo>+</mo> <mn>1</mn> </mrow> </semantics></math> subintervals. (<b>a</b>) uniformly discretized depth interval and (<b>b</b>) discretized depth interval in log space. As the <math display="inline"><semantics> <mrow> <msub> <mi>l</mi> <mi>i</mi> </msub> </mrow> </semantics></math> label approaches the minimum value <span class="html-italic">a</span>, a dense value is reassigned to <math display="inline"><semantics> <mrow> <msubsup> <mi>l</mi> <mi>i</mi> <mo>′</mo> </msubsup> </mrow> </semantics></math>. In contrast, as the <math display="inline"><semantics> <mrow> <msub> <mi>l</mi> <mi>i</mi> </msub> </mrow> </semantics></math> label approaches the maximum value <math display="inline"><semantics> <mi>b</mi> </semantics></math>, a coarse value is reassigned to <math display="inline"><semantics> <mrow> <msubsup> <mi>l</mi> <mi>i</mi> <mo>′</mo> </msubsup> </mrow> </semantics></math>.</p>
Full article ">Figure 6
<p>Qualitative comparison of estimated depth-maps on KITTI. (<b>a</b>) Input images, (<b>b</b>) ground truth, (<b>c</b>) multi-scale convolutional architecture [<a href="#B11-sensors-19-04434" class="html-bibr">11</a>], (<b>d</b>) deep ordinal regression network [<a href="#B14-sensors-19-04434" class="html-bibr">14</a>], (<b>e</b>) with regression of [<a href="#B11-sensors-19-04434" class="html-bibr">11</a>], and (<b>f</b>) proposed lightweight efficient neural network.</p>
Full article ">
27 pages, 3738 KiB  
Article
PEnBayes: A Multi-Layered Ensemble Approach for Learning Bayesian Network Structure from Big Data
by Yan Tang, Jianwu Wang, Mai Nguyen and Ilkay Altintas
Sensors 2019, 19(20), 4400; https://doi.org/10.3390/s19204400 - 11 Oct 2019
Cited by 8 | Viewed by 4562
Abstract
Discovering the Bayesian network (BN) structure from big datasets containing rich causal relationships is becoming increasingly valuable for modeling and reasoning under uncertainties in many areas with big data gathered from sensors due to high volume and fast veracity. Most of the current [...] Read more.
Discovering the Bayesian network (BN) structure from big datasets containing rich causal relationships is becoming increasingly valuable for modeling and reasoning under uncertainties in many areas with big data gathered from sensors due to high volume and fast veracity. Most of the current BN structure learning algorithms have shortcomings facing big data. First, learning a BN structure from the entire big dataset is an expensive task which often ends in failure due to memory constraints. Second, it is quite difficult to select a learner from numerous BN structure learning algorithms to consistently achieve good learning accuracy. Lastly, there is a lack of an intelligent method that merges separately learned BN structures into a well structured BN network. To address these shortcomings, we introduce a novel parallel learning approach called PEnBayes (Parallel Ensemble-based Bayesian network learning). PEnBayes starts with an adaptive data preprocessing phase that calculates the Appropriate Learning Size and intelligently divides a big dataset for fast distributed local structure learning. Then, PEnBayes learns a collection of local BN Structures in parallel using a two-layered weighted adjacent matrix-based structure ensemble method. Lastly, PEnBayes merges the local BN Structures into a global network structure using the structure ensemble method at the global layer. For the experiment, we generate big data sets by simulating sensor data from patient monitoring, transportation, and disease diagnosis domains. The Experimental results show that PEnBayes achieves a significantly improved execution performance with more consistent and stable results compared with three baseline learning algorithms. Full article
(This article belongs to the Special Issue Intelligent Sensor Signal in Machine Learning)
Show Figures

Figure 1

Figure 1
<p>A Bayesian network example - Cancer Network.</p>
Full article ">Figure 2
<p>An Overview of PEnBayes approach.</p>
Full article ">Figure 3
<p>Full PEnBayes Workflow in Kepler.</p>
Full article ">Figure 4
<p>Data Preprocessing Sub-workflow.</p>
Full article ">Figure 5
<p>Local Learner Sub-workflow.</p>
Full article ">Figure 6
<p>Global Ensemble Sub-workflow.</p>
Full article ">Figure 7
<p>Structural Hamming distance (SHD) of different data set sizes using calculated <math display="inline"><semantics> <mrow> <mi>A</mi> <mi>L</mi> <mi>S</mi> </mrow> </semantics></math> (<a href="#sensors-19-04400-t003" class="html-table">Table 3</a>) as reference value (red circle), Child Dataset.</p>
Full article ">Figure 8
<p>Structural Hamming distance (SHD) of different data set sizes using calculated <math display="inline"><semantics> <mrow> <mi>A</mi> <mi>L</mi> <mi>S</mi> </mrow> </semantics></math> (<a href="#sensors-19-04400-t003" class="html-table">Table 3</a>) as reference value (red circle), Insurance Dataset.</p>
Full article ">Figure 9
<p>Structural Hamming distance (SHD) of different data set sizes using calculated <math display="inline"><semantics> <mrow> <mi>A</mi> <mi>L</mi> <mi>S</mi> </mrow> </semantics></math> (<a href="#sensors-19-04400-t003" class="html-table">Table 3</a>) as reference value (red circle), Alarm Dataset.</p>
Full article ">Figure 10
<p>Alarm Set Execution Time.</p>
Full article ">Figure 11
<p>Child Set Execution Time.</p>
Full article ">Figure 12
<p>Insurance Set Execution Time.</p>
Full article ">Figure 13
<p>Alarm Dataset Accuracy Results. Negative values indicate that the algorithm was unsuccessful in learning a network for the dataset.</p>
Full article ">Figure 14
<p>Child Dataset Accuracy Results.</p>
Full article ">Figure 15
<p>Insurance Dataset Accuracy Results.</p>
Full article ">Figure 16
<p>Alarm Set Standard Deviation Results.</p>
Full article ">Figure 17
<p>Alarm Set Execution Time vs. Number of Local Learners.</p>
Full article ">Figure 18
<p>Child Set Execution Time vs. Number of Local Learners.</p>
Full article ">Figure 19
<p>Insurance Set Execution Time vs. Number of Local Learners.</p>
Full article ">
19 pages, 1571 KiB  
Article
Machine Learning for LTE Energy Detection Performance Improvement
by Małgorzata Wasilewska and Hanna Bogucka
Sensors 2019, 19(19), 4348; https://doi.org/10.3390/s19194348 - 8 Oct 2019
Cited by 18 | Viewed by 3940
Abstract
The growing number of radio communication devices and limited spectrum resources are drivers for the development of new techniques of dynamic spectrum access and spectrum sharing. In order to make use of the spectrum opportunistically, the concept of cognitive radio was proposed, where [...] Read more.
The growing number of radio communication devices and limited spectrum resources are drivers for the development of new techniques of dynamic spectrum access and spectrum sharing. In order to make use of the spectrum opportunistically, the concept of cognitive radio was proposed, where intelligent decisions on transmission opportunities are based on spectrum sensing. In this paper, two Machine Learning (ML) algorithms, namely k-Nearest Neighbours and Random Forest, have been proposed to increase spectrum sensing performance. These algorithms have been applied to Energy Detection (ED) and Energy Vector-based data (EV) to detect the presence of a Fourth Generation (4G) Long-Term Evolution (LTE) signal for the purpose of utilizing the available resource blocks by a 5G new radio system. The algorithms capitalize on time, frequency and spatial dependencies in daily communication traffic. Research results show that the ML methods used can significantly improve the spectrum sensing performance if the input training data set is carefully chosen. The input data sets with ED decisions and energy values have been examined, and advantages and disadvantages of their real-life application have been analyzed. Full article
(This article belongs to the Special Issue Intelligent Sensor Signal in Machine Learning)
Show Figures

Figure 1

Figure 1
<p>k-Nearest Neighbors—visualization of the closest data points for different <span class="html-italic">k</span> values.</p>
Full article ">Figure 2
<p>Decision tree—tree with depth 3.</p>
Full article ">Figure 3
<p>System model.</p>
Full article ">Figure 4
<p>LTE Resource Blocks features.</p>
Full article ">Figure 5
<p>Transmitted LTE Resource Blocks.</p>
Full article ">Figure 6
<p>Probability of detection <math display="inline"><semantics> <msub> <mi>P</mi> <mi mathvariant="normal">d</mi> </msub> </semantics></math> for the Energy Detection stage for <math display="inline"><semantics> <mrow> <msub> <mover accent="true"> <mi>P</mi> <mo stretchy="false">¯</mo> </mover> <mi>fa</mi> </msub> <mo>=</mo> <mn>10</mn> <mo>%</mo> </mrow> </semantics></math>, <math display="inline"><semantics> <mrow> <msub> <mover accent="true"> <mi>P</mi> <mo stretchy="false">¯</mo> </mover> <mi>fa</mi> </msub> <mo>=</mo> <mn>2</mn> <mo>%</mo> </mrow> </semantics></math> and <math display="inline"><semantics> <mrow> <msub> <mover accent="true"> <mi>P</mi> <mo stretchy="false">¯</mo> </mover> <mi>fa</mi> </msub> <mo>=</mo> <mn>0.5</mn> <mo>%</mo> </mrow> </semantics></math>.</p>
Full article ">Figure 7
<p>Resulting probability of detection <math display="inline"><semantics> <msub> <mi>P</mi> <mi mathvariant="normal">d</mi> </msub> </semantics></math> and probability of false alarm <math display="inline"><semantics> <msub> <mi>P</mi> <mi>fa</mi> </msub> </semantics></math> of the Energy Detection-based k-Nearest Neighbors method for <math display="inline"><semantics> <mrow> <msub> <mover accent="true"> <mi>P</mi> <mo stretchy="false">¯</mo> </mover> <mi>fa</mi> </msub> <mo>=</mo> <mn>10</mn> <mo>%</mo> </mrow> </semantics></math>.</p>
Full article ">Figure 8
<p>Resulting probability of detection <math display="inline"><semantics> <msub> <mi>P</mi> <mi mathvariant="normal">d</mi> </msub> </semantics></math> and probability of false alarm <math display="inline"><semantics> <msub> <mi>P</mi> <mi>fa</mi> </msub> </semantics></math> of the Energy Detection-based Random Forest method for <math display="inline"><semantics> <mrow> <msub> <mover accent="true"> <mi>P</mi> <mo stretchy="false">¯</mo> </mover> <mi>fa</mi> </msub> <mo>=</mo> <mn>10</mn> <mo>%</mo> </mrow> </semantics></math>.</p>
Full article ">Figure 9
<p>Resulting probability of detection <math display="inline"><semantics> <msub> <mi>P</mi> <mi mathvariant="normal">d</mi> </msub> </semantics></math> and probability of false alarm <math display="inline"><semantics> <msub> <mi>P</mi> <mi>fa</mi> </msub> </semantics></math> of the Energy Vector-based k-Nearest Neighbors method for <math display="inline"><semantics> <mrow> <msub> <mover accent="true"> <mi>P</mi> <mo stretchy="false">¯</mo> </mover> <mi>fa</mi> </msub> <mo>=</mo> <mn>10</mn> <mo>%</mo> </mrow> </semantics></math>.</p>
Full article ">Figure 10
<p>Resulting probability of detection <math display="inline"><semantics> <msub> <mi>P</mi> <mi mathvariant="normal">d</mi> </msub> </semantics></math> and probability of false alarm <math display="inline"><semantics> <msub> <mi>P</mi> <mi>fa</mi> </msub> </semantics></math> of the Energy Vector-based Random Forest method for <math display="inline"><semantics> <mrow> <msub> <mover accent="true"> <mi>P</mi> <mo stretchy="false">¯</mo> </mover> <mi>fa</mi> </msub> <mo>=</mo> <mn>10</mn> <mo>%</mo> </mrow> </semantics></math>.</p>
Full article ">Figure 11
<p>Probability of detection <math display="inline"><semantics> <msub> <mi>P</mi> <mi mathvariant="normal">d</mi> </msub> </semantics></math> comparison of the Energy Detection-based and Energy Vector-based k-Nearest Neighbors and Random Forest methods for <math display="inline"><semantics> <mrow> <msub> <mover accent="true"> <mi>P</mi> <mo stretchy="false">¯</mo> </mover> <mi>fa</mi> </msub> <mo>=</mo> <mn>10</mn> <mo>%</mo> </mrow> </semantics></math>.</p>
Full article ">Figure 12
<p>Probability of detection <math display="inline"><semantics> <msub> <mi>P</mi> <mi mathvariant="normal">d</mi> </msub> </semantics></math> comparison of the Energy Detection-based and Energy Vector-based k-Nearest Neighbors and Random Forest methods for different assumed <math display="inline"><semantics> <mrow> <msub> <mover accent="true"> <mi>P</mi> <mo stretchy="false">¯</mo> </mover> <mi>fa</mi> </msub> </mrow> </semantics></math>.</p>
Full article ">Figure 13
<p>SNR values resulting from the shadowing effect in the considered area.</p>
Full article ">Figure 14
<p><math display="inline"><semantics> <msub> <mi>P</mi> <mi mathvariant="normal">d</mi> </msub> </semantics></math> and <math display="inline"><semantics> <msub> <mi>P</mi> <mi>fa</mi> </msub> </semantics></math> for different locations.</p>
Full article ">Figure 15
<p><math display="inline"><semantics> <msub> <mi>P</mi> <mi mathvariant="normal">d</mi> </msub> </semantics></math> and <math display="inline"><semantics> <msub> <mi>P</mi> <mi>fa</mi> </msub> </semantics></math> for the k-Nearest Neighbors method applied in different locations. (<b>a</b>) <math display="inline"><semantics> <msub> <mi>P</mi> <mi mathvariant="normal">d</mi> </msub> </semantics></math> and <math display="inline"><semantics> <msub> <mi>P</mi> <mi>fa</mi> </msub> </semantics></math> surfaces for Energy Detection-based k-Nearest Neighbors; (<b>b</b>) <math display="inline"><semantics> <msub> <mi>P</mi> <mi mathvariant="normal">d</mi> </msub> </semantics></math> and <math display="inline"><semantics> <msub> <mi>P</mi> <mi>fa</mi> </msub> </semantics></math> surfaces for Energy Vector-based k-Nearest Neighbors.</p>
Full article ">Figure 16
<p>Resulting probabilities <math display="inline"><semantics> <msub> <mi>P</mi> <mi mathvariant="normal">d</mi> </msub> </semantics></math> and <math display="inline"><semantics> <msub> <mi>P</mi> <mi>fa</mi> </msub> </semantics></math> of Energy Detection-based k-Nearest Neighbors compared with Energy Vector-based k-Nearest Neighbors for <math display="inline"><semantics> <mrow> <msub> <mover accent="true"> <mi>P</mi> <mo stretchy="false">¯</mo> </mover> <mi>fa</mi> </msub> <mo>=</mo> <mn>10</mn> <mo>%</mo> </mrow> </semantics></math> for different SNR values with a shadowing channel.</p>
Full article ">Figure 17
<p>Probabilities <math display="inline"><semantics> <msub> <mi>P</mi> <mi mathvariant="normal">d</mi> </msub> </semantics></math> and <math display="inline"><semantics> <msub> <mi>P</mi> <mi>fa</mi> </msub> </semantics></math> in different locations for the applied Random Forest method. (<b>a</b>) <math display="inline"><semantics> <msub> <mi>P</mi> <mi mathvariant="normal">d</mi> </msub> </semantics></math> and <math display="inline"><semantics> <msub> <mi>P</mi> <mi>fa</mi> </msub> </semantics></math> surfaces for Energy Detection-based Random Forest; (<b>b</b>) <math display="inline"><semantics> <msub> <mi>P</mi> <mi mathvariant="normal">d</mi> </msub> </semantics></math> and <math display="inline"><semantics> <msub> <mi>P</mi> <mi>fa</mi> </msub> </semantics></math> surfaces for Energy Vector-based Random Forest.</p>
Full article ">Figure 18
<p>Resulting probabilities <math display="inline"><semantics> <msub> <mi>P</mi> <mi mathvariant="normal">d</mi> </msub> </semantics></math> and <math display="inline"><semantics> <msub> <mi>P</mi> <mi>fa</mi> </msub> </semantics></math> of applied Energy Detection-based Random Forest compared with Energy Vector-based Random Forest for <math display="inline"><semantics> <mrow> <msub> <mover accent="true"> <mi>P</mi> <mo stretchy="false">¯</mo> </mover> <mi>fa</mi> </msub> <mo>=</mo> <mn>10</mn> <mo>%</mo> </mrow> </semantics></math> for different SNR values with a shadowing channel.</p>
Full article ">Figure 19
<p>Probabilities <math display="inline"><semantics> <msub> <mi>P</mi> <mi mathvariant="normal">d</mi> </msub> </semantics></math> and <math display="inline"><semantics> <msub> <mi>P</mi> <mi>fa</mi> </msub> </semantics></math> in different locations for the applied k-Nearest Neighbors, Random Forest, Gaussian Naive Bayes and Support Vector Machine classifier methods. (<b>a</b>) <math display="inline"><semantics> <msub> <mi>P</mi> <mi mathvariant="normal">d</mi> </msub> </semantics></math> and <math display="inline"><semantics> <msub> <mi>P</mi> <mi>fa</mi> </msub> </semantics></math> surfaces for Energy Detection-based Machine Learning algorithms; (<b>b</b>) <math display="inline"><semantics> <msub> <mi>P</mi> <mi mathvariant="normal">d</mi> </msub> </semantics></math> and <math display="inline"><semantics> <msub> <mi>P</mi> <mi>fa</mi> </msub> </semantics></math> surfaces for Energy Vector-based Machine Learning algorithms.</p>
Full article ">
23 pages, 1419 KiB  
Article
An FPGA-Based Neuro-Fuzzy Sensor for Personalized Driving Assistance
by Óscar Mata-Carballeira, Jon Gutiérrez-Zaballa, Inés del Campo and Victoria Martínez
Sensors 2019, 19(18), 4011; https://doi.org/10.3390/s19184011 - 17 Sep 2019
Cited by 15 | Viewed by 4816
Abstract
Advanced driving-assistance systems (ADAS) are intended to automatize driver tasks, as well as improve driving and vehicle safety. This work proposes an intelligent neuro-fuzzy sensor for driving style (DS) recognition, suitable for ADAS enhancement. The development of the driving style intelligent sensor uses [...] Read more.
Advanced driving-assistance systems (ADAS) are intended to automatize driver tasks, as well as improve driving and vehicle safety. This work proposes an intelligent neuro-fuzzy sensor for driving style (DS) recognition, suitable for ADAS enhancement. The development of the driving style intelligent sensor uses naturalistic driving data from the SHRP2 study, which includes data from a CAN bus, inertial measurement unit, and front radar. The system has been successfully implemented using a field-programmable gate array (FPGA) device of the Xilinx Zynq programmable system-on-chip (PSoC). It can mimic the typical timing parameters of a group of drivers as well as tune these typical parameters to model individual DSs. The neuro-fuzzy intelligent sensor provides high-speed real-time active ADAS implementation and is able to personalize its behavior into safe margins without driver intervention. In particular, the personalization procedure of the time headway (THW) parameter for an ACC in steady car following was developed, achieving a performance of 0.53 microseconds. This performance fulfilled the requirements of cutting-edge active ADAS specifications. Full article
(This article belongs to the Special Issue Intelligent Sensor Signal in Machine Learning)
Show Figures

Figure 1

Figure 1
<p>Offline sequence of tasks involved in the design and development of a neuro-fuzzy sensor for advanced driving-assistance system (ADAS) personalization.</p>
Full article ">Figure 2
<p>Block diagram of the field-programmable gate array (FPGA)-based intelligent sensor for online car-following ADAS.</p>
Full article ">Figure 3
<p>Data acquisition systems and sensors installed in the vehicles that participated in the SHRP2-NDS. IR: infrared; SW: software.</p>
Full article ">Figure 4
<p>Representative example of car-following features: (<b>a</b>) time-exposed time headway (TETH) and (<b>b</b>) time-integrated time headway (TITH).</p>
Full article ">Figure 5
<p>Clusters obtained applying the k-means algorithm to the car-following segments; <math display="inline"><semantics> <msub> <mi>THW</mi> <mi>RMS</mi> </msub> </semantics></math>, TETH, and TITH values were normalized.</p>
Full article ">Figure 6
<p>Histogram of <math display="inline"><semantics> <msub> <mi>THW</mi> <mi>RMS</mi> </msub> </semantics></math>, TETH, and TITH values distribution for (<b>a</b>) Cluster 1, (<b>b</b>) Cluster 2, and (<b>c</b>) Cluster 3. Red line represents average value of each distribution.</p>
Full article ">Figure 7
<p>(<b>a</b>) Adaptive neuro-fuzzy inference system (ANFIS) 1 (Cluster 1), (<b>b</b>) ANFIS 2 (Cluster 2), and (<b>c</b>) ANFIS 3 (Cluster 3). Training data shown in red; response of corresponding trained ANFIS shown in blue.</p>
Full article ">Figure 8
<p>Block scheme of the parallel architecture of a three-input ANFIS implemented in the programmable logic (PL) of the programmable system-on-chip (PSoC). Three ANFIS cores, one per cluster, were implemented in the HW partition.</p>
Full article ">Figure 9
<p>Scheme of proposed sum of product architecture that substitutes traditional tree-adder solution.</p>
Full article ">Figure 10
<p>Chronogram of control-signal sequence of the ANFIS core.</p>
Full article ">Figure 11
<p>Rules and membership functions of ANFIS Cluster 1 for a given input.</p>
Full article ">Figure 12
<p>Simulation results of the ANFIS Cluster 1 HW accelerator obtained with the Vivado design suite.</p>
Full article ">Figure 13
<p><math display="inline"><semantics> <msub> <mover accent="true"> <mrow> <mi>T</mi> <mi>H</mi> <mi>W</mi> </mrow> <mo>^</mo> </mover> <mi>i</mi> </msub> </semantics></math> model planes for (<b>a</b>) Cluster 1, (<b>b</b>) Cluster 2, and (<b>c</b>) Cluster 3.</p>
Full article ">
21 pages, 1306 KiB  
Article
Pedestrian Positioning Using a Double-Stacked Particle Filter in Indoor Wireless Networks
by Kwangjae Sung, Hyung Kyu Lee and Hwangnam Kim
Sensors 2019, 19(18), 3907; https://doi.org/10.3390/s19183907 - 10 Sep 2019
Cited by 7 | Viewed by 2957
Abstract
The indoor pedestrian positioning methods are affected by substantial bias and errors because of the use of cheap microelectromechanical systems (MEMS) devices (e.g., gyroscope and accelerometer) and the users’ movements. Moreover, because radio-frequency (RF) signal values are changed drastically due to multipath fading [...] Read more.
The indoor pedestrian positioning methods are affected by substantial bias and errors because of the use of cheap microelectromechanical systems (MEMS) devices (e.g., gyroscope and accelerometer) and the users’ movements. Moreover, because radio-frequency (RF) signal values are changed drastically due to multipath fading and obstruction, the performance of RF-based localization systems may deteriorate in practice. To deal with this problem, various indoor localization methods that integrate the positional information gained from received signal strength (RSS) fingerprinting scheme and the motion of the user inferred by dead reckoning (DR) approach via Bayes filters have been suggested to accomplish more accurate localization results indoors. Among the Bayes filters, while the particle filter (PF) can offer the most accurate positioning performance, it may require substantial computation time due to use of many samples (particles) for high positioning accuracy. This paper introduces a pedestrian localization scheme performed on a mobile phone that leverages the RSS fingerprint-based method, dead reckoning (DR), and improved PF called a double-stacked particle filter (DSPF) in indoor environments. As a key element of our system, the DSPF algorithm is employed to correct the position of the user by fusing noisy location data gained by the RSS fingerprinting and DR schemes. By estimating the position of the user through the proposal distribution and target distribution obtained from multiple measurements, the DSPF method can offer better localization results compared to the Kalman filtering-based methods, and it can achieve competitive localization accuracy compared with PF while offering higher computational efficiency than PF. Experimental results demonstrate that the DSPF algorithm can achieve accurate and reliable localization with higher efficiency in computational cost compared with PF in indoor environments. Full article
(This article belongs to the Special Issue Intelligent Sensor Signal in Machine Learning)
Show Figures

Figure 1

Figure 1
<p>Localization system architecture.</p>
Full article ">Figure 2
<p>User motion model that takes into account the user’s <span class="html-italic">x</span> (east) and <span class="html-italic">y</span> (north) coordinates and heading angles at timesteps <span class="html-italic">k</span> and <math display="inline"><semantics> <mrow> <mi>k</mi> <mo>−</mo> <mn>1</mn> </mrow> </semantics></math>.</p>
Full article ">Figure 3
<p>DSPF algorithm with (<b>a</b>) target and (<b>b</b>) proposal distributions that consist of particles with <span class="html-italic">x</span>-axis and <span class="html-italic">y</span>-axis coordinate values.</p>
Full article ">Figure 4
<p>Floor map for experimental testbeds. RSS fingerprints and user’s direction information are collected in both the hallways and room.</p>
Full article ">Figure 5
<p>Mean packet success rate (PSR) computed using pedestrians’ smartphones in test positions (orange circles in <a href="#sensors-19-03907-f004" class="html-fig">Figure 4</a>) of test case TC2.</p>
Full article ">Figure 6
<p>Positioning error of the localization schemes for test cases (<b>a</b>) TC1 and (<b>b</b>) TC2.</p>
Full article ">Figure 7
<p>Mean positioning error for the localization scheme in which the Bayes filters are performed by DRC3 at each physical location (orange circle and green square in <a href="#sensors-19-03907-f004" class="html-fig">Figure 4</a>) of test cases (<b>a</b>) TC1 and (<b>b</b>) TC2.</p>
Full article ">Figure 8
<p>Pedestrian trajectories calculated by the positioning scheme in which the Bayes filters are performed by DRC3 when striding on the physical positions (orange circles and green squares in <a href="#sensors-19-03907-f004" class="html-fig">Figure 4</a>) of test cases (<b>a</b>) TC1 and (<b>b</b>) TC2 clockwise. The trajectories are indicated with <span class="html-italic">x</span> (east) and <span class="html-italic">y</span> (north) coordinates.</p>
Full article ">Figure 9
<p>Positioning experiment results of test cases TC1 and TC2 executed by positioning scheme DRC3 using DSPF for the values between <math display="inline"><semantics> <mrow> <msub> <mi>N</mi> <mrow> <mi>p</mi> <mi>d</mi> <mi>p</mi> </mrow> </msub> <mo>⊂</mo> <mrow> <mo>{</mo> <mn>10</mn> <mo>,</mo> <mn>140</mn> <mo>}</mo> </mrow> </mrow> </semantics></math> and <math display="inline"><semantics> <mrow> <msub> <mi>N</mi> <mrow> <mi>t</mi> <mi>d</mi> <mi>p</mi> </mrow> </msub> <mo>⊂</mo> <mrow> <mo>{</mo> <msup> <mrow> <mn>10</mn> </mrow> <mn>1</mn> </msup> <mo>,</mo> <msup> <mrow> <mn>10</mn> </mrow> <mn>6</mn> </msup> <mo>}</mo> </mrow> </mrow> </semantics></math>; (<b>a</b>) Curve of the average positioning error and (<b>b</b>) Curve of the average positioning time. Text annotations on the curves indicate optimal values of <math display="inline"><semantics> <mrow> <msub> <mi>N</mi> <mrow> <mi>p</mi> <mi>d</mi> <mi>p</mi> </mrow> </msub> <mo>=</mo> <msup> <mrow> <mn>10</mn> </mrow> <mn>2</mn> </msup> </mrow> </semantics></math> and <math display="inline"><semantics> <mrow> <msub> <mi>N</mi> <mrow> <mi>t</mi> <mi>d</mi> <mi>p</mi> </mrow> </msub> <mo>=</mo> <msup> <mrow> <mn>10</mn> </mrow> <mn>5</mn> </msup> </mrow> </semantics></math> that can achieve a high degree of positioning accuracy at little computational cost in TC1 and TC2.</p>
Full article ">Figure 10
<p>Mean computation time of localization approaches.</p>
Full article ">
17 pages, 1239 KiB  
Article
Time Coherent Full-Body Poses Estimated Using Only Five Inertial Sensors: Deep versus Shallow Learning
by Frank J. Wouda, Matteo Giuberti, Nina Rudigkeit, Bert-Jan F. van Beijnum, Mannes Poel and Peter H. Veltink
Sensors 2019, 19(17), 3716; https://doi.org/10.3390/s19173716 - 27 Aug 2019
Cited by 12 | Viewed by 4763
Abstract
Full-body motion capture typically requires sensors/markers to be placed on each rigid body segment, which results in long setup times and is obtrusive. The number of sensors/markers can be reduced using deep learning or offline methods. However, this requires large training datasets and/or [...] Read more.
Full-body motion capture typically requires sensors/markers to be placed on each rigid body segment, which results in long setup times and is obtrusive. The number of sensors/markers can be reduced using deep learning or offline methods. However, this requires large training datasets and/or sufficient computational resources. Therefore, we investigate the following research question: “What is the performance of a shallow approach, compared to a deep learning one, for estimating time coherent full-body poses using only five inertial sensors?”. We propose to incorporate past/future inertial sensor information into a stacked input vector, which is fed to a shallow neural network for estimating full-body poses. Shallow and deep learning approaches are compared using the same input vector configurations. Additionally, the inclusion of acceleration input is evaluated. The results show that a shallow learning approach can estimate full-body poses with a similar accuracy (~6 cm) to that of a deep learning approach (~7 cm). However, the jerk errors are smaller using the deep learning approach, which can be the effect of explicit recurrent modelling. Furthermore, it is shown that the delay using a shallow learning approach (72 ms) is smaller than that of a deep learning approach (117 ms). Full article
(This article belongs to the Special Issue Intelligent Sensor Signal in Machine Learning)
Show Figures

Figure 1

Figure 1
<p>A sequence of inputs (lower arm (orange circle) /leg (green circle) orientations relative to the pelvis (blue circle) is used to estimate a single output pose (at time <span class="html-italic">i</span>). Size of the input sequence can vary by the number of past (<span class="html-italic">P</span>) and future (<span class="html-italic">F</span>) poses that are taken into account and the distance in time (<math display="inline"><semantics> <mrow> <mo mathvariant="normal">Δ</mo> <mi>t</mi> </mrow> </semantics></math>) between the different inputs. Here, <math display="inline"><semantics> <msub> <mi>j</mi> <mi>p</mi> </msub> </semantics></math> and <math display="inline"><semantics> <msub> <mi>j</mi> <mi>f</mi> </msub> </semantics></math> are used as counters for the past/future poses in time, which have a maximum value of P and F (in this example, <math display="inline"><semantics> <mrow> <mi>P</mi> <mo>=</mo> <mn>2</mn> </mrow> </semantics></math> and <math display="inline"><semantics> <mrow> <mi>F</mi> <mo>=</mo> <mn>2</mn> </mrow> </semantics></math>), respectively. In other words, <math display="inline"><semantics> <mrow> <msub> <mi>j</mi> <mi>p</mi> </msub> <mo>=</mo> <mrow> <mo>{</mo> <mn>1</mn> <mo>,</mo> <mo>.</mo> <mo>.</mo> <mo>.</mo> <mo>,</mo> <mi>P</mi> <mo>}</mo> </mrow> </mrow> </semantics></math> and <math display="inline"><semantics> <mrow> <msub> <mi>j</mi> <mi>f</mi> </msub> <mo>=</mo> <mrow> <mo>{</mo> <mn>1</mn> <mo>,</mo> <mo>.</mo> <mo>.</mo> <mo>.</mo> <mo>,</mo> <mi>F</mi> <mo>}</mo> </mrow> </mrow> </semantics></math>, and <math display="inline"><semantics> <mrow> <mo mathvariant="normal">Δ</mo> <mi>t</mi> </mrow> </semantics></math> is defined as <math display="inline"><semantics> <mrow> <mi>I</mi> <mo>/</mo> <msub> <mi>f</mi> <mi>s</mi> </msub> </mrow> </semantics></math> with <span class="html-italic">I</span> as the sample interval and <math display="inline"><semantics> <msub> <mi>f</mi> <mi>s</mi> </msub> </semantics></math> as the sampling frequency.</p>
Full article ">Figure 2
<p>Processing of the measured sensor accelerations to be suitable input to the recurrent neural network (RNN) and stacked input neural network (SINN).</p>
Full article ">Figure 3
<p>The implemented network architectures for the deep (RNN) and shallow (SINN) learning approaches. Different networks were trained for estimating upper/lower-body poses, which resulted in 8 inputs (2 segment orientations, represented by quaternions, relative to the pelvis) times SIL (stacked input length) poses for both SINN and RNN. Furthermore, a different number of outputs was obtained from the separate networks, namely, 24 for the lower body and 48 for the upper body. Input to the SINN is stacked with adjacent poses from past (P), current (<math display="inline"><semantics> <msub> <mi>t</mi> <mi>i</mi> </msub> </semantics></math>) and future (F) time samples, resulting in a total of L samples that are taken into account. The same sequence can be provided as an input matrix to the RNN, which produces a sequence as output (and the relevant pose can be used). The different types of hidden layers are shown by the various colours with the corresponding number of neurons shown in brackets.</p>
Full article ">Figure 4
<p>Bar plots of the full-body mean (of 6 subjects) joint position (<b>A</b>) and jerk (<b>B</b>) error for the shallow (SINN, left bars for each configuration) and deep (RNN, right bars for each configuration) learning approaches during a gait trial (1 in <a href="#sensors-19-03716-t001" class="html-table">Table 1</a>), standard deviation over the various subjects is displayed by whiskers. The different time windows are presented on the <span class="html-italic">x</span>-axis, where the number of past (<span class="html-italic">P</span>) and future (<span class="html-italic">F</span>) poses are shown. The interval (<span class="html-italic">I</span>) between input poses are marked by the different colours, where the number of samples between input poses is shown. For comparison, the mean joint position error (<b>A</b>) for using only the current pose as input (SINN approach) is <math display="inline"><semantics> <mrow> <mn>0.07</mn> </mrow> </semantics></math> (<math display="inline"><semantics> <mrow> <mo>±</mo> <mn>0.01</mn> </mrow> </semantics></math>) m. For comparison, the baseline using only the current pose as input) mean joint position (<b>A</b>) and jerk (<b>B</b>) errors are shown as the dark blue bars on the left.</p>
Full article ">Figure 5
<p>Bar plots of the full-body mean (of 6 subjects) joint position (<b>A</b>) and jerk (<b>B</b>) error for the shallow (SINN, left bars for each configuration ) and deep (RNN, right bars for each configuration ) learning approaches during an ADL trial (5 in <a href="#sensors-19-03716-t001" class="html-table">Table 1</a>), standard deviation over the various subjects is displayed by whiskers. The different time windows are presented on the <span class="html-italic">x</span>-axis, where the number of past (<span class="html-italic">P</span>) and future (<span class="html-italic">F</span>) poses are shown. The interval (<span class="html-italic">I</span>) between input poses are marked by the different colours, where the number of samples between input poses is shown. For comparison, the mean joint position error (<b>A</b>) for using only the current pose as input (SINN approach) is <math display="inline"><semantics> <mrow> <mn>0.08</mn> </mrow> </semantics></math> (<math display="inline"><semantics> <mrow> <mo>±</mo> <mn>0.01</mn> </mrow> </semantics></math>) m. For comparison, the mean joint jerk error (<b>B</b>) for using only the current pose as input (SINN approach) is <math display="inline"><semantics> <mrow> <mn>1.8</mn> </mrow> </semantics></math> (<math display="inline"><semantics> <mrow> <mo>±</mo> <mn>0.5</mn> </mrow> </semantics></math>) <math display="inline"><semantics> <mrow> <mo>×</mo> <msup> <mn>10</mn> <mn>3</mn> </msup> </mrow> </semantics></math> m/s<sup>3</sup>.</p>
Full article ">Figure 6
<p>Bar plots of the full-body mean (of 6 subjects) joint position (<b>A</b>) and jerk (<b>B</b>) error for the shallow (SINN, left bars for each configuration ) and deep (RNN, right bars for each configuration) learning approaches during a sports trial (10 in <a href="#sensors-19-03716-t001" class="html-table">Table 1</a>), standard deviation over the various subjects is displayed by whiskers. The different time windows are presented on the <span class="html-italic">x</span>-axis, where the number of past (<span class="html-italic">P</span>) and future (<span class="html-italic">F</span>) poses are shown. The interval (<span class="html-italic">I</span>) between input poses are marked by the different colours, where the number of samples between input poses is shown. For comparison, the baseline (L<b>A</b>) for using only the current pose as input mean joint position (<b>A</b>) and jerk (<b>B</b>) errors are shown as the dark blue bars on the left.</p>
Full article ">Figure 7
<p>Bar plots of the mean (of 6 subjects) joint position and jerk error for the shallow (SINN) and deep (RNN) learning approaches (left and right, respectively) using orientation features (<b>O</b> in black) and including accelerations (<b>O + A</b> in white). Three different types of activities are shown namely: gait (<b>A</b>), sports (<b>B</b>) and ADL (<b>C</b>). Standard deviation over the various subjects is displayed by black whiskers. These results were obtained using the following parameters: <math display="inline"><semantics> <mrow> <mi>I</mi> <mo>=</mo> <mn>2</mn> </mrow> </semantics></math>, <math display="inline"><semantics> <mrow> <mi>H</mi> <mo>=</mo> <mn>2</mn> </mrow> </semantics></math> and <math display="inline"><semantics> <mrow> <mi>F</mi> <mo>=</mo> <mn>2</mn> </mrow> </semantics></math>.</p>
Full article ">
19 pages, 9429 KiB  
Article
Logistic Regression for Machine Learning in Process Tomography
by Tomasz Rymarczyk, Edward Kozłowski, Grzegorz Kłosowski and Konrad Niderla
Sensors 2019, 19(15), 3400; https://doi.org/10.3390/s19153400 - 2 Aug 2019
Cited by 138 | Viewed by 12593
Abstract
The main goal of the research presented in this paper was to develop a refined machine learning algorithm for industrial tomography applications. The article presents algorithms based on logistic regression in relation to image reconstruction using electrical impedance tomography (EIT) and ultrasound transmission [...] Read more.
The main goal of the research presented in this paper was to develop a refined machine learning algorithm for industrial tomography applications. The article presents algorithms based on logistic regression in relation to image reconstruction using electrical impedance tomography (EIT) and ultrasound transmission tomography (UST). The test object was a tank filled with water in which reconstructed objects were placed. For both EIT and UST, a novel approach was used in which each pixel of the output image was reconstructed by a separately trained prediction system. Therefore, it was necessary to use many predictive systems whose number corresponds to the number of pixels of the output image. Thanks to this approach the under-completed problem was changed to an over-completed one. To reduce the number of predictors in logistic regression by removing irrelevant and mutually correlated entries, the elastic net method was used. The developed algorithm that reconstructs images pixel-by-pixel is insensitive to the shape, number and position of the reconstructed objects. In order to assess the quality of mappings obtained thanks to the new algorithm, appropriate metrics were used: compatibility ratio (CR) and relative error (RE). The obtained results enabled the assessment of the usefulness of logistic regression in the reconstruction of EIT and UST images. Full article
(This article belongs to the Special Issue Intelligent Sensor Signal in Machine Learning)
Show Figures

Figure 1

Figure 1
<p>The physical model of the reactor with: (<b>a</b>,<b>b</b>) electrical impedance tomography (EIT) electrodes, (<b>c</b>) ultrasound transmission tomography (UST) transducers.</p>
Full article ">Figure 2
<p>Comparison of algorithms for: (<b>a</b>)—single logistic regression unit, (<b>b</b>)—multiple LRS.</p>
Full article ">Figure 3
<p>Model of the EIT system converting electrical signals into a 2D image of the cross-section.</p>
Full article ">Figure 4
<p>Model of elastic net + LRS hybrid subsystem dedicated to a particular pixel ψ<sub>181</sub>.</p>
Full article ">Figure 5
<p>A measuring case generated with the simulation method of EIT with a graph showing the 96 voltage measurements between different pairs of electrodes.</p>
Full article ">Figure 6
<p>Cross-validated MSE of elastic net fit (alpha = 0.9) for pixel ψ<sub>181</sub>.</p>
Full article ">Figure 7
<p>β vs. L1—trace plot of coefficients fit by elastic net (alpha = 0.9) for pixel ψ<sub>181</sub>.</p>
Full article ">Figure 8
<p>β vs. lambda—trace plot of coefficients fit by elastic net (alpha = 0.9) for pixel ψ<sub>181</sub>.</p>
Full article ">Figure 9
<p>EIT image reconstructions for different classification thresholds <span class="html-italic">l.</span></p>
Full article ">Figure 10
<p>UST image reconstructions for classification threshold <span class="html-italic">l</span> = 0.9.</p>
Full article ">Figure 10 Cont.
<p>UST image reconstructions for classification threshold <span class="html-italic">l</span> = 0.9.</p>
Full article ">
18 pages, 3667 KiB  
Article
A Combined Offline and Online Algorithm for Real-Time and Long-Term Classification of Sheep Behaviour: Novel Approach for Precision Livestock Farming
by Jorge A. Vázquez-Diosdado, Veronica Paul, Keith A Ellis, David Coates, Radhika Loomba and Jasmeet Kaler
Sensors 2019, 19(14), 3201; https://doi.org/10.3390/s19143201 - 20 Jul 2019
Cited by 37 | Viewed by 5867
Abstract
Real-time and long-term behavioural monitoring systems in precision livestock farming have huge potential to improve welfare and productivity for the better health of farm animals. However, some of the biggest challenges for long-term monitoring systems relate to “concept drift”, which occurs when systems [...] Read more.
Real-time and long-term behavioural monitoring systems in precision livestock farming have huge potential to improve welfare and productivity for the better health of farm animals. However, some of the biggest challenges for long-term monitoring systems relate to “concept drift”, which occurs when systems are presented with challenging new or changing conditions, and/or in scenarios where training data is not accurately reflective of live sensed data. This study presents a combined offline algorithm and online learning algorithm which deals with concept drift and is deemed by the authors as a useful mechanism for long-term in-the-field monitoring systems. The proposed algorithm classifies three relevant sheep behaviours using information from an embedded edge device that includes tri-axial accelerometer and tri-axial gyroscope sensors. The proposed approach is for the first time reported in precision livestock behavior monitoring and demonstrates improvement in classifying relevant behaviour in sheep, in real-time, under dynamically changing conditions. Full article
(This article belongs to the Special Issue Intelligent Sensor Signal in Machine Learning)
Show Figures

Figure 1

Figure 1
<p>Offline learning scheme. Both train and test sets are disjoint.</p>
Full article ">Figure 2
<p>Online learning scheme.</p>
Full article ">Figure 3
<p>Boxplot of the MeanAMag across the different behaviours.</p>
Full article ">Figure 4
<p>Combined offline and online learning scheme. In this scheme KNN refers to the K-nearest neighbours model (trained offline).</p>
Full article ">Figure 5
<p>Decision tree rules for the combination of both KNN and k-means algorithms.</p>
Full article ">Figure 6
<p>Satellite view of field with drawn and labelled perimeter.</p>
Full article ">Figure 7
<p>Custom-made device (including processor, memory, radio, inertial measurement unit (IMU), etc.).</p>
Full article ">Figure 8
<p>Distribution of the MeanAMag for our previously collected dataset [<a href="#B8-sensors-19-03201" class="html-bibr">8</a>] (dataset 1) and for this study (dataset 2). The plot shows the distribution change between the two studies for the three different behaviours. (<b>a</b>–<b>c</b>) show the distributions of the MeanAMag for walking, standing and lying using collected dataset [<a href="#B8-sensors-19-03201" class="html-bibr">8</a>], (<b>d</b>–<b>f</b>) show the distribution of the MeanAMag for this study.</p>
Full article ">Figure 9
<p>Evolution of the centres of the different classes computed using the k-means online learning algorithm with data collected for this study. The k-means online learning algorithm will discriminate between high, medium and low MeanAMag values.</p>
Full article ">
0 pages, 5279 KiB  
Article
RETRACTED: The Novel Sensor Network Structure for Classification Processing Based on the Machine Learning Method of the ACGAN
by Yuantao Chen, Jiajun Tao, Jin Wang, Xi Chen, Jingbo Xie, Jie Xiong and Kai Yang
Sensors 2019, 19(14), 3145; https://doi.org/10.3390/s19143145 - 17 Jul 2019
Cited by 18 | Viewed by 3997 | Retraction
Abstract
To address the problem of unstable training and poor accuracy in image classification algorithms based on generative adversarial networks (GAN), a novel sensor network structure for classification processing using auxiliary classifier generative adversarial networks (ACGAN) is proposed in this paper. Firstly, the real/fake [...] Read more.
To address the problem of unstable training and poor accuracy in image classification algorithms based on generative adversarial networks (GAN), a novel sensor network structure for classification processing using auxiliary classifier generative adversarial networks (ACGAN) is proposed in this paper. Firstly, the real/fake discrimination of sensor samples in the network has been canceled at the output layer of the discriminative network and only the posterior probability estimation of the sample tag is outputted. Secondly, by regarding the real sensor samples as supervised data and the generative sensor samples as labeled fake data, we have reconstructed the loss function of the generator and discriminator by using the real/fake attributes of sensor samples and the cross-entropy loss function of the label. Thirdly, the pooling and caching method has been introduced into the discriminator to enable more effective extraction of the classification features. Finally, feature matching has been added to the discriminative network to ensure the diversity of the generative sensor samples. Experimental results have shown that the proposed algorithm (CP-ACGAN) achieves better classification accuracy on the MNIST dataset, CIFAR10 dataset and CIFAR100 dataset than other solutions. Moreover, when compared with the ACGAN and CNN classification algorithms, which have the same deep network structure as CP-ACGAN, the proposed method continues to achieve better classification effects and stability than other main existing sensor solutions. Full article
(This article belongs to the Special Issue Intelligent Sensor Signal in Machine Learning)
Show Figures

Figure 1

Figure 1
<p>The original GAN network structure [<a href="#B9-sensors-19-03145" class="html-bibr">9</a>].</p>
Full article ">Figure 2
<p>The ACGAN network structure [<a href="#B16-sensors-19-03145" class="html-bibr">16</a>].</p>
Full article ">Figure 3
<p>The generator structure of ACGAN [<a href="#B16-sensors-19-03145" class="html-bibr">16</a>].</p>
Full article ">Figure 4
<p>Discriminator structure using an ACGAN.</p>
Full article ">Figure 5
<p>The CP-ACGAN network structure.</p>
Full article ">Figure 6
<p>The discriminator structure of the CP-ACGAN.</p>
Full article ">Figure 7
<p>Classification effect of different methods on the MNIST training dataset.</p>
Full article ">Figure 8
<p>Testing accuracy comparison of different methods on the MNIST testing dataset.</p>
Full article ">Figure 9
<p>Comparison of different methods on the MNIST testing dataset.</p>
Full article ">Figure 10
<p>Average classification accuracy of different methods on the CIFAR10 training dataset.</p>
Full article ">Figure 11
<p>Average testing accuracy comparison on the CIFAR10 testing dataset.</p>
Full article ">Figure 12
<p>Average classification accuracy of different methods on the CIFAR100 training dataset.</p>
Full article ">Figure 13
<p>Average testing accuracy comparison on the CIFAR100 testing dataset.</p>
Full article ">
17 pages, 4045 KiB  
Article
A Cascade Ensemble Learning Model for Human Activity Recognition with Smartphones
by Shoujiang Xu, Qingfeng Tang, Linpeng Jin and Zhigeng Pan
Sensors 2019, 19(10), 2307; https://doi.org/10.3390/s19102307 - 19 May 2019
Cited by 32 | Viewed by 5105
Abstract
Human activity recognition (HAR) has gained lots of attention in recent years due to its high demand in different domains. In this paper, a novel HAR system based on a cascade ensemble learning (CELearning) model is proposed. Each layer of the proposed model [...] Read more.
Human activity recognition (HAR) has gained lots of attention in recent years due to its high demand in different domains. In this paper, a novel HAR system based on a cascade ensemble learning (CELearning) model is proposed. Each layer of the proposed model is comprised of Extremely Gradient Boosting Trees (XGBoost), Random Forest, Extremely Randomized Trees (ExtraTrees) and Softmax Regression, and the model goes deeper layer by layer. The initial input vectors sampled from smartphone accelerometer and gyroscope sensor are trained separately by four different classifiers in the first layer, and the probability vectors representing different classes to which each sample belongs are obtained. Both the initial input data and the probability vectors are concatenated together and considered as input to the next layer’s classifiers, and eventually the final prediction is obtained according to the classifiers of the last layer. This system achieved satisfying classification accuracy on two public datasets of HAR based on smartphone accelerometer and gyroscope sensor. The experimental results show that the proposed approach has gained better classification accuracy for HAR compared to existing state-of-the-art methods, and the training process of the model is simple and efficient. Full article
(This article belongs to the Special Issue Intelligent Sensor Signal in Machine Learning)
Show Figures

Figure 1

Figure 1
<p>Overview of HAR system. Handcrafted feature extraction based HAR contains data collection, signal processing, feature extraction and CELearning model. Automatic feature extraction based HAR contains data collection, FFT and CELearning model.</p>
Full article ">Figure 2
<p>CELearning model. Each layer is composed of four basic classifiers which generate the probability vectors as augmented features for next layer’s learning.</p>
Full article ">Figure 3
<p>Class vector generation of the randomized decision trees. A probability vector is obtained from each decision tree and the final probability vector of randomized decision trees is jointly generated by all the decision trees.</p>
Full article ">Figure 4
<p>Augmented features generation of each classifier. K-fold cross validation is used for each classifier to generate K-1 estimated class vectors, which are averaged to obtain a final vector as augmented features.</p>
Full article ">Figure 5
<p>Three-axial linear acceleration and three-axial angular velocity: (<b>a</b>–<b>c</b>) the three-axis data of the accelerometer, respectively; and (<b>d</b>–<b>f</b>) the three-axis data of the gyroscope, respectively.</p>
Full article ">Figure 6
<p>Convergence curves of the proposed model for HAR: (<b>a</b>) the convergence curve of handcrafted feature extraction based HAR; and (<b>b</b>) the convergence curve of automatic feature extraction based HAR.</p>
Full article ">
13 pages, 2108 KiB  
Article
Contextual Action Cues from Camera Sensor for Multi-Stream Action Recognition
by Jongkwang Hong, Bora Cho, Yong Won Hong and Hyeran Byun
Sensors 2019, 19(6), 1382; https://doi.org/10.3390/s19061382 - 20 Mar 2019
Cited by 20 | Viewed by 3894
Abstract
In action recognition research, two primary types of information are appearance and motion information that is learned from RGB images through visual sensors. However, depending on the action characteristics, contextual information, such as the existence of specific objects or globally-shared information in the [...] Read more.
In action recognition research, two primary types of information are appearance and motion information that is learned from RGB images through visual sensors. However, depending on the action characteristics, contextual information, such as the existence of specific objects or globally-shared information in the image, becomes vital information to define the action. For example, the existence of the ball is vital information distinguishing “kicking” from “running”. Furthermore, some actions share typical global abstract poses, which can be used as a key to classify actions. Based on these observations, we propose the multi-stream network model, which incorporates spatial, temporal, and contextual cues in the image for action recognition. We experimented on the proposed method using C3D or inflated 3D ConvNet (I3D) as a backbone network, regarding two different action recognition datasets. As a result, we observed overall improvement in accuracy, demonstrating the effectiveness of our proposed method. Full article
(This article belongs to the Special Issue Intelligent Sensor Signal in Machine Learning)
Show Figures

Figure 1

Figure 1
<p>Examples from UCF sports [<a href="#B11-sensors-19-01382" class="html-bibr">11</a>] dataset, each row representing the image sequence of the video. (<b>a</b>) RGB images from the “Kicking” class. (<b>b</b>) Corresponding optical flow images to (<b>a</b>). (<b>c</b>) RGB images from the “Run” class. (<b>d</b>) Corresponding optical flow images to (<b>c</b>).</p>
Full article ">Figure 2
<p>Overall architecture of the proposed method.</p>
Full article ">Figure 3
<p>Example of an RGB image with corresponding pairwise inputs. (<b>a</b>) “HorseRiding”. (<b>b</b>) “SoccerJugglings”. (<b>c</b>) “HorseRiding”. (<b>d</b>) “SoccerJugglings”.</p>
Full article ">Figure 4
<p>Details about pairwise stream inputs; images of each row are from the same video in different timelines. Solid red line boxes represent the “actors” in the action. Dotted red line boxes represent the “person” who is regarded as an “object”. Solid blue line boxes represent the “object” of “actors”. (<b>a</b>) “Biking”. (<b>b</b>) “SoccerJugglings”. (<b>c</b>) “Basketball”. (<b>d</b>) “IceDancing”.</p>
Full article ">Figure 5
<p>Example of an RGB image with corresponding keypoints inputs. (<b>a</b>) “Basketball”. (<b>b</b>) “Biking”. (<b>c</b>) “IceDancing”. (<b>d</b>) “HandbandPushups”.</p>
Full article ">
18 pages, 1733 KiB  
Article
Estimation of Pedestrian Pose Orientation Using Soft Target Training Based on Teacher–Student Framework
by DuYeong Heo, Jae Yeal Nam and Byoung Chul Ko
Sensors 2019, 19(5), 1147; https://doi.org/10.3390/s19051147 - 6 Mar 2019
Cited by 12 | Viewed by 4165
Abstract
Semi-supervised learning is known to achieve better generalisation than a model learned solely from labelled data. Therefore, we propose a new method for estimating a pedestrian pose orientation using a soft-target method, which is a type of semi-supervised learning method. Because a convolutional [...] Read more.
Semi-supervised learning is known to achieve better generalisation than a model learned solely from labelled data. Therefore, we propose a new method for estimating a pedestrian pose orientation using a soft-target method, which is a type of semi-supervised learning method. Because a convolutional neural network (CNN) based pose orientation estimation requires large numbers of parameters and operations, we apply the teacher–student algorithm to generate a compressed student model with high accuracy and compactness resembling that of the teacher model by combining a deep network with a random forest. After the teacher model is generated using hard target data, the softened outputs (soft-target data) of the teacher model are used for training the student model. Moreover, the orientation of the pedestrian has specific shape patterns, and a wavelet transform is applied to the input image as a pre-processing step owing to its good spatial frequency localisation property and the ability to preserve both the spatial information and gradient information of an image. For a benchmark dataset considering real driving situations based on a single camera, we used the TUD and KITTI datasets. We applied the proposed algorithm to various driving images in the datasets, and the results indicate that its classification performance with regard to the pose orientation is better than that of other state-of-the-art methods based on a CNN. In addition, the computational speed of the proposed student model is faster than that of other deep CNNs owing to the shorter model structure with a smaller number of parameters. Full article
(This article belongs to the Special Issue Intelligent Sensor Signal in Machine Learning)
Show Figures

Figure 1

Figure 1
<p>Teacher–student learning framework using hard- and soft-target data (in order of learning process): (<b>a</b>) dataset A labelled with a hard target input into the (<b>b</b>) teacher deep network and used for training teacher deep network, (<b>c</b>) teacher random forest (RF) is trained using feature maps of teacher deep network, (<b>d</b>) after finishing training of two-teacher model, unlabelled training dataset B input into the trained two-teacher model, (<b>e</b>) the soft output of the two teachers combined into one soft target vector and (<b>f</b>) a soft target dataset <math display="inline"><semantics> <mrow> <msup> <mi mathvariant="normal">B</mi> <mo>∗</mo> </msup> </mrow> </semantics></math> of the teacher network are used for training the target student network, (<b>g</b>) training of the student model composed of student network and student RF, and (<b>h</b>) the final class probability is generated by combining output probability of two-student model.</p>
Full article ">Figure 2
<p>Eight orientations recognized by the proposed system and corresponding examples of pedestrian poses.</p>
Full article ">Figure 3
<p>Confusion matrix based on accuracy (ACC) of proposed method (%).</p>
Full article ">Figure 4
<p>Five possible pairs of experiment results for determining the number of trees for the student RF.</p>
Full article ">Figure 5
<p>Sample orientation classification results for TUD and KITTI datasets using the proposed method: (<b>a</b>) estimation of pedestrian’s body orientation using TUD dataset, (<b>b</b>) results of pose orientation estimation (POE) using KITTI.</p>
Full article ">
12 pages, 8465 KiB  
Article
A Deep Convolutional Neural Network Inspired by Auditory Perception for Underwater Acoustic Target Recognition
by Honghui Yang, Junhao Li, Sheng Shen and Guanghui Xu
Sensors 2019, 19(5), 1104; https://doi.org/10.3390/s19051104 - 4 Mar 2019
Cited by 73 | Viewed by 5489
Abstract
Underwater acoustic target recognition (UATR) using ship-radiated noise faces big challenges due to the complex marine environment. In this paper, inspired by neural mechanisms of auditory perception, a new end-to-end deep neural network named auditory perception inspired Deep Convolutional Neural Network (ADCNN) is [...] Read more.
Underwater acoustic target recognition (UATR) using ship-radiated noise faces big challenges due to the complex marine environment. In this paper, inspired by neural mechanisms of auditory perception, a new end-to-end deep neural network named auditory perception inspired Deep Convolutional Neural Network (ADCNN) is proposed for UATR. In the ADCNN model, inspired by the frequency component perception neural mechanism, a bank of multi-scale deep convolution filters are designed to decompose raw time domain signal into signals with different frequency components. Inspired by the plasticity neural mechanism, the parameters of the deep convolution filters are initialized randomly, and the is n learned and optimized for UATR. The n, max-pooling layers and fully connected layers extract features from each decomposed signal. Finally, in fusion layers, features from each decomposed signal are merged and deep feature representations are extracted to classify underwater acoustic targets. The ADCNN model simulates the deep acoustic information processing structure of the auditory system. Experimental results show that the proposed model can decompose, model and classify ship-radiated noise signals efficiently. It achieves a classification accuracy of 81.96%, which is the highest in the contrast experiments. The experimental results show that auditory perception inspired deep learning method has encouraging potential to improve the classification performance of UATR. Full article
(This article belongs to the Special Issue Intelligent Sensor Signal in Machine Learning)
Show Figures

Figure 1

Figure 1
<p>The architecture of ADCNN.</p>
Full article ">Figure 2
<p>Spectrogram of recordings. (<b>a</b>) Cargo recording; (<b>b</b>) Passenger ship recording; (<b>c</b>) Tanker recording; (<b>d</b>) Environment noise recording.</p>
Full article ">Figure 3
<p>Visualization of the output of each filter. (<b>a</b>) Testing sample of Cargo class; (<b>b</b>) Testing sample of Passenger ship class.</p>
Full article ">Figure 4
<p>Result of t-SNE feature visualization. (<b>a</b>–<b>e</b>) Feature groups of deep filter sub-networks; (<b>f</b>) Features of layer-1; (<b>g</b>) Features of layer-2.</p>
Full article ">Figure 5
<p>ROC curves of the proposed model and its competitors. (<b>a</b>) Cargo class is positive class; (<b>b</b>) Passenger ship class is positive class; (<b>c</b>) Tanker class is positive class; (<b>d</b>) Environment noise class is positive class.</p>
Full article ">
15 pages, 5193 KiB  
Article
Vision Sensor Based Fuzzy System for Intelligent Vehicles
by Kwangsoo Kim, Yangho Kim and Sooyeong Kwak
Sensors 2019, 19(4), 855; https://doi.org/10.3390/s19040855 - 19 Feb 2019
Cited by 5 | Viewed by 3861
Abstract
Those in the automotive industry and many researchers have become interested in the development of pedestrian protection systems in recent years. In particular, vision-based methods for predicting pedestrian intentions are now being actively studied to improve the performance of pedestrian protection systems. In [...] Read more.
Those in the automotive industry and many researchers have become interested in the development of pedestrian protection systems in recent years. In particular, vision-based methods for predicting pedestrian intentions are now being actively studied to improve the performance of pedestrian protection systems. In this paper, we propose a vision-based system that can detect pedestrians using an on-dash camera in the car, and can then analyze their movements to determine the probability of collision. Information about pedestrians, including position, distance, movement direction, and magnitude are extracted using computer vision technologies and, using this information, a fuzzy rule-based system makes a judgement on the pedestrian’s risk level. To verify the function of the proposed system, we built several test datasets, collected by ourselves, in high-density regions where vehicles and pedestrians mix closely. The true positive rate of the experimental results was about 86%, which shows the validity of the proposed system. Full article
(This article belongs to the Special Issue Intelligent Sensor Signal in Machine Learning)
Show Figures

Figure 1

Figure 1
<p>Overview of the proposed system: The system has two main modules, a feature extraction module and a risk analysis module. These take videos from a car dash camera and vehicle speed information from an external GPS module, and output the level of risk to pedestrians.</p>
Full article ">Figure 2
<p>The speed information obtained from an external GPS module is written to a text file with the PC clock time. The speed information is synchronized with the video input from the on-dash camera in a car using the PC clock time afterwards.</p>
Full article ">Figure 3
<p>Detection of pedestrian using the histogram of gradient (HOG) + support vector machine (SVM) method. The pedestrian detected is marked with a green box bounding the pedestrian. The midpoint of the base of this box is used as the pedestrian’s position in the image.</p>
Full article ">Figure 4
<p>Transformation of the image from 2D to 3D. (<b>a</b>) Perspective 2D original image, (<b>b</b>) transformed 3D image: Because the distance should be computed in 3D view or top-view, the original perspective-view image is transformed using the homography transformation. Several distances (5 m, 7 m, 10 m, and 20 m) calculated in the transformed image are indicated on the original image, as shown in (<b>a</b>).</p>
Full article ">Figure 5
<p>Basic configuration of fuzzy logic systems [<a href="#B22-sensors-19-00855" class="html-bibr">22</a>].</p>
Full article ">Figure 6
<p>Membership function for two fuzzy inputs and a fuzzy output. (<b>a</b>) MovDirection, (<b>b</b>) MovMag, (<b>c</b>) Result.</p>
Full article ">Figure 7
<p>Classification of pedestrian’s location based on the vanishing point and two lines: The vanishing point and two lines move accordingly when the vehicle turns left or right.</p>
Full article ">Figure 8
<p>Steps for finding the two lines that will be used as guidelines for classification of the pedestrian’s location: The right panel shows an example of this process where more than two lines are detected from a line detection algorithm, the Hough transform. Among those lines, two lines are selected between specific angles.</p>
Full article ">Figure 9
<p>Membership function for the pedestrian’s position (PedPos): The values of <math display="inline"><semantics> <mrow> <msub> <mi>x</mi> <mn>1</mn> </msub> </mrow> </semantics></math> and <math display="inline"><semantics> <mrow> <msub> <mi>x</mi> <mn>2</mn> </msub> </mrow> </semantics></math> are determined by the <span class="html-italic">x</span>-coordinate of the two lines in the transformed 3D image. The two lines are detected by the process explained in <a href="#sensors-19-00855-f008" class="html-fig">Figure 8</a>.</p>
Full article ">Figure 10
<p>Conceptual classification of the pedestrian’s distance: The distance between a pedestrian and the ego-vehicle is classified as one of three regions, determined by the horizontal line passing through the vanishing point and the stopping distance line.</p>
Full article ">Figure 11
<p>Membership functions for pedestrian’s distance (PedDist): The value of <math display="inline"><semantics> <mrow> <msub> <mi>x</mi> <mn>3</mn> </msub> </mrow> </semantics></math> and <math display="inline"><semantics> <mrow> <msub> <mi>x</mi> <mn>4</mn> </msub> </mrow> </semantics></math> are determined by the x-coordinate of two lines in the transformed 3D image.</p>
Full article ">Figure 12
<p>Confusion matrices for each dataset and total database. (<b>a</b>) ATP_S1 dataset, (<b>b</b>) ATP_S2 dataset, (<b>c</b>) ATP_S3 dataset, (<b>d</b>) ATP_S4 dataset, (<b>e</b>) ATP_S5 dataset, (<b>f</b>) STORE_S1 dataset, (<b>g</b>) UNIV_S1 dataset, (<b>h</b>) UNIV_S2 dataset, (<b>i</b>) Total: The columns of the matrix denote ground truths made by humans, and the rows of the matrix denote decisions made by the proposed system. All correct predictions are located in the diagonal of the table.</p>
Full article ">Figure 13
<p>Some working examples of the proposed system. (<b>a</b>) Frame 51 of the APT_S1 dataset, (<b>b</b>) frame 2321 of the APT_S1 dataset, (<b>c</b>) frame 2980 of the UNIV_S1 dataset, (<b>d</b>) frame 3273 of the UNIV_S1 dataset. The results are illustrated with green bounding boxes around the detected pedestrians, the red arrows indicate the direction of pedestrian movement, and the red risk level, above the green box, was obtained from the fuzzy system. Along with the information around the boxes, the frame number and vehicle speed, as well as the calculated stopping distance at that time, are displayed on the top border above the image. The estimated pedestrian position, distance from the vehicle, and the direction of movement are displayed on the border below the image.</p>
Full article ">Figure 14
<p>Two examples of failure of the proposed system: (<b>a</b>) Frame 2759 of the APT_S1 dataset. Due to the blurred image, many pedestrians were not detected. (<b>b</b>) Frame 526 of the APT_S1 dataset. Incorrect feature extraction and direction of movement for the right pedestrian led to an incorrect judgement. That is, she was quite safe actually, but got a high risk level.</p>
Full article ">

Review

Jump to: Research

34 pages, 797 KiB  
Review
Multiple Physiological Signals Fusion Techniques for Improving Heartbeat Detection: A Review
by Javier Tejedor, Constantino A. García, David G. Márquez, Rafael Raya and Abraham Otero
Sensors 2019, 19(21), 4708; https://doi.org/10.3390/s19214708 - 29 Oct 2019
Cited by 21 | Viewed by 5618
Abstract
This paper presents a review of the techniques found in the literature that aim to achieve a robust heartbeat detection from fusing multi-modal physiological signals (e.g., electrocardiogram (ECG), blood pressure (BP), artificial blood pressure (ABP), stroke volume (SV), photoplethysmogram (PPG), electroencephalogram (EEG), electromyogram [...] Read more.
This paper presents a review of the techniques found in the literature that aim to achieve a robust heartbeat detection from fusing multi-modal physiological signals (e.g., electrocardiogram (ECG), blood pressure (BP), artificial blood pressure (ABP), stroke volume (SV), photoplethysmogram (PPG), electroencephalogram (EEG), electromyogram (EMG), and electrooculogram (EOG), among others). Techniques typically employ ECG, BP, and ABP, of which usage has been shown to obtain the best performance under challenging conditions. SV, PPG, EMG, EEG, and EOG signals can help increase performance when included within the fusion. Filtering, signal normalization, and resampling are common preprocessing steps. Delay correction between the heartbeats obtained over some of the physiological signals must also be considered, and signal-quality assessment to retain the best signal/s must be considered as well. Fusion is usually accomplished by exploiting regularities in the RR intervals; by selecting the most promising signal for the detection at every moment; by a voting process; or by performing simultaneous detection and fusion using Bayesian techniques, hidden Markov models, or neural networks. Based on the results of the review, guidelines to facilitate future comparison of the performance of the different proposals are given and promising future lines of research are pointed out. Full article
(This article belongs to the Special Issue Intelligent Sensor Signal in Machine Learning)
Show Figures

Figure 1

Figure 1
<p>Typical steps of heartbeat detection: “SQA” stands for signal-quality assessment.</p>
Full article ">Figure 2
<p>Examples of signals of interest for heartbeat detection: Blue vertical lines show the heartbeat annotations. This figure corresponds to slp04 record from MIT-BIH Polysomnographic Database.</p>
Full article ">Figure 3
<p>Results for the Physionet 2014 challenge database over (<b>a</b>) training and (<b>b</b>) test III data: <math display="inline"><semantics> <mover> <mrow> <mi>S</mi> <mi>e</mi> </mrow> <mo>¯</mo> </mover> </semantics></math> is represented as mean-Se and <math display="inline"><semantics> <mover> <mrow> <mi>P</mi> <mi>P</mi> <mi>V</mi> </mrow> <mo>¯</mo> </mover> </semantics></math> is represented as mean-PPV. “Ref.” stands for reference.</p>
Full article ">Figure 4
<p>Results for the Physionet 2014 follow-up challenge database over (<b>a</b>) training and (<b>b</b>) test data: <math display="inline"><semantics> <mover> <mrow> <mi>S</mi> <mi>e</mi> </mrow> <mo>¯</mo> </mover> </semantics></math> is represented as mean-Se and <math display="inline"><semantics> <mover> <mrow> <mi>P</mi> <mi>P</mi> <mi>V</mi> </mrow> <mo>¯</mo> </mover> </semantics></math> is represented as mean-PPV.</p>
Full article ">
Back to TopTop