[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (30)

Search Parameters:
Keywords = on-device machine learning

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
14 pages, 1578 KiB  
Article
A Platform of Federated Learning Management for Enhanced Mobile Collaboration
by Farkhod Yusubov and KangYoon Lee
Electronics 2024, 13(20), 4104; https://doi.org/10.3390/electronics13204104 - 18 Oct 2024
Cited by 1 | Viewed by 788
Abstract
Federated learning (FL) has emerged as a crucial technology in today’s data-centric environment, enabling decentralized machine learning while safeguarding user privacy. This study introduces “Federated Learning ML Operations (FedOps) Mobile”, a novel FL framework optimized for the dynamic and heterogeneous ecosystem of mobile [...] Read more.
Federated learning (FL) has emerged as a crucial technology in today’s data-centric environment, enabling decentralized machine learning while safeguarding user privacy. This study introduces “Federated Learning ML Operations (FedOps) Mobile”, a novel FL framework optimized for the dynamic and heterogeneous ecosystem of mobile devices. FedOps Mobile addresses the inherent challenges of FL—such as system scalability, device heterogeneity, and operational efficiency—through advanced on-device training using TensorFlow Lite and CoreML. The framework’s innovative approach includes sophisticated client selection mechanisms that assess device readiness and capabilities, ensuring equitable and efficient participation across the network. Additionally, FedOps Mobile leverages remote device control for seamless task management and continuous learning, all without compromising the user experience. The main contribution of this study is the demonstration that federated learning across heterogeneous devices, especially those using different operating systems, can be both practical and efficient using the FedOps Mobile framework. This was validated through experiments that evaluated three key areas: operational efficiency, model personalization, and resource optimization in multi-device settings. The results showed that the proposed method excels in client selection, energy consumption, and model optimization, establishing a new benchmark for federated learning in diverse and complex environments. Full article
Show Figures

Figure 1

Figure 1
<p>Architecture of FedOps. 1. Platform manager—manages device-specific functionalities and ensures seamless integration and operation of various system components. 2. Network Manager-handles all network-related functionalities, such as registering the client with the FL server and managing network communication. 3. Native Platforms implements machine learning functions, such as Fit(), Evaluate(), and Get weights(), that are crucial for training and evaluating models on the client device. 4. Server side (Microk8s Environment), hosted in a Microk8s environment, plays a pivotal role in managing and coordinating the FL process. 5. FL server, central to the FL process, is the aggregation of models and data analysis. 6. Server Manager manages the overall server operations, including client performance monitoring and data management.</p>
Full article ">Figure 2
<p>Explanation of how FedOps Mobile works after installing the application on mobile. 1. The system administrator registers a new task by sending an FCM message. 2. The online devices send resource information to the FRD. 3. The client selection function is triggered to select devices based on the client selection algorithm 4. Selected device identifications are sent to the FCM. 5. The FCM sends a new message to call selected devices for training. 6. Selected devices connect to the server to obtain the initial global model. 7. The server gives the initial global model to the devices.</p>
Full article ">Figure 3
<p>Performance Increase (Accuracy) Across Rounds for Each Experiment.</p>
Full article ">Figure 4
<p>Energy Consumption Over Rounds for Each Experiment.</p>
Full article ">
18 pages, 18528 KiB  
Article
Data Poisoning Attack against Neural Network-Based On-Device Learning Anomaly Detector by Physical Attacks on Sensors
by Takahito Ino, Kota Yoshida, Hiroki Matsutani and Takeshi Fujino
Sensors 2024, 24(19), 6416; https://doi.org/10.3390/s24196416 - 3 Oct 2024
Viewed by 3105
Abstract
In this paper, we introduce a security approach for on-device learning Edge AIs designed to detect abnormal conditions in factory machines. Since Edge AIs are easily accessible by an attacker physically, there are security risks due to physical attacks. In particular, there is [...] Read more.
In this paper, we introduce a security approach for on-device learning Edge AIs designed to detect abnormal conditions in factory machines. Since Edge AIs are easily accessible by an attacker physically, there are security risks due to physical attacks. In particular, there is a concern that the attacker may tamper with the training data of the on-device learning Edge AIs to degrade the task accuracy. Few risk assessments have been reported. It is important to understand these security risks before considering countermeasures. In this paper, we demonstrate a data poisoning attack against an on-device learning Edge AI. Our attack target is an on-device learning anomaly detection system. The system adopts MEMS accelerometers to measure the vibration of factory machines and detect anomalies. The anomaly detector also adopts a concept drift detection algorithm and multiple models to accommodate multiple normal patterns. For the attack, we used a method in which measurements are tampered with by exposing the MEMS accelerometer to acoustic waves of a specific frequency. The acceleration data falsified by this method were trained on an anomaly detector, and the result was that the abnormal state could not be detected. Full article
Show Figures

Figure 1

Figure 1
<p>Autoencoder-based anomaly detector.</p>
Full article ">Figure 2
<p>Overview of ELM.</p>
Full article ">Figure 3
<p>Overview of concept drift detection algorithm. (<b>a</b>) Trained centroids are sequentially calculated during training. (<b>b</b>) Test centroids are sequentially calculated during inference. (<b>c</b>) When concept drift occurs, the test centroid moves away from the train centroid. (<b>d</b>) When the test centroid exceeds the threshold, a concept drift is detected and a new instance is created. The new instance computes its own train centroid from the latest data (training data).</p>
Full article ">Figure 4
<p>Expected drift rate behavior. (<b>a</b>) Concept drift does not occur; (<b>b</b>) Concept drift occurs.</p>
Full article ">Figure 5
<p>Behavior of multi-instance on-device learning anomaly detector.</p>
Full article ">Figure 6
<p>Behavior of anomaly detector without attack.</p>
Full article ">Figure 7
<p>Behavior of anomaly detector with data poisoning attack.</p>
Full article ">Figure 8
<p>Experimental setup. (<b>a</b>) Overall setup; (<b>b</b>) Cooling fan and speaker.</p>
Full article ">Figure 9
<p>Block diagram of experimental setup.</p>
Full article ">Figure 10
<p>Observed frequency spectrum while both cooling fans are stopped.</p>
Full article ">Figure 11
<p>Relationship between irradiated acoustic wave frequency (in audible range), observed peak frequency, and amplitude.</p>
Full article ">Figure 12
<p>Relationship between irradiated acoustic wave frequency (in ultrasonic range), observed peak frequency, and amplitude.</p>
Full article ">Figure 13
<p>Effects of sound pressure for observed peak amplitude (frequency of acoustic waves: 3000 Hz).</p>
Full article ">Figure 14
<p>Samples of observed data. (<b>a</b>) Normal state; (<b>b</b>) Abnormal state; (<b>c</b>) Poisoned state.</p>
Full article ">Figure 15
<p>Error and drift rate without data poisoning attack.</p>
Full article ">Figure 16
<p>Error and drift rate with data poisoning attack.</p>
Full article ">
33 pages, 14331 KiB  
Article
A Virtual Machine Platform Providing Machine Learning as a Programmable and Distributed Service for IoT and Edge On-Device Computing: Architecture, Transformation, and Evaluation of Integer Discretization
by Stefan Bosse
Algorithms 2024, 17(8), 356; https://doi.org/10.3390/a17080356 - 15 Aug 2024
Viewed by 941
Abstract
Data-driven models used for predictive classification and regression tasks are commonly computed using floating-point arithmetic and powerful computers. We address constraints in distributed sensor networks like the IoT, edge, and material-integrated computing, providing only low-resource embedded computers with sensor data that are acquired [...] Read more.
Data-driven models used for predictive classification and regression tasks are commonly computed using floating-point arithmetic and powerful computers. We address constraints in distributed sensor networks like the IoT, edge, and material-integrated computing, providing only low-resource embedded computers with sensor data that are acquired and processed locally. Sensor networks are characterized by strong heterogeneous systems. This work introduces and evaluates a virtual machine architecture that provides ML as a service layer (MLaaS) on the node level and addresses very low-resource distributed embedded computers (with less than 20 kB of RAM). The VM provides a unified ML instruction set architecture that can be programmed to implement decision trees, ANN, and CNN model architectures using scaled integer arithmetic only. Models are trained primarily offline using floating-point arithmetic, finally converted by an iterative scaling and transformation process, demonstrated in this work by two tests based on simulated and synthetic data. This paper is an extended version of the FedCSIS 2023 conference paper providing new algorithms and ML applications, including ANN/CNN-based regression and classification tasks studying the effects of discretization on classification and regression accuracy. Full article
(This article belongs to the Special Issue Algorithms for Network Systems and Applications)
Show Figures

Figure 1

Figure 1
<p>ARM Cortex M0-based sensor node (STM32L031) implementing the REXA VM for material-integrated GUW sensing with NFC for energy transfer and bidirectional communication with only 8 kB of RAM and 32 kB of ROM.</p>
Full article ">Figure 2
<p>Principle REXA VM network architecture using different wired and wireless communication technologies.</p>
Full article ">Figure 3
<p>Basic REXA-VM architecture with integrated JIT compiler, stacks, and byte-code processor [<a href="#B11-algorithms-17-00356" class="html-bibr">11</a>,<a href="#B12-algorithms-17-00356" class="html-bibr">12</a>].</p>
Full article ">Figure 4
<p>(<b>Left</b>) Incremental growing code segment (single-tasking), persistent code cannot be removed. (<b>Right</b>) Dynamically partitioned code segments using code frames and linking code frames due to fragmentation.</p>
Full article ">Figure 5
<p>Exploding output values for negative x-values (<span class="html-italic">e</span><sup>−x</sup> term) and positive x-values (<span class="html-italic">e</span><sup>x</sup> term) of the exponential function.</p>
Full article ">Figure 6
<p>Relative discretization error of integer-scaled LUT-based approximation of the <span class="html-italic">log10</span> function for different Δ<span class="html-italic">x</span> values (1,2,4) and LUT sizes of 90, 45, and 23, respectively.</p>
Full article ">Figure 7
<p>Relative discretization error of integer-scaled LUT-interpolated approximation of the <span class="html-italic">sigmoid</span> function using the discretized <span class="html-italic">log10</span> LUT-interpolation function for different LUT resolutions and sigmoid segment ranges <span class="html-italic">R</span>. The small error plots show only positive x values.</p>
Full article ">Figure 8
<p>Relative discretization error of integer-scaled LUT-interpolated approximation of the <span class="html-italic">tanh</span> function using the discretized <span class="html-italic">log10</span> LUT-interpolation function.</p>
Full article ">Figure 9
<p>Phase 1 transformation (CNN). (<b>Top</b>) Transformation of 3-dim tensors into multiple vectors for convolutional and pooling layers and flattening of multiple vectors from last convolutional or pooling layer into one vector for the input of a fully connected neuronal layer. (<b>Bottom</b>) Convolutional and pooling operations factorized into sequential and accumulated vector operations.</p>
Full article ">Figure 10
<p>Scaling architectures for (<b>Top</b>) functional nodes, i.e., neurons; (<b>Bottom</b>) convolution or pooling operation.</p>
Full article ">Figure 11
<p>Accumulative scaled convolution or multi-vector input (flattening) neural network operation based on a product–sum calculation. Each accumulative iteration uses a different input scaling <span class="html-italic">s</span><sub>d</sub> normalization with respect to the output scaling <span class="html-italic">s</span>.</p>
Full article ">Figure 12
<p>The ML model transformation pipeline creating an intermediate USM and then creating a sequence of MLISA vector operations.</p>
Full article ">Figure 13
<p>GUW signal simulation using a 2 dim viscoelastic wave propagation model. (<b>Left</b>) Simulation set-up. (<b>Right</b>) Some example signals with and without damage (blue areas show damage features).</p>
Full article ">Figure 14
<p>Down-sampled GUW signal from simulation and low-pass-filtered rectified (envelope approximation) signal as input for the CNN (damage at position x = 100, y = 100).</p>
Full article ">Figure 15
<p>Foo/FooFP model analysis of the GUW regression CNN model. The classification error was always zero.</p>
Full article ">Figure 16
<p>(<b>Top</b>) Analysis of the ANN FP and DS models comparing RMSE and E<sub>max</sub> values for different configurations of the activation function approximation, including an FPU replacement. (<b>Bottom</b>) Selected prediction results are shown with discontinuities in the top plot using ActDS configuration 5 and without using the FPU replacement for the tanh function.</p>
Full article ">
15 pages, 977 KiB  
Article
Dynamic Bandwidth Slicing in Passive Optical Networks to Empower Federated Learning
by Alaelddin F. Y. Mohammed, Joohyung Lee and Sangdon Park
Sensors 2024, 24(15), 5000; https://doi.org/10.3390/s24155000 - 2 Aug 2024
Viewed by 922
Abstract
Federated Learning (FL) is a decentralized machine learning method in which individual devices compute local models based on their data. In FL, devices periodically share newly trained updates with the central server, rather than submitting their raw data. The key characteristics of FL, [...] Read more.
Federated Learning (FL) is a decentralized machine learning method in which individual devices compute local models based on their data. In FL, devices periodically share newly trained updates with the central server, rather than submitting their raw data. The key characteristics of FL, including on-device training and aggregation, make it interesting for many communication domains. Moreover, the potential of new systems facilitating FL in sixth generation (6G) enabled Passive Optical Networks (PON), presents a promising opportunity for integration within this domain. This article focuses on the interaction between FL and PON, exploring approaches for effective bandwidth management, particularly in addressing the complexity introduced by FL traffic. In the PON standard, advanced bandwidth management is proposed by allocating multiple upstream grants utilizing the Dynamic Bandwidth Allocation (DBA) algorithm to be allocated for an Optical Network Unit (ONU). However, there is a lack of research on studying the utilization of multiple grant allocation. In this paper, we address this limitation by introducing a novel DBA approach that efficiently allocates PON bandwidth for FL traffic generation and demonstrates how multiple grants can benefit from the enhanced capacity of implementing PON in carrying out FL flows. Simulations conducted in this study show that the proposed solution outperforms state-of-the-art solutions in several network performance metrics, particularly in reducing upstream delay. This improvement holds great promise for enabling real-time data-intensive services that will be key components of 6G environments. Furthermore, our discussion outlines the potential for the integration of FL and PON as an operational reality capable of supporting 6G networking. Full article
(This article belongs to the Section Communications)
Show Figures

Figure 1

Figure 1
<p>FL integration over PON in 6G.</p>
Full article ">Figure 2
<p>System overview.</p>
Full article ">Figure 3
<p>Bandwidth allocation on a DBA cycle <math display="inline"><semantics> <msub> <mi>T</mi> <mi>c</mi> </msub> </semantics></math>.</p>
Full article ">Figure 4
<p>The process model of the (<b>a</b>) Client, (<b>b</b>) ONU, and (<b>c</b>) OLT.</p>
Full article ">Figure 5
<p>Bandwidth allocation example on a <math display="inline"><semantics> <msub> <mi>T</mi> <mi>c</mi> </msub> </semantics></math> for 8 and 16 ONUs in PON.</p>
Full article ">Figure 6
<p>Evaluating the FL upstream delay.</p>
Full article ">Figure 7
<p>Fairness under different numbers on ONUs.</p>
Full article ">
20 pages, 8592 KiB  
Article
On-Device Semi-Supervised Activity Detection: A New Privacy-Aware Personalized Health Monitoring Approach
by Avirup Roy, Hrishikesh Dutta, Amit Kumar Bhuyan and Subir Biswas
Sensors 2024, 24(14), 4444; https://doi.org/10.3390/s24144444 - 9 Jul 2024
Viewed by 1056
Abstract
This paper presents an on-device semi-supervised human activity detection system that can learn and predict human activity patterns in real time. The clinical objective is to monitor and detect the unhealthy sedentary lifestyle of a user. The proposed semi-supervised learning (SSL) framework uses [...] Read more.
This paper presents an on-device semi-supervised human activity detection system that can learn and predict human activity patterns in real time. The clinical objective is to monitor and detect the unhealthy sedentary lifestyle of a user. The proposed semi-supervised learning (SSL) framework uses sparsely labelled user activity events acquired from Inertial Measurement Unit sensors installed as wearable devices. The proposed cluster-based learning model in this approach is trained with data from the same target user, thus preserving data privacy while providing personalized activity detection services. Two different cluster labelling strategies, namely, population-based and distance-based strategies, are employed to achieve the desired classification performance. The proposed system is shown to be highly accurate and computationally efficient for different algorithmic parameters, which is relevant in the context of limited computing resources on typical wearable devices. Extensive experimentation and simulation study have been conducted on multi-user human activity data from the public domain in order to analyze the trade-off between classification accuracy and computation complexity of the proposed learning paradigm with different algorithmic hyper-parameters. With 4.17 h of training time for 8000 activity episodes, the proposed SSL approach consumes at most 20 KB of CPU memory space, while providing a maximum accuracy of 90% and 100% classification rates. Full article
(This article belongs to the Special Issue IMU and Innovative Sensors for Healthcare)
Show Figures

Figure 1

Figure 1
<p>A wearable on-device human activity detection system.</p>
Full article ">Figure 2
<p>Feature distribution of the three classes.</p>
Full article ">Figure 3
<p>Pre-processing and classification pipeline.</p>
Full article ">Figure 4
<p>Iterative semi-supervised learning framework.</p>
Full article ">Figure 5
<p>Example cluster model evolution with incoming episodes self-trained using iterative semi-supervised learning paradigm.</p>
Full article ">Figure 6
<p>(<b>a</b>–<b>h</b>) Impacts of PCA on accuracy parameters for population-based SSL; (<b>a</b>–<b>c</b>): true positive for the three classes; (<b>d</b>–<b>f</b>) false positive for the three classes; (<b>g</b>) overall accuracy; (<b>h</b>) classification rate.</p>
Full article ">Figure 7
<p>(<b>a</b>–<b>h</b>) Impacts of PCA on accuracy parameters for distance-based SSL; (<b>a</b>–<b>c</b>) true positive for the three classes; (<b>d</b>–<b>f</b>) false positive for the three classes; (<b>g</b>) overall accuracy; (<b>h</b>) classification rate.</p>
Full article ">Figure 8
<p>Impact of PCA on complexity for population-based SSL; (<b>a</b>) computational time; (<b>b</b>) CPU memory usage.</p>
Full article ">Figure 9
<p>Impact of PCA on complexity for distance-based SSL; (<b>a</b>) computational time; (<b>b</b>) CPU memory usage.</p>
Full article ">Figure 10
<p>Post-convergence accuracy parameters’ results for (<b>a</b>–<b>c</b>) population-based SSL and (<b>d</b>–<b>f</b>) distance-based SSL, with varying <math display="inline"><semantics> <mrow> <mi>α</mi> </mrow> </semantics></math>.</p>
Full article ">Figure 11
<p>Total learning time and CPU memory usage for all the episodes using (<b>a</b>) population-based SSL and (<b>b</b>) distance-based SSL, with varying <math display="inline"><semantics> <mrow> <mi>α</mi> </mrow> </semantics></math>.</p>
Full article ">Figure 12
<p>Post-convergence accuracy parameters’ results for (<b>a</b>–<b>c</b>) population-based SSL and (<b>d</b>–<b>f</b>) distance-based SSL, with varying <span class="html-italic">Nc</span>.</p>
Full article ">Figure 13
<p>Total learning time and CPU memory usage for all the episodes using (<b>a</b>) population-based SSL and (<b>b</b>) distance-based SSL, with varying <math display="inline"><semantics> <mrow> <msub> <mrow> <mi>N</mi> </mrow> <mrow> <mi>c</mi> </mrow> </msub> </mrow> </semantics></math>.</p>
Full article ">Figure 14
<p>Post-convergence accuracy parameters’ results for (<b>a</b>–<b>c</b>) population-based SSL and (<b>d</b>–<b>f</b>) distance-based SSL for different clustering methods.</p>
Full article ">Figure 15
<p>Total learning time and CPU memory usage for 8262 episodes using (<b>a</b>) population-based SSL and (<b>b</b>) distance-based SSL for different clustering methods.</p>
Full article ">
18 pages, 5560 KiB  
Article
Towards Full Forward On-Tiny-Device Learning: A Guided Search for a Randomly Initialized Neural Network
by Danilo Pau, Andrea Pisani and Antonio Candelieri
Algorithms 2024, 17(1), 22; https://doi.org/10.3390/a17010022 - 5 Jan 2024
Viewed by 2144
Abstract
In the context of TinyML, many research efforts have been devoted to designing forward topologies to support On-Device Learning. Reaching this target would bring numerous advantages, including reductions in latency and computational complexity, stronger privacy, data safety and robustness to adversarial attacks, higher [...] Read more.
In the context of TinyML, many research efforts have been devoted to designing forward topologies to support On-Device Learning. Reaching this target would bring numerous advantages, including reductions in latency and computational complexity, stronger privacy, data safety and robustness to adversarial attacks, higher resilience against concept drift, etc. However, On-Device Learning on resource constrained devices poses severe limitations to computational power and memory. Therefore, deploying Neural Networks on tiny devices appears to be prohibitive, since their backpropagation-based training is too memory demanding for their embedded assets. Using Extreme Learning Machines based on Convolutional Neural Networks might be feasible and very convenient, especially for Feature Extraction tasks. However, it requires searching for a randomly initialized topology that achieves results as good as those achieved by the backpropagated model. This work proposes a novel approach for automatically composing an Extreme Convolutional Feature Extractor, based on Neural Architecture Search and Bayesian Optimization. It was applied to the CIFAR-10 and MNIST datasets for evaluation. Two search spaces have been defined, as well as a search strategy that has been tested with two surrogate models, Gaussian Process and Random Forest. A performance estimation strategy was defined, keeping the feature set computed by the MLCommons-Tiny benchmark ResNet as a reference model. In as few as 1200 search iterations, the proposed strategy was able to achieve a topology whose extracted features scored a mean square error equal to 0.64 compared to the reference set. Further improvements are required, with a target of at least one order of magnitude decrease in mean square error for improved classification accuracy. The code is made available via GitHub to allow for the reproducibility of the results reported in this paper. Full article
Show Figures

Figure 1

Figure 1
<p>Atomic block for type 1 neural topology.</p>
Full article ">Figure 2
<p>Atomic block for type 2 neural topology. The numbers in the chart indicate the order for the subsequent subtraction operation.</p>
Full article ">Figure 3
<p>Flowchart describing the experimental setup.</p>
Full article ">Figure 4
<p>Progression of the search over the CIFAR-10 dataset, with type 1 neural topology and GP surrogate model.</p>
Full article ">Figure 5
<p>Growth of the optimum over number of iterations for the search with type 1 neural topology and GP surrogate model applied to the CIFAR-10 dataset.</p>
Full article ">Figure 6
<p>Progression of the search over the MNIST dataset with type 1 neural topology and GP surrogate model.</p>
Full article ">Figure 7
<p>Growth of the optimum over number of iterations for the search with type 1 neural topology and GP surrogate model applied to the MNIST dataset.</p>
Full article ">Figure 8
<p>Progression of the search with type 1 neural topology and RF surrogate model.</p>
Full article ">Figure 9
<p>Growth of the optimum found over the number of iterations for the search with type 1 neural topology and RF surrogate model.</p>
Full article ">Figure 10
<p>Schema representing the best performing topology found in all searches for the CIFAR-10 dataset. Every group of three squares represents a block; the first square describes its kernel size, the second its number of filters and the third represents the presence/absence of regularization within the block. See <a href="#algorithms-17-00022-f001" class="html-fig">Figure 1</a> for the composition of a singular block.</p>
Full article ">Figure 11
<p>Progression of the search over the MNIST dataset with type 1 neural topology and RF surrogate model.</p>
Full article ">Figure 12
<p>Growth of the optimum over number of iterations for the search with type 1 neural topology and RF surrogate model applied to the MNIST dataset.</p>
Full article ">Figure 13
<p>Progression of the search with type 2 neural topology and RF surrogate model.</p>
Full article ">Figure 14
<p>Growth of the optimum over number of iterations for the search with type 2 neural topology and RF surrogate model.</p>
Full article ">Figure 15
<p>Progression of the search over the MNIST dataset with type 2 neural topology and RF surrogate model.</p>
Full article ">Figure 16
<p>Growth of the optimum over number of iterations for the search with type 2 neural topology and RF surrogate model applied to the MNIST dataset.</p>
Full article ">Figure 17
<p>Schema representing the best performing topology found in all searches for the MNIST dataset. Every group of three squares represents a block; the first square describes its kernel size, the second its number of filters and the third represents the presence/absence of regularization within the block. See <a href="#algorithms-17-00022-f002" class="html-fig">Figure 2</a> for the composition of a singular block.</p>
Full article ">
17 pages, 1466 KiB  
Article
Smart Buildings: Water Leakage Detection Using TinyML
by Othmane Atanane, Asmaa Mourhir, Nabil Benamar and Marco Zennaro
Sensors 2023, 23(22), 9210; https://doi.org/10.3390/s23229210 - 16 Nov 2023
Cited by 6 | Viewed by 3688
Abstract
The escalating global water usage and the increasing strain on major cities due to water shortages highlights the critical need for efficient water management practices. In water-stressed regions worldwide, significant water wastage is primarily attributed to leakages, inefficient use, and aging infrastructure. Undetected [...] Read more.
The escalating global water usage and the increasing strain on major cities due to water shortages highlights the critical need for efficient water management practices. In water-stressed regions worldwide, significant water wastage is primarily attributed to leakages, inefficient use, and aging infrastructure. Undetected water leakages in buildings’ pipelines contribute to the water waste problem. To address this issue, an effective water leak detection method is required. In this paper, we explore the application of edge computing in smart buildings to enhance water management. By integrating sensors and embedded Machine Learning models, known as TinyML, smart water management systems can collect real-time data, analyze it, and make accurate decisions for efficient water utilization. The transition to TinyML enables faster and more cost-effective local decision-making, reducing the dependence on centralized entities. In this work, we propose a solution that can be adapted for effective leakage detection in real-world scenarios with minimum human intervention using TinyML. We follow an approach that is similar to a typical machine learning lifecycle in production, spanning stages including data collection, training, hyperparameter tuning, offline evaluation and model optimization for on-device resource efficiency before deployment. In this work, we considered an existing water leakage acoustic dataset for polyvinyl chloride pipelines. To prepare the acoustic data for analysis, we performed preprocessing to transform it into scalograms. We devised a water leak detection method by applying transfer learning to five distinct Convolutional Neural Network (CNN) variants, which are namely EfficientNet, ResNet, AlexNet, MobileNet V1, and MobileNet V2. The CNN models were found to be able to detect leakages where a maximum testing accuracy, recall, precision, and F1 score of 97.45%, 98.57%, 96.70%, and 97.63%, respectively, were observed using the EfficientNet model. To enable seamless deployment on the Arduino Nano 33 BLE edge device, the EfficientNet model is compressed using quantization resulting in a low inference time of 1932 ms, a peak RAM usage of 255.3 kilobytes, and a flash usage requirement of merely 48.7 kilobytes. Full article
Show Figures

Figure 1

Figure 1
<p>The systematic approach undertaken to achieve this study.</p>
Full article ">Figure 2
<p>Schematic layout of the experimental pipeline setup (reprinted, with permission, from [<a href="#B14-sensors-23-09210" class="html-bibr">14</a>] @ 2020 Elsevier).</p>
Full article ">Figure 3
<p>NonLeak|Unburied|Scalogram 1.</p>
Full article ">Figure 4
<p>NonLeak|Unburied|Scalogram 7.</p>
Full article ">Figure 5
<p>NonLeak|Unburied|Scalogram 18.</p>
Full article ">Figure 6
<p>NonLeak|Buried|Scalogram 1.</p>
Full article ">Figure 7
<p>NonLeak|Buried|Scalogram 23.</p>
Full article ">Figure 8
<p>NonLeak|Buried|Scalogram 25.</p>
Full article ">Figure 9
<p>Leak|Unburied|Scalogram 5.</p>
Full article ">Figure 10
<p>Leak|Unburied|Scalogram 17.</p>
Full article ">Figure 11
<p>Leak|Unburied|Scalogram 23.</p>
Full article ">Figure 12
<p>Leak|Buried|Scalogram 6.</p>
Full article ">Figure 13
<p>Leak|Buried|Scalogram 12.</p>
Full article ">Figure 14
<p>Leak|Buried|Scalogram 19.</p>
Full article ">
19 pages, 1278 KiB  
Article
Deep-Learning-Based Classification of Cyclic-Alternating-Pattern Sleep Phases
by Yoav Kahana, Aviad Aberdam, Alon Amar and Israel Cohen
Entropy 2023, 25(10), 1395; https://doi.org/10.3390/e25101395 - 28 Sep 2023
Viewed by 1423
Abstract
Determining the cyclic-alternating-pattern (CAP) phases in sleep using electroencephalography (EEG) signals is crucial for assessing sleep quality. However, most current methods for CAP classification primarily rely on classical machine learning techniques, with limited implementation of deep-learning-based tools. Furthermore, these methods often require manual [...] Read more.
Determining the cyclic-alternating-pattern (CAP) phases in sleep using electroencephalography (EEG) signals is crucial for assessing sleep quality. However, most current methods for CAP classification primarily rely on classical machine learning techniques, with limited implementation of deep-learning-based tools. Furthermore, these methods often require manual feature extraction. Herein, we propose a fully automatic deep-learning-based algorithm that leverages convolutional neural network architectures to classify the EEG signals via their time-frequency representations. Through our investigation, we explored using time-frequency analysis techniques and found that Wigner-based representations outperform the commonly used short-time Fourier transform for CAP classification. Additionally, our algorithm incorporates contextual information of the EEG signals and employs data augmentation techniques specifically designed to preserve the time-frequency structure. The model is developed using EEG signals of healthy subjects from the publicly available CAP sleep database (CAPSLPDB) on Physionet. An experimental study demonstrates that our algorithm surpasses existing machine-learning-based methods, achieving an accuracy of 77.5% on a balanced test set and 81.8% when evaluated on an unbalanced test set. Notably, the proposed algorithm exhibits efficiency and scalability, making it suitable for on-device implementation to enhance CAP identification procedures. Full article
(This article belongs to the Special Issue Signal and Information Processing in Networks)
Show Figures

Figure 1

Figure 1
<p>Proposed CAP Classification Workflow: The EEG signal is segmented, transformed into a 2D time-frequency representation (TFR), and fed into a convolutional neural network (CNN) architecture for A/B-phase classification.</p>
Full article ">Figure 2
<p>A demonstration of the cyclic alternating pattern (CAP) in sleep.</p>
Full article ">Figure 3
<p>Proposed method scheme.</p>
Full article ">Figure 4
<p>Incorporated Contextual Information: Each signal second extends to include preceding and subsequent seconds, labeled by its central (main data) second. The figure illustrates A-phase (<b>left</b>) and B-phase (<b>right</b>) 5 s data segments.</p>
Full article ">Figure 5
<p>Example of (<b>a</b>) 5 s 1D-EEG segment from channel E4-A1 and its corresponding time-frequency representations (TFRs): (<b>b</b>) spectrogram (SPEC), (<b>c</b>) Wigner–Ville distribution (WVD), and (<b>d</b>) smoothed pseudo-Wigner–Ville distribution (SPWVD). The WVD exhibits a distinct energy concentration when compared to SPEC, albeit with the tradeoff of noticeable cross-term patterns.</p>
Full article ">Figure 6
<p>Comparison of performance achieved using different time-frequency representations (TFRs) and window sizes. The blue line corresponds to the SPWVD transform, the red line to the WVD, and the green line to the spectrogram (SPEC). The validation and test data are depicted as solid and dashed lines, respectively.</p>
Full article ">Figure 7
<p>Accuracy comparison of various data augmentation techniques. The figure shows the performance of four strategies: no data augmentation (blue), proposed TF-augmentations only (cyan), random time-shifts only (red), and employment of both time-shifts and TF-augmentation (green).</p>
Full article ">Figure 8
<p>Confusion matrix of CAP detection using 9 s segment’s length, SPWVD, and TF-augmentations.</p>
Full article ">Figure A1
<p>Accuracy and loss vs. epoch number for the training (blue) and validation (orange) sets.</p>
Full article ">
22 pages, 987 KiB  
Article
PePTM: An Efficient and Accurate Personalized P2P Learning Algorithm for Home Thermal Modeling
by Karim Boubouh, Robert Basmadjian, Omid Ardakanian, Alexandre Maurer and Rachid Guerraoui
Energies 2023, 16(18), 6594; https://doi.org/10.3390/en16186594 - 13 Sep 2023
Cited by 1 | Viewed by 1166
Abstract
Nowadays, the integration of home automation systems with smart thermostats is a common trend, designed to enhance resident comfort and conserve energy. The introduction of smart thermostats that can run machine learning algorithms has opened the door for on-device training, enabling customized thermal [...] Read more.
Nowadays, the integration of home automation systems with smart thermostats is a common trend, designed to enhance resident comfort and conserve energy. The introduction of smart thermostats that can run machine learning algorithms has opened the door for on-device training, enabling customized thermal experiences in homes. However, leveraging the flexibility offered by on-device learning has been hindered by the absence of a tailored learning scheme that allows for accurate on-device training of thermal models. Traditional centralized learning (CL) and federated learning (FL) schemes rely on a central server that controls the learning experience, compromising the home’s privacy and requiring significant energy to operate. To address these challenges, we propose PePTM, a personalized peer-to-peer thermal modeling algorithm that generates tailored thermal models for each home, offering a controlled learning experience with a minimal training energy footprint while preserving the home’s privacy, an aspect difficult to achieve in both CL and FL. PePTM consists of local and collaborative learning phases that enable each home to train its thermal model and collaboratively improve it with a set of similar homes in a peer-to-peer fashion. To showcase the effectiveness of PePTM, we use a year’s worth of data from US homes to train thermal models using the RNN time-series model and compare the data across three learning schemes: CL, FL, and PePTM, in terms of model performance and the training energy footprint. Our experimental results show that PePTM is significantly energy-efficient, requiring 695 and 40 times less training energy than CL and FL, respectively, while maintaining comparable performance. We believe that PePTM sets the stage for new avenues for on-device thermal model training, providing a personalized thermal experience with reduced energy consumption and enhanced privacy. Full article
(This article belongs to the Section G: Energy and Buildings)
Show Figures

Figure 1

Figure 1
<p>The proposed temporal abstraction is visually represented through the utilization of 15 min time intervals. Comparable homes are assembled into cluster <span class="html-italic">i</span>, composed of N homes. The data collected from these homes are condensed and downsampled on the server for <span class="html-small-caps">CL</span>, while for <span class="html-small-caps">FL</span> and <span class="html-small-caps">PePTM</span>, downsampling occurs on the device.</p>
Full article ">Figure 2
<p>Graphical presentation of the considered centralized (<b>left</b>) and distributed (<b>right</b>) approaches. The latter case is used to conduct experiments for Federated (<b>left</b>) and Peer-to-Peer (<b>right</b>) Learning schemes. In both schemes, worker nodes are implemented in the form of mobile devices.</p>
Full article ">Figure 3
<p>Accuracy and energy consumption of thermal model training under <span class="html-small-caps">CL</span> using different temporal abstraction scenarios.</p>
Full article ">Figure 4
<p>Accuracy and energy consumption of thermal model training with and without using our proposed temporal and spatial abstraction techniques under <span class="html-small-caps">CL</span>, <span class="html-small-caps">FL</span>, and <span class="html-small-caps">PePTM</span>. (<b>a</b>) Accuracy of thermal models using data from all homes, compared to training a thermal model for each cluster of homes. (<b>b</b>) Energy consumption of model training in <span class="html-small-caps">CL</span>, <span class="html-small-caps">FL</span>, and <span class="html-small-caps">PePTM</span> schemes without using temporal abstraction.</p>
Full article ">Figure 5
<p>Different configurations of <span class="html-small-caps">PePTM</span> to determine the optimal number of rounds and sample size to achieve accurate and efficient thermal model training. The light-colored areas present the standard deviations of RMSE for the 207 homes. (<b>a</b>) Accuracy in terms of RMSE for <span class="html-small-caps">PePTM</span>, with 500 rounds and 1 h temporal abstraction. (<b>b</b>) Accuracy in terms of RMSE for <span class="html-small-caps">PePTM</span>, with 500 rounds and 15 min temporal abstraction.</p>
Full article ">Figure 6
<p>Different configurations of <span class="html-small-caps">PePTM</span> to determine the optimal number of rounds and sample size to achieve accurate and efficient thermal model training. The light-colored areas present the standard deviation of RMSE for the 207 homes. (<b>a</b>) Accuracy in terms of RMSE for <span class="html-small-caps">PePTM</span> with 300 rounds, using a batch size of <math display="inline"><semantics> <mrow> <mi>s</mi> <mo>=</mo> <mn>512</mn> </mrow> </semantics></math> and graph density of <math display="inline"><semantics> <mrow> <mi>ρ</mi> <mo>=</mo> <mn>0.05</mn> </mrow> </semantics></math>. (<b>b</b>) Accuracy in terms of RMSE for <span class="html-small-caps">PePTM</span> using different values of graph density <math display="inline"><semantics> <mi>ρ</mi> </semantics></math>.</p>
Full article ">Figure 7
<p>Performance results in terms of RMSE (<b>a</b>) and MAE (<b>b</b>) of training an RNN thermal model in <span class="html-small-caps">CL</span>, <span class="html-small-caps">FL</span>, and <span class="html-small-caps">PePTM</span>, using different temporal abstractions.</p>
Full article ">Figure 8
<p>Energy consumption of training an RNN thermal model in <span class="html-small-caps">CL</span>, <span class="html-small-caps">FL</span>, and <span class="html-small-caps">PePTM</span> using different temporal abstractions. Note that the y-axis is on a logarithmic scale, as the energy use of <span class="html-small-caps">CL</span> and <span class="html-small-caps">FL</span> is significantly larger than that of <span class="html-small-caps">PePTM</span>.</p>
Full article ">
22 pages, 1377 KiB  
Article
Agent-Based Collaborative Random Search for Hyperparameter Tuning and Global Function Optimization
by Ahmad Esmaeili, Zahra Ghorrati and Eric T. Matson
Systems 2023, 11(5), 228; https://doi.org/10.3390/systems11050228 - 5 May 2023
Cited by 9 | Viewed by 2236
Abstract
Hyperparameter optimization is one of the most tedious yet crucial steps in training machine learning models. There are numerous methods for this vital model-building stage, ranging from domain-specific manual tuning guidelines suggested by the oracles to the utilization of general purpose black-box optimization [...] Read more.
Hyperparameter optimization is one of the most tedious yet crucial steps in training machine learning models. There are numerous methods for this vital model-building stage, ranging from domain-specific manual tuning guidelines suggested by the oracles to the utilization of general purpose black-box optimization techniques. This paper proposes an agent-based collaborative technique for finding near-optimal values for any arbitrary set of hyperparameters (or decision variables) in a machine learning model (or a black-box function optimization problem). The developed method forms a hierarchical agent-based architecture for the distribution of the searching operations at different dimensions and employs a cooperative searching procedure based on an adaptive width-based random sampling technique to locate the optima. The behavior of the presented model, specifically against changes in its design parameters, is investigated in both machine learning and global function optimization applications, and its performance is compared with that of two randomized tuning strategies that are commonly used in practice. Moreover, we have compared the performance of the proposed approach against particle swarm optimization (PSO) and simulated annealing (SA) methods in function optimization to provide additional insights into its exploration in the search space. According to the empirical results, the proposed model outperformed the compared random-based methods in almost all tasks conducted, notably in a higher number of dimensions and in the presence of limited on-device computational resources. Full article
Show Figures

Figure 1

Figure 1
<p>Hierarchical structure built for <math display="inline"><semantics> <mrow> <msub> <mi mathvariant="bold">λ</mi> <mi mathvariant="bold-italic">o</mi> </msub> <mo>=</mo> <mrow> <mo>{</mo> <msub> <mi>λ</mi> <mn>1</mn> </msub> <mo>,</mo> <msub> <mi>λ</mi> <mn>2</mn> </msub> <mo>,</mo> <msub> <mi>λ</mi> <mn>3</mn> </msub> <mo>,</mo> <msub> <mi>λ</mi> <mn>4</mn> </msub> <mo>,</mo> <msub> <mi>λ</mi> <mn>5</mn> </msub> <mo>,</mo> <msub> <mi>λ</mi> <mn>6</mn> </msub> <mo>}</mo> </mrow> </mrow> </semantics></math>, where the primary and complementary hyperparameters of each node are, respectively, highlighted in green and orange, and the labels are the indexes of <math display="inline"><semantics> <msub> <mi>λ</mi> <mi>i</mi> </msub> </semantics></math>.</p>
Full article ">Figure 2
<p>A toy example demonstrating three iterations of running the proposed method for tuning two hyperparameters <math display="inline"><semantics> <msub> <mi>λ</mi> <mn>1</mn> </msub> </semantics></math> and <math display="inline"><semantics> <msub> <mi>λ</mi> <mn>2</mn> </msub> </semantics></math> using terminal agents <math display="inline"><semantics> <msubsup> <mi>g</mi> <mrow> <msub> <mi>λ</mi> <mn>1</mn> </msub> </mrow> <mn>1</mn> </msubsup> </semantics></math> and <math display="inline"><semantics> <msubsup> <mi>g</mi> <mrow> <msub> <mi>λ</mi> <mn>2</mn> </msub> </mrow> <mn>1</mn> </msubsup> </semantics></math>, respectively. It is assumed that for each agent, <math display="inline"><semantics> <mrow> <mi>b</mi> <mo>=</mo> <mn>3</mn> </mrow> </semantics></math>.</p>
Full article ">Figure 3
<p>Average performance of the C-support vector classification (SVC) (first row) and stochastic gradient descent (SGD) (second row) classifiers on two synthetic classification datasets based on the accuracy measure. The error bars in each plot are calculated based on the standard error.</p>
Full article ">Figure 4
<p>Average performance of the passive aggressive (first row) and elastic net (second row) regression algorithms on two synthetic regression datasets based on the mean squared error (MSE) measure. The error bars in each plot are calculated based on the standard error.</p>
Full article ">Figure 5
<p>Average values of the Hartmann function optimized under variable iterations, budgets, explorations, and connection thresholds. Each row of the figure pertains to a particular dimension size, and the error bars are calculated based on the standard error.</p>
Full article ">Figure 6
<p>Average values of the Rastrigin function optimized under variable iterations, budgets, explorations, and connection thresholds. Each row of the figure pertains to a particular dimension size, and the error bars are calculated based on the standard error.</p>
Full article ">Figure 7
<p>Average values of the Styblinski–Tang function optimized under variable iterations, budgets, explorations, and connection thresholds. Each row of the figure pertains to a particular dimension size, and the error bars are calculated based on the standard error.</p>
Full article ">Figure 8
<p>Average values of the toy mean absolute error function optimized under variable iterations, budgets, explorations, and connection thresholds. Each row of the figure pertains to a particular dimension size, and the error bars are calculated based on the standard error.</p>
Full article ">Figure 9
<p>Average function values of four objective functions optimized under a variable number of iterations. Each row of the figure pertains to a particular dimension size, and the error bars are calculated based on the standard error.</p>
Full article ">Figure 10
<p>Average function values of four objective functions optimized under a variable number of budget values. Each row of the figure pertains to a particular dimension size, and the error bars are calculated based on the standard error.</p>
Full article ">
22 pages, 1424 KiB  
Article
An Impact Localization Solution Using Embedded Intelligence—Methodology and Experimental Verification via a Resource-Constrained IoT Device
by Ioannis Katsidimas, Vassilis Kostopoulos, Thanasis Kotzakolios, Sotiris E. Nikoletseas, Stefanos H. Panagiotou and Constantinos Tsakonas
Sensors 2023, 23(2), 896; https://doi.org/10.3390/s23020896 - 12 Jan 2023
Cited by 4 | Viewed by 2788
Abstract
Recent advances both in hardware and software have facilitated the embedded intelligence (EI) research field, and enabled machine learning and decision-making integration in resource-scarce IoT devices and systems, realizing “conscious” and self-explanatory objects (smart objects). In the context of the broad use of [...] Read more.
Recent advances both in hardware and software have facilitated the embedded intelligence (EI) research field, and enabled machine learning and decision-making integration in resource-scarce IoT devices and systems, realizing “conscious” and self-explanatory objects (smart objects). In the context of the broad use of WSNs in advanced IoT applications, this is the first work to provide an extreme-edge system, to address structural health monitoring (SHM) on polymethyl methacrylate (PPMA) thin-plate. To the best of our knowledge, state-of-the-art solutions primarily utilize impact positioning methods based on the time of arrival of the stress wave, while in the last decade machine learning data analysis has been performed, by more expensive and resource-abundant equipment than general/development purpose IoT devices, both for the collection and the inference stages of the monitoring system. In contrast to the existing systems, we propose a methodology and a system, implemented by a low-cost device, with the benefit of performing an online and on-device impact localization service from an agnostic perspective, regarding the material and the sensors’ location (as none of those attributes are used). Thus, a design of experiments and the corresponding methodology to build an experimental time-series dataset for impact detection and localization is proposed, using ceramic piezoelectric transducers (PZTs). The system is excited with a steel ball, varying the height from which it is released. Based on TinyML technology for embedding intelligence in low-power devices, we implement and validate random forest and shallow neural network models to localize in real-time (less than 400 ms latency) any occurring impacts on the structure, achieving higher than 90% accuracy. Full article
(This article belongs to the Special Issue Wireless Sensor Networks in Industrial Applications)
Show Figures

Figure 1

Figure 1
<p>Thin plate model with the sensors and the impact localization areas. Letters A, B, C, D correspond to the respective sensors, as they are denoted later in the paper as SensorA, SensorB, etc.</p>
Full article ">Figure 2
<p>Ceramic piezoelectric transducer CEB-35D26.</p>
Full article ">Figure 3
<p>Two-dimensional representation of impact coordinates and fall height combinations (with different colors), based on Sobol sequences.</p>
Full article ">Figure 4
<p>Visualization of the steel ball drop setup.</p>
Full article ">Figure 5
<p>Image of the experimental plate setup.</p>
Full article ">Figure 6
<p>PZT modules connected to Arduino NANO 33 BLE.</p>
Full article ">Figure 7
<p>Sampling procedure.</p>
Full article ">Figure 8
<p>Raw sensor values over the first 1 k samples.</p>
Full article ">Figure 9
<p>Shallow neural net architecture.</p>
Full article ">Figure 10
<p>Confusion matrices for random forest (<b>left</b>) and shallow neural network (<b>right</b>).</p>
Full article ">
6 pages, 1969 KiB  
Proceeding Paper
CMOS-MEMS Gas Sensor Dubbed GMOS for SelectiveAnalysis of Gases with Tiny Edge Machine Learning
by Adir Krayden, Maayan Schohet, Oz Shmueli, Dima Shlenkevitch, Tanya Blank, Sara Stolyarova and Yael Nemirovsky
Eng. Proc. 2022, 27(1), 81; https://doi.org/10.3390/ecsa-9-13316 - 1 Nov 2022
Cited by 2 | Viewed by 1506
Abstract
Embedded machine learning, TinyML, is a relatively new and fast-growing field of ML, enabling on-device sensor data analytics at low power requirements. This paper presents possible improvements to GMOS, a gas sensor, using TinyML technology. GMOS is a low-cost catalytic gas sensor, fabricated [...] Read more.
Embedded machine learning, TinyML, is a relatively new and fast-growing field of ML, enabling on-device sensor data analytics at low power requirements. This paper presents possible improvements to GMOS, a gas sensor, using TinyML technology. GMOS is a low-cost catalytic gas sensor, fabricated with the standard CMOS-SOI process, based on a suspended thermal transistor MOS (TMOS). Exothermic combustion reactions lead to temperature increases, which modify the suspended transistor’s (used as the sensing element) current-voltage characteristics. We were able to use GMOS measurements for gas classification (both for gas types, as well as concentration), resulting in high-proficiency gas detection at a low cost. Our preliminary results show great successes in the detection of ethanol and acetone gases. Moreover, we believe the method could be generalized to more gas types, concentrations, and gas mixes in future research. Full article
Show Figures

Figure 1

Figure 1
<p>Comparison between Ethanol and Acetone gases. For the same catalyst, different <italic>T*</italic> are obtained.</p>
Full article ">Figure 2
<p>A schematic depiction of the full system: An Arduino controller requests samples from GMOS over UART and lights a LED bulb according to its gas classification.</p>
Full article ">Figure 3
<p>A GMOS reaction in the presence of 100 PPM Acetone, sampled.</p>
Full article ">Figure 4
<p>The fully connected neural network architecture includes an input layer (1 × 24), two hidden layers (30, 20 neurons) with ReLU activation functions, and an output layer (1 × 3) with a Softmax activation function.</p>
Full article ">Figure 5
<p>Standard deviation feature on a minibatch. Good separation was achieved (little overlap between Acetone and ‘No Gas’ classes).</p>
Full article ">Figure 6
<p>The confusion matrix. 100% success was achieved on the test set.</p>
Full article ">
21 pages, 2055 KiB  
Article
Leveraging IoT-Aware Technologies and AI Techniques for Real-Time Critical Healthcare Applications
by Angela-Tafadzwa Shumba, Teodoro Montanaro, Ilaria Sergi, Luca Fachechi, Massimo De Vittorio and Luigi Patrono
Sensors 2022, 22(19), 7675; https://doi.org/10.3390/s22197675 - 10 Oct 2022
Cited by 42 | Viewed by 6157
Abstract
Personalised healthcare has seen significant improvements due to the introduction of health monitoring technologies that allow wearable devices to unintrusively monitor physiological parameters such as heart health, blood pressure, sleep patterns, and blood glucose levels, among others. Additionally, utilising advanced sensing technologies based [...] Read more.
Personalised healthcare has seen significant improvements due to the introduction of health monitoring technologies that allow wearable devices to unintrusively monitor physiological parameters such as heart health, blood pressure, sleep patterns, and blood glucose levels, among others. Additionally, utilising advanced sensing technologies based on flexible and innovative biocompatible materials in wearable devices allows high accuracy and precision measurement of biological signals. Furthermore, applying real-time Machine Learning algorithms to highly accurate physiological parameters allows precise identification of unusual patterns in the data to provide health event predictions and warnings for timely intervention. However, in the predominantly adopted architectures, health event predictions based on Machine Learning are typically obtained by leveraging Cloud infrastructures characterised by shortcomings such as delayed response times and privacy issues. Fortunately, recent works highlight that a new paradigm based on Edge Computing technologies and on-device Artificial Intelligence significantly improve the latency and privacy issues. Applying this new paradigm to personalised healthcare architectures can significantly improve their efficiency and efficacy. Therefore, this paper reviews existing IoT healthcare architectures that utilise wearable devices and subsequently presents a scalable and modular system architecture to leverage emerging technologies to solve identified shortcomings. The defined architecture includes ultrathin, skin-compatible, flexible, high precision piezoelectric sensors, low-cost communication technologies, on-device intelligence, Edge Intelligence, and Edge Computing technologies. To provide development guidelines and define a consistent reference architecture for improved scalable wearable IoT-based critical healthcare architectures, this manuscript outlines the essential functional and non-functional requirements based on deductions from existing architectures and emerging technology trends. The presented system architecture can be applied to many scenarios, including ambient assisted living, where continuous surveillance and issuance of timely warnings can afford independence to the elderly and chronically ill. We conclude that the distribution and modularity of architecture layers, local AI-based elaboration, and data packaging consistency are the more essential functional requirements for critical healthcare application use cases. We also identify fast response time, utility, comfort, and low cost as the essential non-functional requirements for the defined system architecture. Full article
(This article belongs to the Special Issue IoT Multi Sensors)
Show Figures

Figure 1

Figure 1
<p>Typical Cloud-based architecture.</p>
Full article ">Figure 2
<p>Proposed system architecture.</p>
Full article ">Figure 3
<p>Intelligent Data Acquisition Layer.</p>
Full article ">Figure 4
<p>Autoencoder with separated Encoder and Decoder.</p>
Full article ">Figure 5
<p>Simple Autoencoder structure.</p>
Full article ">Figure 6
<p>Test set-up.</p>
Full article ">Figure 7
<p>Original input vs. reconstructed decoder output.</p>
Full article ">Figure 8
<p>Custom BLE−enabled sensing module. (<b>a</b>) Block diagram, (<b>b</b>) prototype device to scale.</p>
Full article ">
25 pages, 1007 KiB  
Article
Lightweight On-Device Detection of Android Malware Based on the Koodous Platform and Machine Learning
by Mateusz Krzysztoń, Bartosz Bok, Marcin Lew and Andrzej Sikora
Sensors 2022, 22(17), 6562; https://doi.org/10.3390/s22176562 - 31 Aug 2022
Cited by 5 | Viewed by 2539
Abstract
Currently, Android is the most popular operating system among mobile devices. However, as the number of devices with the Android operating system increases, so does the danger of using them. This is especially important as smartphones increasingly authenticate critical activities(e-banking, e-identity). BotSense Mobile [...] Read more.
Currently, Android is the most popular operating system among mobile devices. However, as the number of devices with the Android operating system increases, so does the danger of using them. This is especially important as smartphones increasingly authenticate critical activities(e-banking, e-identity). BotSense Mobile is a tool already integrated with some critical applications (e-banking, e-identity) to increase user safety. In this paper, we focus on the novel functionality of BotSense Mobile: the detection of malware applications on a user device. In addition to the standard blacklist approach, we propose a machine learning-based model for unknown malicious application detection. The lightweight neural network model is deployed on an edge device to avoid sending sensitive user data outside the device. For the same reason, manifest-related features can be used by the detector only. We present a comprehensive empirical analysis of malware detection conducted on recent data (May–June, 2022) from the Koodous platform, which is a collaborative platform where over 70 million Android applications were collected. The research highlighted the problem of machine learning model aging. We evaluated the lightweight model on recent Koodous data and obtained f1=0.77 and high precision (0.9). Full article
(This article belongs to the Special Issue Cybersecurity in the Internet of Things)
Show Figures

Figure 1

Figure 1
<p>BotSense mobile deployment scheme.</p>
Full article ">Figure 2
<p>Malware detection module architecture.</p>
Full article ">Figure 3
<p>Data distribution in the Koodous platform from 2016 to 2021 (note a logarithmic scale).</p>
Full article ">Figure 4
<p>Neural network structure (<b>left</b>) with and (<b>right</b>) without dropout layer.</p>
Full article ">Figure 5
<p>The architecture of the automated model generation mechanism.</p>
Full article ">Figure 6
<p>The process of assessing a newly installed application as malicious or benign.</p>
Full article ">Figure 7
<p>Labeled data distribution in Koodous from 28 April to 21 June.</p>
Full article ">Figure 8
<p>Labeled malicious data distribution in Koodous from the 28 April to the 21 June.</p>
Full article ">Figure 9
<p>Comparisonof three updating model strategies (left-side axis) and the number of samples on each day of the experiment (right-side axis). Gradient boosting classifier was used to build classifiers.</p>
Full article ">Figure 10
<p>Top 20% trials of the hyperparameters search.</p>
Full article ">Figure 11
<p>Comparing the results of chosen neural network and gradient boosting classifier.</p>
Full article ">Figure 12
<p>RMSE score for the predictor output predicting the degree of maliciousness of the application.</p>
Full article ">Figure 13
<p>The RSME value obtained by the regression model in variants without and with applying bias value.</p>
Full article ">
15 pages, 2183 KiB  
Article
FedDP: A Privacy-Protecting Theft Detection Scheme in Smart Grids Using Federated Learning
by Muhammad Mansoor Ashraf, Muhammad Waqas, Ghulam Abbas, Thar Baker, Ziaul Haq Abbas and Hisham Alasmary
Energies 2022, 15(17), 6241; https://doi.org/10.3390/en15176241 - 26 Aug 2022
Cited by 23 | Viewed by 2928
Abstract
In smart grids (SGs), the systematic utilization of consumer energy data while maintaining its privacy is of paramount importance. This research addresses this problem by energy theft detection while preserving the privacy of client data. In particular, this research identifies centralized models as [...] Read more.
In smart grids (SGs), the systematic utilization of consumer energy data while maintaining its privacy is of paramount importance. This research addresses this problem by energy theft detection while preserving the privacy of client data. In particular, this research identifies centralized models as more accurate in predicting energy theft in SGs but with no or significantly less data protection. Current research proposes a novel federated learning (FL) framework, namely FedDP, to tackle this issue. The proposed framework enables various clients to benefit from on-device prediction with very little communication overhead and to learn from the experience of other clients with the help of a central server (CS). Furthermore, for the accurate identification of energy theft, the use of a novel federated voting classifier (FVC) is proposed. FVC uses the majority voting-based consensus of traditional machine learning (ML) classifiers namely, random forests (RF), k-nearest neighbors (KNN), and bagging classifiers (BG). To the best of our knowledge, conventional ML classifiers have never been used in a federated manner for energy theft detection in SGs. Finally, substantial experiments are performed on the real-world energy consumption dataset. Results illustrate that the proposed model can accurately and efficiently detect energy theft in SGs while guaranteeing the security of client data. Full article
(This article belongs to the Section F5: Artificial Intelligence and Smart Energy)
Show Figures

Figure 1

Figure 1
<p>Federated learning enables smart grids networks for theft detection.</p>
Full article ">Figure 2
<p>Abstract overview of the proposed framework. Highlighted area depicts the proposed novel federated voting classifier (FVC).</p>
Full article ">Figure 3
<p>An illustration of the FedDP architecture. The shaded area represents the working in the specific phase.</p>
Full article ">Figure 4
<p>Functionality of the proposed FVC model.</p>
Full article ">Figure 5
<p>Comparison of log loss between different clients and centralized model.</p>
Full article ">Figure 6
<p>Comparison of average RMSE between different clients and centralized model.</p>
Full article ">Figure 7
<p>Comparison of previous literature with FVC in terms of log loss.</p>
Full article ">Figure 8
<p>Comparison of previous literature with FVC in terms of RMSE.</p>
Full article ">Figure 9
<p>Federated time comparison of FVC with previous research.</p>
Full article ">
Back to TopTop