[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (64)

Search Parameters:
Keywords = on-device learning

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
28 pages, 462 KiB  
Review
A Joint Survey in Decentralized Federated Learning and TinyML: A Brief Introduction to Swarm Learning
by Evangelia Fragkou and Dimitrios Katsaros
Future Internet 2024, 16(11), 413; https://doi.org/10.3390/fi16110413 - 8 Nov 2024
Viewed by 1507
Abstract
TinyML/DL is a new subfield of ML that allows for the deployment of ML algorithms on low-power devices to process their own data. The lack of resources restricts the aforementioned devices to running only inference tasks (static TinyML), while training is handled by [...] Read more.
TinyML/DL is a new subfield of ML that allows for the deployment of ML algorithms on low-power devices to process their own data. The lack of resources restricts the aforementioned devices to running only inference tasks (static TinyML), while training is handled by a more computationally efficient system, such as the cloud. In recent literature, the focus has been on conducting real-time on-device training tasks (Reformable TinyML) while being wirelessly connected. With data processing being shift to edge devices, the development of decentralized federated learning (DFL) schemes becomes justified. Within these setups, nodes work together to train a neural network model, eliminating the necessity of a central coordinator. Ensuring secure communication among nodes is of utmost importance for protecting data privacy during edge device training. Swarm Learning (SL) emerges as a DFL paradigm that promotes collaborative learning through peer-to-peer interaction, utilizing edge computing and blockchain technology. While SL provides a robust defense against adversarial attacks, it comes at a high computational expense. In this survey, we emphasize the current literature regarding both DFL and TinyML/DL fields. We explore the obstacles encountered by resource-starved devices in this collaboration and provide a brief overview of the potential of transitioning to Swarm Learning. Full article
Show Figures

Figure 1

Figure 1
<p>Screening process [<a href="#B17-futureinternet-16-00413" class="html-bibr">17</a>] of relevant academic works.</p>
Full article ">Figure 2
<p>Different topologies in FL: (<b>a</b>) star-shaped topology, (<b>b</b>) ring-shaped topology, and (<b>c</b>) the topology of a mesh network.</p>
Full article ">Figure 3
<p>Objectives in DFL.</p>
Full article ">
24 pages, 9406 KiB  
Article
Lightweight Digit Recognition in Smart Metering System Using Narrowband Internet of Things and Federated Learning
by Vladimir Nikić, Dušan Bortnik, Milan Lukić, Dejan Vukobratović and Ivan Mezei
Future Internet 2024, 16(11), 402; https://doi.org/10.3390/fi16110402 - 31 Oct 2024
Viewed by 2269
Abstract
Replacing mechanical utility meters with digital ones is crucial due to the numerous benefits they offer, including increased time resolution in measuring consumption, remote monitoring capabilities for operational efficiency, real-time data for informed decision-making, support for time-of-use billing, and integration with smart grids, [...] Read more.
Replacing mechanical utility meters with digital ones is crucial due to the numerous benefits they offer, including increased time resolution in measuring consumption, remote monitoring capabilities for operational efficiency, real-time data for informed decision-making, support for time-of-use billing, and integration with smart grids, leading to enhanced customer service, reduced energy waste, and progress towards environmental sustainability goals. However, the cost associated with replacing mechanical meters with their digital counterparts is a key factor contributing to the relatively slow roll-out of such devices. In this paper, we present a low-cost and power-efficient solution for retrofitting the existing metering infrastructure, based on state-of-the-art communication and artificial intelligence technologies. The edge device we developed contains a camera for capturing images of a dial meter, a 32-bit microcontroller capable of running the digit recognition algorithm, and an NB-IoT module with (E)GPRS fallback, which enables nearly ubiquitous connectivity even in difficult radio conditions. Our digit recognition methodology, based on the on-device training and inference, augmented with federated learning, achieves a high level of accuracy (97.01%) while minimizing the energy consumption and associated communication overhead (87 μWh per day on average). Full article
Show Figures

Figure 1

Figure 1
<p>Proposed SM architecture used for old TM retrofitting.</p>
Full article ">Figure 2
<p>The system architecture consists of a collection of deployed SMs that communicate via MNO with the cloud.</p>
Full article ">Figure 3
<p>Distribution of ML models on edge devices, which provides the basis for FL.</p>
Full article ">Figure 4
<p>Design and components of edge device is depicted on the left, whereas right images display fabricated devices.</p>
Full article ">Figure 5
<p>View of metering device through camera lens of the edge device used for creation of datasets.</p>
Full article ">Figure 6
<p>Conversion to B/W image format using a fixed threshold: TH = 128 (<b>left</b>), TH = 192 (<b>right</b>).</p>
Full article ">Figure 7
<p>Initial image taken using camera on edge device, intermediate image after B/W conversion and the final image without artifacts.</p>
Full article ">Figure 8
<p>Scheme used for detecting differences between two digit images.</p>
Full article ">Figure 9
<p>Proposed CNN architecture used for digit recognition.</p>
Full article ">Figure 10
<p>The first test case comprises two distinct scenarios representing different training methodologies, one incorporating federated learning (FL) and the other without its use. (<b>a</b>) Scenario 1, which does not use FL. (<b>b</b>) Scenario 2, which utilizes averaging methodology for FL.</p>
Full article ">Figure 11
<p>Second test case displaying training scheme where second batch of devices is trained based on results of training on first batch.</p>
Full article ">Figure 12
<p>Power consumption profile of image capture + preprocessing + inference.</p>
Full article ">Figure 13
<p>Power consumption profile of data packet transmission via NB-IoT.</p>
Full article ">
14 pages, 1578 KiB  
Article
A Platform of Federated Learning Management for Enhanced Mobile Collaboration
by Farkhod Yusubov and KangYoon Lee
Electronics 2024, 13(20), 4104; https://doi.org/10.3390/electronics13204104 - 18 Oct 2024
Cited by 1 | Viewed by 788
Abstract
Federated learning (FL) has emerged as a crucial technology in today’s data-centric environment, enabling decentralized machine learning while safeguarding user privacy. This study introduces “Federated Learning ML Operations (FedOps) Mobile”, a novel FL framework optimized for the dynamic and heterogeneous ecosystem of mobile [...] Read more.
Federated learning (FL) has emerged as a crucial technology in today’s data-centric environment, enabling decentralized machine learning while safeguarding user privacy. This study introduces “Federated Learning ML Operations (FedOps) Mobile”, a novel FL framework optimized for the dynamic and heterogeneous ecosystem of mobile devices. FedOps Mobile addresses the inherent challenges of FL—such as system scalability, device heterogeneity, and operational efficiency—through advanced on-device training using TensorFlow Lite and CoreML. The framework’s innovative approach includes sophisticated client selection mechanisms that assess device readiness and capabilities, ensuring equitable and efficient participation across the network. Additionally, FedOps Mobile leverages remote device control for seamless task management and continuous learning, all without compromising the user experience. The main contribution of this study is the demonstration that federated learning across heterogeneous devices, especially those using different operating systems, can be both practical and efficient using the FedOps Mobile framework. This was validated through experiments that evaluated three key areas: operational efficiency, model personalization, and resource optimization in multi-device settings. The results showed that the proposed method excels in client selection, energy consumption, and model optimization, establishing a new benchmark for federated learning in diverse and complex environments. Full article
Show Figures

Figure 1

Figure 1
<p>Architecture of FedOps. 1. Platform manager—manages device-specific functionalities and ensures seamless integration and operation of various system components. 2. Network Manager-handles all network-related functionalities, such as registering the client with the FL server and managing network communication. 3. Native Platforms implements machine learning functions, such as Fit(), Evaluate(), and Get weights(), that are crucial for training and evaluating models on the client device. 4. Server side (Microk8s Environment), hosted in a Microk8s environment, plays a pivotal role in managing and coordinating the FL process. 5. FL server, central to the FL process, is the aggregation of models and data analysis. 6. Server Manager manages the overall server operations, including client performance monitoring and data management.</p>
Full article ">Figure 2
<p>Explanation of how FedOps Mobile works after installing the application on mobile. 1. The system administrator registers a new task by sending an FCM message. 2. The online devices send resource information to the FRD. 3. The client selection function is triggered to select devices based on the client selection algorithm 4. Selected device identifications are sent to the FCM. 5. The FCM sends a new message to call selected devices for training. 6. Selected devices connect to the server to obtain the initial global model. 7. The server gives the initial global model to the devices.</p>
Full article ">Figure 3
<p>Performance Increase (Accuracy) Across Rounds for Each Experiment.</p>
Full article ">Figure 4
<p>Energy Consumption Over Rounds for Each Experiment.</p>
Full article ">
18 pages, 18528 KiB  
Article
Data Poisoning Attack against Neural Network-Based On-Device Learning Anomaly Detector by Physical Attacks on Sensors
by Takahito Ino, Kota Yoshida, Hiroki Matsutani and Takeshi Fujino
Sensors 2024, 24(19), 6416; https://doi.org/10.3390/s24196416 - 3 Oct 2024
Viewed by 3105
Abstract
In this paper, we introduce a security approach for on-device learning Edge AIs designed to detect abnormal conditions in factory machines. Since Edge AIs are easily accessible by an attacker physically, there are security risks due to physical attacks. In particular, there is [...] Read more.
In this paper, we introduce a security approach for on-device learning Edge AIs designed to detect abnormal conditions in factory machines. Since Edge AIs are easily accessible by an attacker physically, there are security risks due to physical attacks. In particular, there is a concern that the attacker may tamper with the training data of the on-device learning Edge AIs to degrade the task accuracy. Few risk assessments have been reported. It is important to understand these security risks before considering countermeasures. In this paper, we demonstrate a data poisoning attack against an on-device learning Edge AI. Our attack target is an on-device learning anomaly detection system. The system adopts MEMS accelerometers to measure the vibration of factory machines and detect anomalies. The anomaly detector also adopts a concept drift detection algorithm and multiple models to accommodate multiple normal patterns. For the attack, we used a method in which measurements are tampered with by exposing the MEMS accelerometer to acoustic waves of a specific frequency. The acceleration data falsified by this method were trained on an anomaly detector, and the result was that the abnormal state could not be detected. Full article
Show Figures

Figure 1

Figure 1
<p>Autoencoder-based anomaly detector.</p>
Full article ">Figure 2
<p>Overview of ELM.</p>
Full article ">Figure 3
<p>Overview of concept drift detection algorithm. (<b>a</b>) Trained centroids are sequentially calculated during training. (<b>b</b>) Test centroids are sequentially calculated during inference. (<b>c</b>) When concept drift occurs, the test centroid moves away from the train centroid. (<b>d</b>) When the test centroid exceeds the threshold, a concept drift is detected and a new instance is created. The new instance computes its own train centroid from the latest data (training data).</p>
Full article ">Figure 4
<p>Expected drift rate behavior. (<b>a</b>) Concept drift does not occur; (<b>b</b>) Concept drift occurs.</p>
Full article ">Figure 5
<p>Behavior of multi-instance on-device learning anomaly detector.</p>
Full article ">Figure 6
<p>Behavior of anomaly detector without attack.</p>
Full article ">Figure 7
<p>Behavior of anomaly detector with data poisoning attack.</p>
Full article ">Figure 8
<p>Experimental setup. (<b>a</b>) Overall setup; (<b>b</b>) Cooling fan and speaker.</p>
Full article ">Figure 9
<p>Block diagram of experimental setup.</p>
Full article ">Figure 10
<p>Observed frequency spectrum while both cooling fans are stopped.</p>
Full article ">Figure 11
<p>Relationship between irradiated acoustic wave frequency (in audible range), observed peak frequency, and amplitude.</p>
Full article ">Figure 12
<p>Relationship between irradiated acoustic wave frequency (in ultrasonic range), observed peak frequency, and amplitude.</p>
Full article ">Figure 13
<p>Effects of sound pressure for observed peak amplitude (frequency of acoustic waves: 3000 Hz).</p>
Full article ">Figure 14
<p>Samples of observed data. (<b>a</b>) Normal state; (<b>b</b>) Abnormal state; (<b>c</b>) Poisoned state.</p>
Full article ">Figure 15
<p>Error and drift rate without data poisoning attack.</p>
Full article ">Figure 16
<p>Error and drift rate with data poisoning attack.</p>
Full article ">
33 pages, 14331 KiB  
Article
A Virtual Machine Platform Providing Machine Learning as a Programmable and Distributed Service for IoT and Edge On-Device Computing: Architecture, Transformation, and Evaluation of Integer Discretization
by Stefan Bosse
Algorithms 2024, 17(8), 356; https://doi.org/10.3390/a17080356 - 15 Aug 2024
Viewed by 941
Abstract
Data-driven models used for predictive classification and regression tasks are commonly computed using floating-point arithmetic and powerful computers. We address constraints in distributed sensor networks like the IoT, edge, and material-integrated computing, providing only low-resource embedded computers with sensor data that are acquired [...] Read more.
Data-driven models used for predictive classification and regression tasks are commonly computed using floating-point arithmetic and powerful computers. We address constraints in distributed sensor networks like the IoT, edge, and material-integrated computing, providing only low-resource embedded computers with sensor data that are acquired and processed locally. Sensor networks are characterized by strong heterogeneous systems. This work introduces and evaluates a virtual machine architecture that provides ML as a service layer (MLaaS) on the node level and addresses very low-resource distributed embedded computers (with less than 20 kB of RAM). The VM provides a unified ML instruction set architecture that can be programmed to implement decision trees, ANN, and CNN model architectures using scaled integer arithmetic only. Models are trained primarily offline using floating-point arithmetic, finally converted by an iterative scaling and transformation process, demonstrated in this work by two tests based on simulated and synthetic data. This paper is an extended version of the FedCSIS 2023 conference paper providing new algorithms and ML applications, including ANN/CNN-based regression and classification tasks studying the effects of discretization on classification and regression accuracy. Full article
(This article belongs to the Special Issue Algorithms for Network Systems and Applications)
Show Figures

Figure 1

Figure 1
<p>ARM Cortex M0-based sensor node (STM32L031) implementing the REXA VM for material-integrated GUW sensing with NFC for energy transfer and bidirectional communication with only 8 kB of RAM and 32 kB of ROM.</p>
Full article ">Figure 2
<p>Principle REXA VM network architecture using different wired and wireless communication technologies.</p>
Full article ">Figure 3
<p>Basic REXA-VM architecture with integrated JIT compiler, stacks, and byte-code processor [<a href="#B11-algorithms-17-00356" class="html-bibr">11</a>,<a href="#B12-algorithms-17-00356" class="html-bibr">12</a>].</p>
Full article ">Figure 4
<p>(<b>Left</b>) Incremental growing code segment (single-tasking), persistent code cannot be removed. (<b>Right</b>) Dynamically partitioned code segments using code frames and linking code frames due to fragmentation.</p>
Full article ">Figure 5
<p>Exploding output values for negative x-values (<span class="html-italic">e</span><sup>−x</sup> term) and positive x-values (<span class="html-italic">e</span><sup>x</sup> term) of the exponential function.</p>
Full article ">Figure 6
<p>Relative discretization error of integer-scaled LUT-based approximation of the <span class="html-italic">log10</span> function for different Δ<span class="html-italic">x</span> values (1,2,4) and LUT sizes of 90, 45, and 23, respectively.</p>
Full article ">Figure 7
<p>Relative discretization error of integer-scaled LUT-interpolated approximation of the <span class="html-italic">sigmoid</span> function using the discretized <span class="html-italic">log10</span> LUT-interpolation function for different LUT resolutions and sigmoid segment ranges <span class="html-italic">R</span>. The small error plots show only positive x values.</p>
Full article ">Figure 8
<p>Relative discretization error of integer-scaled LUT-interpolated approximation of the <span class="html-italic">tanh</span> function using the discretized <span class="html-italic">log10</span> LUT-interpolation function.</p>
Full article ">Figure 9
<p>Phase 1 transformation (CNN). (<b>Top</b>) Transformation of 3-dim tensors into multiple vectors for convolutional and pooling layers and flattening of multiple vectors from last convolutional or pooling layer into one vector for the input of a fully connected neuronal layer. (<b>Bottom</b>) Convolutional and pooling operations factorized into sequential and accumulated vector operations.</p>
Full article ">Figure 10
<p>Scaling architectures for (<b>Top</b>) functional nodes, i.e., neurons; (<b>Bottom</b>) convolution or pooling operation.</p>
Full article ">Figure 11
<p>Accumulative scaled convolution or multi-vector input (flattening) neural network operation based on a product–sum calculation. Each accumulative iteration uses a different input scaling <span class="html-italic">s</span><sub>d</sub> normalization with respect to the output scaling <span class="html-italic">s</span>.</p>
Full article ">Figure 12
<p>The ML model transformation pipeline creating an intermediate USM and then creating a sequence of MLISA vector operations.</p>
Full article ">Figure 13
<p>GUW signal simulation using a 2 dim viscoelastic wave propagation model. (<b>Left</b>) Simulation set-up. (<b>Right</b>) Some example signals with and without damage (blue areas show damage features).</p>
Full article ">Figure 14
<p>Down-sampled GUW signal from simulation and low-pass-filtered rectified (envelope approximation) signal as input for the CNN (damage at position x = 100, y = 100).</p>
Full article ">Figure 15
<p>Foo/FooFP model analysis of the GUW regression CNN model. The classification error was always zero.</p>
Full article ">Figure 16
<p>(<b>Top</b>) Analysis of the ANN FP and DS models comparing RMSE and E<sub>max</sub> values for different configurations of the activation function approximation, including an FPU replacement. (<b>Bottom</b>) Selected prediction results are shown with discontinuities in the top plot using ActDS configuration 5 and without using the FPU replacement for the tanh function.</p>
Full article ">
14 pages, 9929 KiB  
Article
Diagnosis of Pressure Ulcer Stage Using On-Device AI
by Yujee Chang, Jun Hyung Kim, Hyun Woo Shin, Changjin Ha, Seung Yeob Lee and Taesik Go
Appl. Sci. 2024, 14(16), 7124; https://doi.org/10.3390/app14167124 - 14 Aug 2024
Viewed by 1384
Abstract
Pressure ulcers are serious healthcare concerns, especially for the elderly with reduced mobility. Severe pressure ulcers are accompanied by pain, degrading patients’ quality of life. Thus, speedy and accurate detection and classification of pressure ulcers are vital for timely treatment. The conventional visual [...] Read more.
Pressure ulcers are serious healthcare concerns, especially for the elderly with reduced mobility. Severe pressure ulcers are accompanied by pain, degrading patients’ quality of life. Thus, speedy and accurate detection and classification of pressure ulcers are vital for timely treatment. The conventional visual examination method requires professional expertise for diagnosing pressure ulcer severity but it is difficult for the lay carer in domiciliary settings. In this study, we present a mobile healthcare platform incorporated with a light-weight deep learning model to exactly detect pressure ulcer regions and classify pressure ulcers into six severities such as stage 1–4, deep tissue pressure injury, and unstageable. YOLOv8 models were trained and tested using 2800 annotated pressure ulcer images. Among the five tested YOLOv8 models, the YOLOv8m model exhibited promising detection performance with overall classification accuracy of 84.6% and a mAP@50 value of 90.8%. The mobile application (app) was also developed applying the trained YOLOv8m model. The mobile app returned the diagnostic result within a short time (≒3 s). Accordingly, the proposed on-device AI app can contribute to early diagnosis and systematic management of pressure ulcers. Full article
Show Figures

Figure 1

Figure 1
<p>YOLOv8 model architecture. <span class="html-italic">d</span>, <span class="html-italic">w</span>, and <span class="html-italic">r</span> indicate the depth multiple, width multiple, and ratio of each module, respectively. <span class="html-italic">k</span>, <span class="html-italic">s</span>, and <span class="html-italic">p</span> denote kernel size, stride, and padding number, respectively.</p>
Full article ">Figure 2
<p>YOLOv8m training results.</p>
Full article ">Figure 3
<p>Confusion matrix of pressure ulcer classification by the trained YOLOv8m model.</p>
Full article ">Figure 4
<p>Representative pressure ulcer detection results by the trained YOLOv8m model. (<b>a</b>–<b>c</b>) Single stage detection in each image. (<b>d</b>–<b>f</b>) Multiple stage detections in each image.</p>
Full article ">Figure 5
<p>Development of a mobile app for detecting pressure ulcers. (<b>a</b>) Log-in page. (<b>b</b>) App home page. (<b>c</b>) Detection results after inspection. (<b>d</b>) Provision of instructions and information according to the results.</p>
Full article ">Figure 6
<p>Failure cases of detecting stage 2. (<b>a</b>,<b>d</b>) Original images. (<b>b</b>,<b>e</b>) Ground–truth labels. (<b>c</b>,<b>f</b>) Prediction results.</p>
Full article ">
15 pages, 977 KiB  
Article
Dynamic Bandwidth Slicing in Passive Optical Networks to Empower Federated Learning
by Alaelddin F. Y. Mohammed, Joohyung Lee and Sangdon Park
Sensors 2024, 24(15), 5000; https://doi.org/10.3390/s24155000 - 2 Aug 2024
Viewed by 922
Abstract
Federated Learning (FL) is a decentralized machine learning method in which individual devices compute local models based on their data. In FL, devices periodically share newly trained updates with the central server, rather than submitting their raw data. The key characteristics of FL, [...] Read more.
Federated Learning (FL) is a decentralized machine learning method in which individual devices compute local models based on their data. In FL, devices periodically share newly trained updates with the central server, rather than submitting their raw data. The key characteristics of FL, including on-device training and aggregation, make it interesting for many communication domains. Moreover, the potential of new systems facilitating FL in sixth generation (6G) enabled Passive Optical Networks (PON), presents a promising opportunity for integration within this domain. This article focuses on the interaction between FL and PON, exploring approaches for effective bandwidth management, particularly in addressing the complexity introduced by FL traffic. In the PON standard, advanced bandwidth management is proposed by allocating multiple upstream grants utilizing the Dynamic Bandwidth Allocation (DBA) algorithm to be allocated for an Optical Network Unit (ONU). However, there is a lack of research on studying the utilization of multiple grant allocation. In this paper, we address this limitation by introducing a novel DBA approach that efficiently allocates PON bandwidth for FL traffic generation and demonstrates how multiple grants can benefit from the enhanced capacity of implementing PON in carrying out FL flows. Simulations conducted in this study show that the proposed solution outperforms state-of-the-art solutions in several network performance metrics, particularly in reducing upstream delay. This improvement holds great promise for enabling real-time data-intensive services that will be key components of 6G environments. Furthermore, our discussion outlines the potential for the integration of FL and PON as an operational reality capable of supporting 6G networking. Full article
(This article belongs to the Section Communications)
Show Figures

Figure 1

Figure 1
<p>FL integration over PON in 6G.</p>
Full article ">Figure 2
<p>System overview.</p>
Full article ">Figure 3
<p>Bandwidth allocation on a DBA cycle <math display="inline"><semantics> <msub> <mi>T</mi> <mi>c</mi> </msub> </semantics></math>.</p>
Full article ">Figure 4
<p>The process model of the (<b>a</b>) Client, (<b>b</b>) ONU, and (<b>c</b>) OLT.</p>
Full article ">Figure 5
<p>Bandwidth allocation example on a <math display="inline"><semantics> <msub> <mi>T</mi> <mi>c</mi> </msub> </semantics></math> for 8 and 16 ONUs in PON.</p>
Full article ">Figure 6
<p>Evaluating the FL upstream delay.</p>
Full article ">Figure 7
<p>Fairness under different numbers on ONUs.</p>
Full article ">
15 pages, 2218 KiB  
Review
A Survey on Neuromorphic Architectures for Running Artificial Intelligence Algorithms
by Seham Al Abdul Wahid, Arghavan Asad and Farah Mohammadi
Electronics 2024, 13(15), 2963; https://doi.org/10.3390/electronics13152963 - 26 Jul 2024
Cited by 1 | Viewed by 2304
Abstract
Neuromorphic computing, a brain-inspired non-Von Neumann computing system, addresses the challenges posed by the Moore’s law memory wall phenomenon. It has the capability to enhance performance while maintaining power efficiency. Neuromorphic chip architecture requirements vary depending on the application and optimising it for [...] Read more.
Neuromorphic computing, a brain-inspired non-Von Neumann computing system, addresses the challenges posed by the Moore’s law memory wall phenomenon. It has the capability to enhance performance while maintaining power efficiency. Neuromorphic chip architecture requirements vary depending on the application and optimising it for large-scale applications remains a challenge. Neuromorphic chips are programmed using spiking neural networks which provide them with important properties such as parallelism, asynchronism, and on-device learning. Widely used spiking neuron models include the Hodgkin–Huxley Model, Izhikevich model, integrate-and-fire model, and spike response model. Hardware implementation platforms of the chip follow three approaches: analogue, digital, or a combination of both. Each platform can be implemented using various memory topologies which interconnect with the learning mechanism. Current neuromorphic computing systems typically use the unsupervised learning spike timing-dependent plasticity algorithms. However, algorithms such as voltage-dependent synaptic plasticity have the potential to enhance performance. This review summarises the potential neuromorphic chip architecture specifications and highlights which applications they are suitable for. Full article
(This article belongs to the Special Issue Neuromorphic Device, Circuits, and Systems)
Show Figures

Figure 1

Figure 1
<p>Four main spiking neuron models used in neuromorphic chips.</p>
Full article ">Figure 2
<p>Neuromorphic architecture characterisation diagram.</p>
Full article ">Figure 3
<p>Neuromorphic computing learning methods characterisation diagram.</p>
Full article ">Figure 4
<p>Von Neumann architecture versus neuromorphic architecture.</p>
Full article ">Figure 5
<p>Example of SNN and information transmission between neurons through synapses.</p>
Full article ">Figure 6
<p>Backpropagation algorithm network structure.</p>
Full article ">Figure 7
<p>Spike-timing-dependent plasticity architecture where the weights are adjusted based on the spike timings of the pre-synaptic neurons (i) and post-synaptic neurons (j).</p>
Full article ">Figure 8
<p>NeuroTower architecture with depiction of stacked memory.</p>
Full article ">
25 pages, 8864 KiB  
Article
A Real-Time and Privacy-Preserving Facial Expression Recognition System Using an AI-Powered Microcontroller
by Jiajin Zhang, Xiaolong Xie, Guoying Peng, Li Liu, Hongyu Yang, Rong Guo, Juntao Cao and Jianke Yang
Electronics 2024, 13(14), 2791; https://doi.org/10.3390/electronics13142791 - 16 Jul 2024
Cited by 1 | Viewed by 1649
Abstract
This study proposes an edge computing-based facial expression recognition system that is low cost, low power, and privacy preserving. It utilizes a minimally obtrusive cap-based system designed for the continuous and real-time monitoring of a user’s facial expressions. The proposed method focuses on [...] Read more.
This study proposes an edge computing-based facial expression recognition system that is low cost, low power, and privacy preserving. It utilizes a minimally obtrusive cap-based system designed for the continuous and real-time monitoring of a user’s facial expressions. The proposed method focuses on detecting facial skin deformations accompanying changes in facial expressions. A multi-zone time-of-flight (ToF) depth sensor VL53L5CX, featuring an 8 × 8 depth image, is integrated into the front brim of the cap to measure the distance between the sensor and the user’s facial skin surface. The distance values corresponding to seven universal facial expressions (neutral, happy, disgust, anger, surprise, fear, and sad) are transmitted to a low-power STM32F476 microcontroller (MCU) as an edge device for data preprocessing and facial expression classification tasks utilizing an on-device pre-trained deep learning model. Performance evaluation of the system is conducted through experiments utilizing data collected from 20 subjects. Four deep learning algorithms, including Convolutional Neural Networks (CNN), Recurrent Neural Networks (RNN), Long Short-Term Memory (LSTM) networks, and Deep Neural Networks (DNN), are assessed. These algorithms demonstrate high accuracy, with CNN yielding the best result, achieving an accuracy of 89.20% at a frame rate of 15 frames per second (fps) and a maximum latency of 2 ms. Full article
Show Figures

Figure 1

Figure 1
<p>System overview, showing the hardware connections and data processing pipeline.</p>
Full article ">Figure 2
<p>VL53L5CX sensor.</p>
Full article ">Figure 3
<p>Facial muscle distances matrix captured by the sensor (8 × 8 Configuration).</p>
Full article ">Figure 4
<p>The facial skin surface detection area of the VL53L5CX.</p>
Full article ">Figure 5
<p>SensorTile kit.</p>
Full article ">Figure 6
<p>Cap with hardware distribution.</p>
Full article ">Figure 7
<p>Depth images for facial expressions. A camera image (<b>left</b>) and a corresponding ToF depth image (<b>right</b>).</p>
Full article ">Figure 7 Cont.
<p>Depth images for facial expressions. A camera image (<b>left</b>) and a corresponding ToF depth image (<b>right</b>).</p>
Full article ">Figure 8
<p>Experimental setting.</p>
Full article ">Figure 9
<p>Structure of the CNN model used for classification.</p>
Full article ">Figure 10
<p>Structure of the Simple RNN model used for classification.</p>
Full article ">Figure 11
<p>Structure of the LSTM model used for classification.</p>
Full article ">Figure 12
<p>Structure of the DNN model used for classification.</p>
Full article ">Figure 13
<p>Deep learning model quantization and porting.</p>
Full article ">Figure 14
<p>Accuracy graphs of the validation datasets.</p>
Full article ">Figure 15
<p>Loss of the validation datasets.</p>
Full article ">Figure 16
<p>CNN model AUC-ROC curve.</p>
Full article ">Figure 17
<p>RNN model AUC-ROC curve.</p>
Full article ">Figure 18
<p>LSTM model AUC-ROC curve.</p>
Full article ">Figure 19
<p>DNN model AUC-ROC curve.</p>
Full article ">Figure 20
<p>Confusion matrix generated using the CNN algorithm.</p>
Full article ">Figure 21
<p>Confusion matrix generated using the RNN algorithm.</p>
Full article ">Figure 22
<p>Confusion matrix generated using the LSTM algorithm.</p>
Full article ">Figure 23
<p>Confusion matrix generated using the DNN algorithm.</p>
Full article ">Figure 24
<p>Screenshots of smartphone application. (<b>a</b>) Recognition page; (<b>b</b>) Recording page.</p>
Full article ">Figure 25
<p>Watching films.</p>
Full article ">Figure 26
<p>Riding a bus.</p>
Full article ">Figure 27
<p>Walking.</p>
Full article ">
20 pages, 8592 KiB  
Article
On-Device Semi-Supervised Activity Detection: A New Privacy-Aware Personalized Health Monitoring Approach
by Avirup Roy, Hrishikesh Dutta, Amit Kumar Bhuyan and Subir Biswas
Sensors 2024, 24(14), 4444; https://doi.org/10.3390/s24144444 - 9 Jul 2024
Viewed by 1056
Abstract
This paper presents an on-device semi-supervised human activity detection system that can learn and predict human activity patterns in real time. The clinical objective is to monitor and detect the unhealthy sedentary lifestyle of a user. The proposed semi-supervised learning (SSL) framework uses [...] Read more.
This paper presents an on-device semi-supervised human activity detection system that can learn and predict human activity patterns in real time. The clinical objective is to monitor and detect the unhealthy sedentary lifestyle of a user. The proposed semi-supervised learning (SSL) framework uses sparsely labelled user activity events acquired from Inertial Measurement Unit sensors installed as wearable devices. The proposed cluster-based learning model in this approach is trained with data from the same target user, thus preserving data privacy while providing personalized activity detection services. Two different cluster labelling strategies, namely, population-based and distance-based strategies, are employed to achieve the desired classification performance. The proposed system is shown to be highly accurate and computationally efficient for different algorithmic parameters, which is relevant in the context of limited computing resources on typical wearable devices. Extensive experimentation and simulation study have been conducted on multi-user human activity data from the public domain in order to analyze the trade-off between classification accuracy and computation complexity of the proposed learning paradigm with different algorithmic hyper-parameters. With 4.17 h of training time for 8000 activity episodes, the proposed SSL approach consumes at most 20 KB of CPU memory space, while providing a maximum accuracy of 90% and 100% classification rates. Full article
(This article belongs to the Special Issue IMU and Innovative Sensors for Healthcare)
Show Figures

Figure 1

Figure 1
<p>A wearable on-device human activity detection system.</p>
Full article ">Figure 2
<p>Feature distribution of the three classes.</p>
Full article ">Figure 3
<p>Pre-processing and classification pipeline.</p>
Full article ">Figure 4
<p>Iterative semi-supervised learning framework.</p>
Full article ">Figure 5
<p>Example cluster model evolution with incoming episodes self-trained using iterative semi-supervised learning paradigm.</p>
Full article ">Figure 6
<p>(<b>a</b>–<b>h</b>) Impacts of PCA on accuracy parameters for population-based SSL; (<b>a</b>–<b>c</b>): true positive for the three classes; (<b>d</b>–<b>f</b>) false positive for the three classes; (<b>g</b>) overall accuracy; (<b>h</b>) classification rate.</p>
Full article ">Figure 7
<p>(<b>a</b>–<b>h</b>) Impacts of PCA on accuracy parameters for distance-based SSL; (<b>a</b>–<b>c</b>) true positive for the three classes; (<b>d</b>–<b>f</b>) false positive for the three classes; (<b>g</b>) overall accuracy; (<b>h</b>) classification rate.</p>
Full article ">Figure 8
<p>Impact of PCA on complexity for population-based SSL; (<b>a</b>) computational time; (<b>b</b>) CPU memory usage.</p>
Full article ">Figure 9
<p>Impact of PCA on complexity for distance-based SSL; (<b>a</b>) computational time; (<b>b</b>) CPU memory usage.</p>
Full article ">Figure 10
<p>Post-convergence accuracy parameters’ results for (<b>a</b>–<b>c</b>) population-based SSL and (<b>d</b>–<b>f</b>) distance-based SSL, with varying <math display="inline"><semantics> <mrow> <mi>α</mi> </mrow> </semantics></math>.</p>
Full article ">Figure 11
<p>Total learning time and CPU memory usage for all the episodes using (<b>a</b>) population-based SSL and (<b>b</b>) distance-based SSL, with varying <math display="inline"><semantics> <mrow> <mi>α</mi> </mrow> </semantics></math>.</p>
Full article ">Figure 12
<p>Post-convergence accuracy parameters’ results for (<b>a</b>–<b>c</b>) population-based SSL and (<b>d</b>–<b>f</b>) distance-based SSL, with varying <span class="html-italic">Nc</span>.</p>
Full article ">Figure 13
<p>Total learning time and CPU memory usage for all the episodes using (<b>a</b>) population-based SSL and (<b>b</b>) distance-based SSL, with varying <math display="inline"><semantics> <mrow> <msub> <mrow> <mi>N</mi> </mrow> <mrow> <mi>c</mi> </mrow> </msub> </mrow> </semantics></math>.</p>
Full article ">Figure 14
<p>Post-convergence accuracy parameters’ results for (<b>a</b>–<b>c</b>) population-based SSL and (<b>d</b>–<b>f</b>) distance-based SSL for different clustering methods.</p>
Full article ">Figure 15
<p>Total learning time and CPU memory usage for 8262 episodes using (<b>a</b>) population-based SSL and (<b>b</b>) distance-based SSL for different clustering methods.</p>
Full article ">
20 pages, 1613 KiB  
Article
Energy-Efficient Edge and Cloud Image Classification with Multi-Reservoir Echo State Network and Data Processing Units
by E. J. López-Ortiz, M. Perea-Trigo, L. M. Soria-Morillo, J. A. Álvarez-García and J. J. Vegas-Olmos
Sensors 2024, 24(11), 3640; https://doi.org/10.3390/s24113640 - 4 Jun 2024
Viewed by 794
Abstract
In an era dominated by Internet of Things (IoT) devices, software-as-a-service (SaaS) platforms, and rapid advances in cloud and edge computing, the demand for efficient and lightweight models suitable for resource-constrained devices such as data processing units (DPUs) has surged. Traditional deep learning [...] Read more.
In an era dominated by Internet of Things (IoT) devices, software-as-a-service (SaaS) platforms, and rapid advances in cloud and edge computing, the demand for efficient and lightweight models suitable for resource-constrained devices such as data processing units (DPUs) has surged. Traditional deep learning models, such as convolutional neural networks (CNNs), pose significant computational and memory challenges, limiting their use in resource-constrained environments. Echo State Networks (ESNs), based on reservoir computing principles, offer a promising alternative with reduced computational complexity and shorter training times. This study explores the applicability of ESN-based architectures in image classification and weather forecasting tasks, using benchmarks such as the MNIST, FashionMnist, and CloudCast datasets. Through comprehensive evaluations, the Multi-Reservoir ESN (MRESN) architecture emerges as a standout performer, demonstrating its potential for deployment on DPUs or home stations. In exploiting the dynamic adaptability of MRESN to changing input signals, such as weather forecasts, continuous on-device training becomes feasible, eliminating the need for static pre-trained models. Our results highlight the importance of lightweight models such as MRESN in cloud and edge computing applications where efficiency and sustainability are paramount. This study contributes to the advancement of efficient computing practices by providing novel insights into the performance and versatility of MRESN architectures. By facilitating the adoption of lightweight models in resource-constrained environments, our research provides a viable alternative for improved efficiency and scalability in modern computing paradigms. Full article
(This article belongs to the Special Issue Feature Papers in the 'Sensor Networks' Section 2024)
Show Figures

Figure 1

Figure 1
<p>ESN training. The symbol <math display="inline"><semantics> <mrow> <mo>+</mo> <mspace width="-0.166667em"/> <mspace width="-0.166667em"/> <mspace width="-0.166667em"/> <mo>+</mo> </mrow> </semantics></math> is used to express a vector concatenation. <math display="inline"><semantics> <mi>λ</mi> </semantics></math> is used to express the computation of the new state. The length of training is expressed by T.</p>
Full article ">Figure 2
<p>ESN Test phase. The output layer processes the new input beside the state it produces in the network to obtain the predicted output. Symbol <math display="inline"><semantics> <mrow> <mo>+</mo> <mspace width="-0.166667em"/> <mspace width="-0.166667em"/> <mspace width="-0.166667em"/> <mo>+</mo> </mrow> </semantics></math> is used to express vector concatenation, while ⨂ is used to express matrix multiplication.</p>
Full article ">Figure 3
<p>ESN with specialized output layers. Argmax or softmax could be used as the final step. Symbol <math display="inline"><semantics> <mrow> <mo>+</mo> <mspace width="-0.166667em"/> <mspace width="-0.166667em"/> <mspace width="-0.166667em"/> <mo>+</mo> </mrow> </semantics></math> is used to express vector concatenation, while ⨂ is used to express matrix multiplication.</p>
Full article ">Figure 4
<p>MRESN architecture. Test phase. Symbol <math display="inline"><semantics> <mrow> <mo>+</mo> <mspace width="-0.166667em"/> <mspace width="-0.166667em"/> <mspace width="-0.166667em"/> <mo>+</mo> </mrow> </semantics></math> is used to express vector concatenation, while ⨂ is used to express matrix multiplication.</p>
Full article ">Figure 5
<p>(<b>a</b>) MNIST, (<b>b</b>) FashionMNIST.</p>
Full article ">Figure 6
<p>CloudCast sample 728 × 728 pixels. Europe region. Coloured regions correspond to the first four classes.</p>
Full article ">Figure 7
<p>Execution times on CPU vs. GPU vs. DPU for 8 × 1000 nodes MRESN.</p>
Full article ">Figure 8
<p>Effect of neighboring pixels on the target pixel over time. The dashed lines represent successive time steps. Certain pixels exert immediate influence on the target (green cell), while others require multiple steps for their information to reach the target.</p>
Full article ">Figure 9
<p>ESN grid search for CloudCast dataset. Big influence of spectral radius on results.</p>
Full article ">Figure 10
<p>PSO results for ESN architecture.</p>
Full article ">Figure 11
<p>PSO over MRESN architecture.</p>
Full article ">
13 pages, 1561 KiB  
Article
Forward Learning of Large Language Models by Consumer Devices
by Danilo Pietro Pau and Fabrizio Maria Aymone
Electronics 2024, 13(2), 402; https://doi.org/10.3390/electronics13020402 - 18 Jan 2024
Cited by 1 | Viewed by 2671
Abstract
Large Language Models achieve state of art performances on a broad variety of Natural Language Processing tasks. In the pervasive IoT era, their deployment on edge devices is more compelling than ever. However, their gigantic model footprint has hindered on-device learning applications which [...] Read more.
Large Language Models achieve state of art performances on a broad variety of Natural Language Processing tasks. In the pervasive IoT era, their deployment on edge devices is more compelling than ever. However, their gigantic model footprint has hindered on-device learning applications which enable AI models to continuously learn and adapt to changes over time. Back-propagation, in use by the majority of deep learning frameworks, is computationally intensive and requires storing intermediate activations into memory to cope with the model’s weights update. Recently, “Forward-only algorithms” have been proposed since they are biologically plausible alternatives. By applying more “forward” passes, this class of algorithms can achieve memory reductions with respect to more naive forward-only approaches and by removing the need to store intermediate activations. This comes at the expense of increased computational complexity. This paper considered three Large Language Model: DistilBERT, GPT-3 Small and AlexaTM. It investigated quantitatively any improvements about memory usage and computational complexity brought by known approaches named PEPITA and MEMPEPITA with respect to backpropagation. For low number of tokens in context, and depending on the model, PEPITA increases marginally or reduces substantially arithmetic operations. On the other hand, for large number of tokens in context, PEPITA reduces computational complexity by 30% to 50%. MEMPEPITA increases PEPITA’s complexity by one third. About memory, PEPITA and backpropagation, require a comparable amount of memory to store activations, while MEMPEPITA reduces it by 50% to 94% with the benefits being more evident for architectures with a long sequence of blocks. In various real case scenarios, MEMPEPITA’s memory reduction was essential for meeting the tight memory requirements of 128 MB equipped edge consumer devices, which are commonly available as smartphone and industrial application multi processors. Full article
Show Figures

Figure 1

Figure 1
<p>Computational and Memory complexity of the three LLMs analysed. Subfigures (<b>a</b>–<b>i</b>) depict respectively the MACCs, FLOPs and activations’ footprint of DistilBert, GPT-3 Small and AlexaTM when trained with BP, PEP and MPEP.</p>
Full article ">Figure 2
<p>Latency estimation on the three LLMs.</p>
Full article ">Figure 3
<p>Memory footprint estimation on the three LLMs.</p>
Full article ">
18 pages, 5560 KiB  
Article
Towards Full Forward On-Tiny-Device Learning: A Guided Search for a Randomly Initialized Neural Network
by Danilo Pau, Andrea Pisani and Antonio Candelieri
Algorithms 2024, 17(1), 22; https://doi.org/10.3390/a17010022 - 5 Jan 2024
Viewed by 2144
Abstract
In the context of TinyML, many research efforts have been devoted to designing forward topologies to support On-Device Learning. Reaching this target would bring numerous advantages, including reductions in latency and computational complexity, stronger privacy, data safety and robustness to adversarial attacks, higher [...] Read more.
In the context of TinyML, many research efforts have been devoted to designing forward topologies to support On-Device Learning. Reaching this target would bring numerous advantages, including reductions in latency and computational complexity, stronger privacy, data safety and robustness to adversarial attacks, higher resilience against concept drift, etc. However, On-Device Learning on resource constrained devices poses severe limitations to computational power and memory. Therefore, deploying Neural Networks on tiny devices appears to be prohibitive, since their backpropagation-based training is too memory demanding for their embedded assets. Using Extreme Learning Machines based on Convolutional Neural Networks might be feasible and very convenient, especially for Feature Extraction tasks. However, it requires searching for a randomly initialized topology that achieves results as good as those achieved by the backpropagated model. This work proposes a novel approach for automatically composing an Extreme Convolutional Feature Extractor, based on Neural Architecture Search and Bayesian Optimization. It was applied to the CIFAR-10 and MNIST datasets for evaluation. Two search spaces have been defined, as well as a search strategy that has been tested with two surrogate models, Gaussian Process and Random Forest. A performance estimation strategy was defined, keeping the feature set computed by the MLCommons-Tiny benchmark ResNet as a reference model. In as few as 1200 search iterations, the proposed strategy was able to achieve a topology whose extracted features scored a mean square error equal to 0.64 compared to the reference set. Further improvements are required, with a target of at least one order of magnitude decrease in mean square error for improved classification accuracy. The code is made available via GitHub to allow for the reproducibility of the results reported in this paper. Full article
Show Figures

Figure 1

Figure 1
<p>Atomic block for type 1 neural topology.</p>
Full article ">Figure 2
<p>Atomic block for type 2 neural topology. The numbers in the chart indicate the order for the subsequent subtraction operation.</p>
Full article ">Figure 3
<p>Flowchart describing the experimental setup.</p>
Full article ">Figure 4
<p>Progression of the search over the CIFAR-10 dataset, with type 1 neural topology and GP surrogate model.</p>
Full article ">Figure 5
<p>Growth of the optimum over number of iterations for the search with type 1 neural topology and GP surrogate model applied to the CIFAR-10 dataset.</p>
Full article ">Figure 6
<p>Progression of the search over the MNIST dataset with type 1 neural topology and GP surrogate model.</p>
Full article ">Figure 7
<p>Growth of the optimum over number of iterations for the search with type 1 neural topology and GP surrogate model applied to the MNIST dataset.</p>
Full article ">Figure 8
<p>Progression of the search with type 1 neural topology and RF surrogate model.</p>
Full article ">Figure 9
<p>Growth of the optimum found over the number of iterations for the search with type 1 neural topology and RF surrogate model.</p>
Full article ">Figure 10
<p>Schema representing the best performing topology found in all searches for the CIFAR-10 dataset. Every group of three squares represents a block; the first square describes its kernel size, the second its number of filters and the third represents the presence/absence of regularization within the block. See <a href="#algorithms-17-00022-f001" class="html-fig">Figure 1</a> for the composition of a singular block.</p>
Full article ">Figure 11
<p>Progression of the search over the MNIST dataset with type 1 neural topology and RF surrogate model.</p>
Full article ">Figure 12
<p>Growth of the optimum over number of iterations for the search with type 1 neural topology and RF surrogate model applied to the MNIST dataset.</p>
Full article ">Figure 13
<p>Progression of the search with type 2 neural topology and RF surrogate model.</p>
Full article ">Figure 14
<p>Growth of the optimum over number of iterations for the search with type 2 neural topology and RF surrogate model.</p>
Full article ">Figure 15
<p>Progression of the search over the MNIST dataset with type 2 neural topology and RF surrogate model.</p>
Full article ">Figure 16
<p>Growth of the optimum over number of iterations for the search with type 2 neural topology and RF surrogate model applied to the MNIST dataset.</p>
Full article ">Figure 17
<p>Schema representing the best performing topology found in all searches for the MNIST dataset. Every group of three squares represents a block; the first square describes its kernel size, the second its number of filters and the third represents the presence/absence of regularization within the block. See <a href="#algorithms-17-00022-f002" class="html-fig">Figure 2</a> for the composition of a singular block.</p>
Full article ">
17 pages, 2303 KiB  
Article
The Fusion of Wide Field Optical Coherence Tomography and AI: Advancing Breast Cancer Surgical Margin Visualization
by Yanir Levy, David Rempel, Mark Nguyen, Ali Yassine, Maggie Sanati-Burns, Payal Salgia, Bryant Lim, Sarah L. Butler, Andrew Berkeley and Ersin Bayram
Life 2023, 13(12), 2340; https://doi.org/10.3390/life13122340 - 14 Dec 2023
Cited by 2 | Viewed by 3382
Abstract
This study explores the integration of Wide Field Optical Coherence Tomography (WF-OCT) with an AI-driven clinical decision support system, with the goal of enhancing productivity and decision making in breast cancer surgery margin assessment. A computationally efficient convolutional neural network (CNN)-based binary classifier [...] Read more.
This study explores the integration of Wide Field Optical Coherence Tomography (WF-OCT) with an AI-driven clinical decision support system, with the goal of enhancing productivity and decision making in breast cancer surgery margin assessment. A computationally efficient convolutional neural network (CNN)-based binary classifier is developed using 585 WF-OCT margin scans from 151 subjects. The CNN model swiftly identifies suspicious areas within margins with an on-device inference time of approximately 10 ms for a 420 × 2400 image. In independent testing on 155 pathology-confirmed margins, including 31 positive margins from 29 patients, the classifier achieved an AUROC of 0.976, a sensitivity of 0.93, and a specificity of 0.98. At the margin level, the deep learning model accurately identified 96.8% of pathology-positive margins. These results highlight the clinical viability of AI-enhanced margin visualization using WF-OCT in breast cancer surgery and its potential to decrease reoperation rates due to residual tumors. Full article
Show Figures

Figure 1

Figure 1
<p>WF-OCT image of breast tissue (<b>top</b>) and the corresponding digital pathology image (<b>bottom</b>). The arrow in the pathology image points to ductal carcinoma in situ (DCIS), and the same DCIS is clearly visible in the WF-OCT image.</p>
Full article ">Figure 2
<p>Workflow for model development and performance assessment.</p>
Full article ">Figure 3
<p>Break-down of the training, validation, and test datasets at margin level along with the total statistics. Training and validation sets are used to train and fine-tune the model, while the test set is blinded to the model for independent performance verification.</p>
Full article ">Figure 4
<p>Schematic representation of the WF-OCT imaging framework, showing the hierarchical relationship of the margin (red arrow), composed of sequential WF-Bscans (orange arrows), and a patch (blue box) formed by a sliding window (yellow arrow) over a B-scan.</p>
Full article ">Figure 5
<p>High-level data labeling workflow using a customized validated labeling tool.</p>
Full article ">Figure 6
<p>Architecture of the CNN in ImgAssist<sup>TM</sup>.</p>
Full article ">Figure 7
<p>A composite diagram illustrating the multifaceted image analysis process: (<b>A</b>) demonstrates the clustering algorithm, retaining only adjacent patches (green) exceeding a set classification threshold, non-suspicious (red) or single (yellow) patches are discarded. (<b>B</b>) Details the selection of a ‘Key Thumbnail’ using a moving average maximum (MA<sub>MAX</sub>) method, which identifies the top three contiguous patches with the highest average probability in a cluster; the patch with the maximum local value within this subset is then designated as the ‘Key Thumbnail’. (<b>C</b>) Displays the Thumbnail Display Page on the OCT device’s user interface (UI), where clusters with higher confidence are prioritized at the top. ‘Key Thumbnails’ serve as the most representative image of a cluster, providing clinicians with a concise ‘highlight reel’ of suspicious areas within a margin, thereby streamlining the review process, minimizing information overload, and reducing clinician fatigue.</p>
Full article ">Figure 8
<p>The suspicious thumbnail image on the left is followed by the gradient-weighted Class Activation Maps, which uses the global average of the gradients flowing into the feature maps of the last convolutional layer, a measure that focuses on which features in the image are contributing to the model prediction. The accompanying heatmap overlay on the right provides transparency to the model’s decision making.</p>
Full article ">
21 pages, 1494 KiB  
Article
A Privacy and Energy-Aware Federated Framework for Human Activity Recognition
by Ahsan Raza Khan, Habib Ullah Manzoor, Fahad Ayaz, Muhammad Ali Imran and Ahmed Zoha
Sensors 2023, 23(23), 9339; https://doi.org/10.3390/s23239339 - 22 Nov 2023
Cited by 7 | Viewed by 1656
Abstract
Human activity recognition (HAR) using wearable sensors enables continuous monitoring for healthcare applications. However, the conventional centralised training of deep learning models on sensor data poses challenges related to privacy, communication costs, and on-device efficiency. This paper proposes a federated learning framework integrating [...] Read more.
Human activity recognition (HAR) using wearable sensors enables continuous monitoring for healthcare applications. However, the conventional centralised training of deep learning models on sensor data poses challenges related to privacy, communication costs, and on-device efficiency. This paper proposes a federated learning framework integrating spiking neural networks (SNNs) with long short-term memory (LSTM) networks for energy-efficient and privacy-preserving HAR. The hybrid spiking-LSTM (S-LSTM) model synergistically combines the event-driven efficiency of SNNs and the sequence modelling capability of LSTMs. The model is trained using surrogate gradient learning and backpropagation through time, enabling fully supervised end-to-end learning. Extensive evaluations of two public datasets demonstrate that the proposed approach outperforms LSTM, CNN, and S-CNN models in accuracy and energy efficiency. For instance, the proposed S-LSTM achieved an accuracy of 97.36% and 89.69% for indoor and outdoor scenarios, respectively. Furthermore, the results also showed a significant improvement in energy efficiency of 32.30%, compared to simple LSTM. Additionally, we highlight the significance of personalisation in HAR, where fine-tuning with local data enhances model accuracy by up to 9% for individual users. Full article
(This article belongs to the Special Issue Human Movement Monitoring Using Wearable Sensor Technology)
Show Figures

Figure 1

Figure 1
<p>Conceptual framework of centralised indoor HAR using wearable sensors.</p>
Full article ">Figure 2
<p>Conceptual FL framework for HAR using wearable sensing in the outdoors.</p>
Full article ">Figure 3
<p>Spiking neurons propagation process.</p>
Full article ">Figure 4
<p>Proposed hybrid S-LSTM model where input LSTM layer activated by LIF.</p>
Full article ">Figure 5
<p>Learning curve for UCI-dataset, trained for 500 communication rounds.</p>
Full article ">Figure 6
<p>The confusion matrix for four DL models compared in this study. The index represents the activity where the label corresponding to the activities are (1) walking, (2) walking upstairs, (3) walking downstairs, (4) sitting, (5) standing, and (6) lying.</p>
Full article ">Figure 7
<p>Learning curve for Real-World dataset, trained for 500 communication rounds.</p>
Full article ">Figure 8
<p>The confusion matrix for four DL models compared for Real-World data set. The index represents the activity where the label corresponding to the activities are: (1) climbing down, (2) climbing up, (3) jumping, (4) lying, (5) running, (6) sitting, (7) standing, (8) walking.</p>
Full article ">Figure 9
<p>Learning curve for Real-World dataset, with 50% random choosing of participant trained for 500 communication rounds.</p>
Full article ">Figure 10
<p>Accuracy comparison graph for global and personalised models for each client using the local test set. The personalised accuracy was obtained after fine-tuning using the local dataset.</p>
Full article ">
Back to TopTop