[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (256)

Search Parameters:
Keywords = one-class

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
18 pages, 3792 KiB  
Article
Dynamic Classifier Auditing by Unsupervised Anomaly Detection Methods: An Application in Packaging Industry Predictive Maintenance
by Fernando Mateo, Joan Vila-Francés, Emilio Soria-Olivas, Marcelino Martínez-Sober, Juan Gómez-Sanchis and Antonio José Serrano-López
Appl. Sci. 2025, 15(2), 882; https://doi.org/10.3390/app15020882 (registering DOI) - 17 Jan 2025
Viewed by 239
Abstract
Predictive maintenance in manufacturing industry applications is a challenging research field. Packaging machines are widely used in a large number of logistic companies’ warehouses and must be working uninterruptedly. Traditionally, preventive maintenance strategies have been carried out to improve the performance of these [...] Read more.
Predictive maintenance in manufacturing industry applications is a challenging research field. Packaging machines are widely used in a large number of logistic companies’ warehouses and must be working uninterruptedly. Traditionally, preventive maintenance strategies have been carried out to improve the performance of these machines. However, these kinds of policies do not take into account the information provided by the sensors implemented in the machines. This paper presents an expert system for the automatic estimation of work orders to implement predictive maintenance policies for packaging machines. The central innovation lies in a two-stage process: a classifier generates a binary decision on whether a machine requires maintenance, and an unsupervised anomaly detection module subsequently audits the classifier’s probabilistic output to refine and interpret its predictions. By leveraging the classifier to condense sensor data and applying anomaly detection to its output, the system optimizes the decision reliability. Three anomaly detection methods were evaluated: One-Class Support Vector Machine (OCSVM), Minimum Covariance Determinant (MCD), and a majority (hard) voting ensemble of the two. All anomaly detection methods improved the baseline classifier’s performance, with the majority voting ensemble achieving the highest F1 score. Full article
(This article belongs to the Special Issue Artificial Intelligence Applications in Industry)
Show Figures

Figure 1

Figure 1
<p>Automatic fixed wrapping machine F-2200 (Aranco, Valencia, Spain). Source: <a href="https://www.aranco.com/en/services/sie-wrappers/fixed-wrapping-machine-f-2200" target="_blank">https://www.aranco.com/en/services/sie-wrappers/fixed-wrapping-machine-f-2200</a> (accessed on 16 December 2024).</p>
Full article ">Figure 2
<p>Predictive maintenance expert system. The main contribution of this work is the improvement of the existing classification module by means of unsupervised AD techniques (AD module).</p>
Full article ">Figure 3
<p>Anomaly detection streaming procedure for online auditing of the classifier.</p>
Full article ">Figure 4
<p>Proposed architecture for the AD module.</p>
Full article ">Figure 5
<p>Auditing example for one of the machines using OCSVM. The output from the classifier and the AD model output are represented as different lines, and the shaded part of each signal indicates if the threshold to call for a work order has been exceeded.</p>
Full article ">Figure 6
<p>Box plots illustrating the differences between the F1 scores distributions obtained by the compared methods.</p>
Full article ">
25 pages, 2548 KiB  
Article
Efficient Real-Time Anomaly Detection in IoT Networks Using One-Class Autoencoder and Deep Neural Network
by Aya G. Ayad, Mostafa M. El-Gayar, Noha A. Hikal and Nehal A. Sakr
Electronics 2025, 14(1), 104; https://doi.org/10.3390/electronics14010104 - 30 Dec 2024
Viewed by 523
Abstract
In the face of growing Internet of Things (IoT) security challenges, traditional Intrusion Detection Systems (IDSs) fall short due to IoT devices’ unique characteristics and constraints. This paper presents an effective, lightweight detection model that strengthens IoT security by addressing the high dimensionality [...] Read more.
In the face of growing Internet of Things (IoT) security challenges, traditional Intrusion Detection Systems (IDSs) fall short due to IoT devices’ unique characteristics and constraints. This paper presents an effective, lightweight detection model that strengthens IoT security by addressing the high dimensionality of IoT data. This model merges an asymmetric stacked autoencoder with a Deep Neural Network (DNN), applying one-class learning. It achieves a high detection rate with minimal false positives in a short time. Compared with state-of-the-art approaches based on the BoT-IoT dataset, it shows a higher detection rate of up to 96.27% in 0.27 s. Also, the model achieves an accuracy of 99.99%, precision of 99.21%, and f1 score of 97.69%. These results demonstrate the effectiveness and significance of the proposed model, confirming its potential for reliable deployment in real IoT security problems. Full article
(This article belongs to the Special Issue AI in Cybersecurity, 2nd Edition)
Show Figures

Figure 1

Figure 1
<p>Framework of the proposed model.</p>
Full article ">Figure 2
<p>Structure of autoencoder.</p>
Full article ">Figure 3
<p>Structure of asymmetric stacked autoencoder.</p>
Full article ">Figure 4
<p>Determination of the optimal threshold based on the intersection points between precision and recall curve.</p>
Full article ">Figure 5
<p>Example subset of the BoT-IoT dataset.</p>
Full article ">Figure 6
<p>Testbed setup for the Bot-IoT dataset.</p>
Full article ">Figure 7
<p>In-depth analysis of class-wise metrics for (<b>a</b>) DNN, (<b>b</b>) Sigmoid, (<b>c</b>) OCSVM, (<b>d</b>) IF.</p>
Full article ">Figure 8
<p>Roc curve for the proposed model using 20-dimensional feature space.</p>
Full article ">Figure 9
<p>Confusion matrix for the proposed model using 20-dimensional feature space.</p>
Full article ">
26 pages, 2059 KiB  
Article
Continual Semi-Supervised Malware Detection
by Matthew Chin and Roberto Corizzo
Mach. Learn. Knowl. Extr. 2024, 6(4), 2829-2854; https://doi.org/10.3390/make6040135 - 10 Dec 2024
Viewed by 812
Abstract
Detecting malware has become extremely important with the increasing exposure of computational systems and mobile devices to online services. However, the rapidly evolving nature of malicious software makes this task particularly challenging. Despite the significant number of machine learning works for malware detection [...] Read more.
Detecting malware has become extremely important with the increasing exposure of computational systems and mobile devices to online services. However, the rapidly evolving nature of malicious software makes this task particularly challenging. Despite the significant number of machine learning works for malware detection proposed in the last few years, limited interest has been devoted to continual learning approaches, which could allow models to showcase effective performance in challenging and dynamic scenarios while being computationally efficient. Moreover, most of the research works proposed thus far adopt a fully supervised setting, which relies on fully labelled data and appears to be impractical in a rapidly evolving malware landscape. In this paper, we address malware detection from a continual semi-supervised one-class learning perspective, which only requires normal/benign data and empowers models with a greater degree of flexibility, allowing them to detect multiple malware types with different morphology. Specifically, we assess the effectiveness of two replay strategies on anomaly detection models and analyze their performance in continual learning scenarios with three popular malware detection datasets (CIC-AndMal2017, CIC-MalMem-2022, and CIC-Evasive-PDFMal2022). Our evaluation shows that replay-based strategies can achieve competitive performance in terms of continual ROC-AUC with respect to the considered baselines and bring new perspectives and insights on this topic. Full article
Show Figures

Figure 1

Figure 1
<p>Semi-supervised one-class continual malware detection workflow consisting of model training and evaluation phases. Training concepts <math display="inline"><semantics> <msub> <mi>C</mi> <mi>i</mi> </msub> </semantics></math> contain exclusively normal data, whereas evaluation concepts <math display="inline"><semantics> <msub> <mi>E</mi> <mi>i</mi> </msub> </semantics></math> contain normal and anomalous data points. The experience replay component updates the replay buffer <span class="html-italic">R</span> each time a new training concept <math display="inline"><semantics> <msub> <mi>C</mi> <mi>i</mi> </msub> </semantics></math> is presented, according to the budget <span class="html-italic">B</span>. Two strategies are used: <span class="html-italic">Random</span> and <span class="html-italic">Selective</span>.</p>
Full article ">Figure 2
<p>Example of model evaluation in continual malware detection: concept-level ROC-AUC as a heatmap (<b>a</b>) and as a line plot showing performance over time (<b>b</b>)—CIC-MalMem-2022 dataset—Strategy: Cumulative—Scenario: A—Model: LOF. The results show that learning a new concept without forgetting previous concepts leads to a more comprehensive and robust model, which results in a performance improvement on all other concepts.</p>
Full article ">Figure 3
<p>Continual ROC-AUC performance (CIC-MalMem-2022 dataset) with different strategies (naive: left; ER—Best-performing variant: center; cumulative: right) on single tasks/concepts after learning each task in two scenarios (A, C) with the best-performing one-class model (LOF).</p>
Full article ">Figure 4
<p>Continual ROC-AUC performance (CIC-Evasive-PDFMal2022 dataset) with different strategies (naive: left; ER—Best-performing variant: center; cumulative: right) on single tasks/concepts after learning each task in two scenarios (A, C) with the best-performing one-class model (ABOD).</p>
Full article ">Figure 5
<p>Continual ROC-AUC performance (CIC-AndMal2017 dataset) with different strategies (naive: left; ER—Best-performing variant: center; cumulative: right) on single tasks/concepts after learning each task in two scenarios (A, C) with the best-performing one-class models (ABOD and LOF, respectively).</p>
Full article ">Figure A1
<p>Visualization of extracted concepts via t-SNE: CIC-MalMem-2022 dataset. Normal class (<b>left</b>) and anomaly class (<b>right</b>).</p>
Full article ">Figure A2
<p>Visualization of extracted concepts via t-SNE: CIC-Evasive-PDFMal2022 dataset. Normal class (<b>left</b>) and anomaly class (<b>right</b>).</p>
Full article ">Figure A3
<p>Visualization of extracted concepts via t-SNE: CIC-AndMal-2017 dataset. Normal class (<b>left</b>) and anomaly class (<b>right</b>).</p>
Full article ">
21 pages, 4145 KiB  
Article
UniFlow: Unified Normalizing Flow for Unsupervised Multi-Class Anomaly Detection
by Jianmei Zhong and Yanzhi Song
Information 2024, 15(12), 791; https://doi.org/10.3390/info15120791 - 10 Dec 2024
Viewed by 683
Abstract
Multi-class anomaly detection is more efficient and less resource-consuming in industrial anomaly detection scenes that involve multiple categories or exhibit large intra-class diversity. However, most industrial image anomaly detection methods are developed for one-class anomaly detection, which typically suffer significant performance drops in [...] Read more.
Multi-class anomaly detection is more efficient and less resource-consuming in industrial anomaly detection scenes that involve multiple categories or exhibit large intra-class diversity. However, most industrial image anomaly detection methods are developed for one-class anomaly detection, which typically suffer significant performance drops in multi-class scenarios. Research specifically targeting multi-class anomaly detection remains relatively limited. In this work, we propose a powerful unified normalizing flow for multi-class anomaly detection, which we call UniFlow. A multi-cognitive visual adapter (Mona) is employed in our method as the feature adaptation layer to adapt image features for both the multi-class anomaly detection task and the normalizing flow model, facilitating the learning of general knowledge of normal images across multiple categories. We adopt multi-cognitive convolutional networks with high capacity to construct the coupling layers within the normalizing flow model for more effective multi-class distribution modeling. In addition, we employ a multi-scale feature fusion module to aggregate features from various levels, thereby obtaining fused features with enhanced expressive capabilities. UniFlow achieves a class-average image-level AUROC of 99.1% and a class-average pixel-level AUROC of 98.0% on MVTec AD, outperforming the SOTA multi-class anomaly detection methods. Extensive experiments on three benchmark datasets, MVTec AD, VisA, and BTAD, demonstrate the efficacy and superiority of our unified normalizing flow in multi-class anomaly detection. Full article
Show Figures

Graphical abstract

Graphical abstract
Full article ">Figure 1
<p>One-class anomaly detection versus multi-class anomaly detection.</p>
Full article ">Figure 2
<p><b>Overview of UniFlow.</b> Mona FA: Mona feature adaptation. Ds: Downsample. MC Affine Layer: Multi-cognitive affine coupling layer. MC Additive Layer: Multi-cognitive additive coupling layer. Two Mona feature adaptation layers are employed to adapt the features from two stages. We downsample the feature maps extracted in the second stage to half of their original size via average pooling. Notably, Gaussian noise is added to the features only during the training phase.</p>
Full article ">Figure 3
<p>(<b>a</b>) The architecture of Mona. (<b>b</b>) The architecture of multi-cognitive convolutional module. (<b>c</b>) Mona-tuning in each SwinBlock.</p>
Full article ">Figure 4
<p>The architecture of the multi-cognitive additive coupling layer. Multi-cognitive Conv: Multi-cognitive convolutional module in <a href="#information-15-00791-f003" class="html-fig">Figure 3</a>b. <math display="inline"><semantics> <mrow> <mn>1</mn> <mo>×</mo> <mn>1</mn> </mrow> </semantics></math> Conv: A convolutional layer with a kernel size of <math display="inline"><semantics> <mrow> <mn>1</mn> <mo>×</mo> <mn>1</mn> </mrow> </semantics></math> and a stride of 1, where the output retains the same shape as the input. Global Affine: Further scaling and translation applied to the global output. Channel permute: Shuffling the order of the channel dimension.</p>
Full article ">Figure 5
<p>The architecture of the multi-cognitive affine coupling layer. Multi-cognitive Conv: Multi-cognitive convolutional module in <a href="#information-15-00791-f003" class="html-fig">Figure 3</a>b. <math display="inline"><semantics> <mrow> <mn>3</mn> <mo>×</mo> <mn>3</mn> </mrow> </semantics></math> Conv: A convolutional layer with a kernel size of <math display="inline"><semantics> <mrow> <mn>3</mn> <mo>×</mo> <mn>3</mn> </mrow> </semantics></math> and a stride of 1, where the output has the same spatial dimensions as the input but with twice the number of channels. Global Affine: Further scaling and translation applied to the global output. Channel permute: Shuffling the order of the channel dimension.</p>
Full article ">Figure 6
<p>Visualization results of anomaly localization for examples from MVTec AD [<a href="#B18-information-15-00791" class="html-bibr">18</a>] and BTAD [<a href="#B20-information-15-00791" class="html-bibr">20</a>].</p>
Full article ">Figure 7
<p>Visualization results of anomaly localization for examples from VisA [<a href="#B19-information-15-00791" class="html-bibr">19</a>].</p>
Full article ">Figure 8
<p>The impact of varying feature jittering probability on anomaly detection and localization performance on MVTec AD.</p>
Full article ">Figure 9
<p>Framework of the convolutional feature adaptation layer.</p>
Full article ">
24 pages, 9923 KiB  
Article
Application of Online Anomaly Detection Using One-Class Classification to the Z24 Bridge
by Amro Abdrabo
Sensors 2024, 24(23), 7866; https://doi.org/10.3390/s24237866 - 9 Dec 2024
Viewed by 611
Abstract
The usage of anomaly detection is of critical importance to numerous domains, including structural health monitoring (SHM). In this study, we examine an online setting for damage detection in the Z24 bridge. We evaluate and compare the performance of the elliptic envelope, incremental [...] Read more.
The usage of anomaly detection is of critical importance to numerous domains, including structural health monitoring (SHM). In this study, we examine an online setting for damage detection in the Z24 bridge. We evaluate and compare the performance of the elliptic envelope, incremental one-class support vector classification, local outlier factor, half-space trees, and entropy-guided envelopes. Our findings demonstrate that XGBoost exhibits enhanced performance in identifying a limited set of significant features. Additionally, we present a novel approach to manage drift through the application of entropy measures to structural state instances. The study is the first to assess the applicability of one-class classification for anomaly detection on the short-term structural health data of the Z24 bridge. Full article
(This article belongs to the Section Physical Sensors)
Show Figures

Figure 1

Figure 1
<p>Reduction of a feature proportional to a stiffness that undergoes natural degradation. As stiffness is affected by damage, this is also a damage-sensitive feature (DSF).</p>
Full article ">Figure 2
<p>Wireframe of the Z24 bridge on ANSYS 2023 R1. The Z-axis defines the longitudinal direction. The Y-axis defines vertical, and the X-axis is transversal. Sensor R1 is at location B. Sensor R2 is at A, and sensor R3 isat C. Only these sensors are used, as they were the only sensors which were not moved during the short-term monitoring campaign.</p>
Full article ">Figure 3
<p>Reduction in natural frequency of second mode (near 5 Hz), from the power spectral density along R2T, visible due to damage progression.</p>
Full article ">Figure 4
<p>A t-SNE visualization on the training set. Labels correspond to the state from <a href="#sensors-24-07866-t001" class="html-table">Table 1</a>.</p>
Full article ">Figure 5
<p>PCA applied over different feature subsets. (<b>a</b>) Features from <a href="#sensors-24-07866-t002" class="html-table">Table 2</a> applied to all channels (R1V, R2L, R2V, R2T, R3V). (<b>b</b>) Features from the PSD of the channels. (<b>c</b>) Features derived from the transmittance with respect to R1V of the remaining channels.</p>
Full article ">Figure 6
<p>Supervised feature selection framework. Models <span class="html-italic">M</span> are used for feature selection, and <math display="inline"><semantics> <msup> <mi>M</mi> <mo>′</mo> </msup> </semantics></math> are used for assessing the quality of found features. Here, <span class="html-italic">M</span> can be KNN, SVC, or XGBoost, where <math display="inline"><semantics> <msup> <mi>M</mi> <mo>′</mo> </msup> </semantics></math> is correspondingly relief feature selection, logistic regression, and XGBoost, respectively.</p>
Full article ">Figure 7
<p>Four most important features extracted by XGBoost in third step of <a href="#sensors-24-07866-f006" class="html-fig">Figure 6</a>.</p>
Full article ">Figure 8
<p>Four most important features extracted by logistic regression in third step of <a href="#sensors-24-07866-f006" class="html-fig">Figure 6</a>.</p>
Full article ">Figure 9
<p>Four most important features extracted by Relief in third step of <a href="#sensors-24-07866-f006" class="html-fig">Figure 6</a>.</p>
Full article ">Figure 10
<p>PCA of top four features of ReliefF from <a href="#sensors-24-07866-f009" class="html-fig">Figure 9</a>, XGBoost from <a href="#sensors-24-07866-f007" class="html-fig">Figure 7</a>, and logistic regression from <a href="#sensors-24-07866-f008" class="html-fig">Figure 8</a>. Damaged state points are in red, while healthy state points are in blue.</p>
Full article ">Figure 11
<p>CV-scores obtained from step 3 of <a href="#sensors-24-07866-f006" class="html-fig">Figure 6</a> for supervised classification methods. XGBoost achieves the highest CV-score for 4 features at 0.95.</p>
Full article ">Figure 12
<p>Standardization and PCA on data from first state.</p>
Full article ">Figure 13
<p>After PCA and standardization have been initialized in <a href="#sensors-24-07866-f012" class="html-fig">Figure 12</a>, the OCC model is trained using the procedure shown here. For cross-validation, the performance on the omitted healthy scenario is evaluated instead of the unseen damage set.</p>
Full article ">Figure 14
<p>Example of a half-space tree for two-dimensional data. Red and blue regions correspond to anomalous and non-anomalous regions, respectively.</p>
Full article ">Figure 15
<p>XGBoost demonstrated superior disentanglement with regard to PCA components. All PCA components approximately have a one-to-one mapping to original features. (<b>a</b>) Weight magnitudes of PCA eigenvectors, used in <math display="inline"><semantics> <mi>ψ</mi> </semantics></math>, on the original features, denoted <math display="inline"><semantics> <mi>ϕ</mi> </semantics></math>. (<b>b</b>) Correlation matrix of the original features.</p>
Full article ">Figure 16
<p>Projection of the decision surface for (<b>a</b>) elliptic envelope, (<b>b</b>) one-class SVC, (<b>c</b>) half-space trees, (<b>d</b>) local outlier factor, and (<b>e</b>) the last elliptic envelope formed in entropy-guided envelopes, respectively, onto two-dimensional PCA space. Negative regions are regions where the model outputs no anomaly, while positive regions are anomalous.</p>
Full article ">Figure 17
<p>Accuracy categorized by damage type. (<b>a</b>) Accuracy for concrete damage detection. (<b>b</b>) Accuracy for tendons and anchor damage detection. (<b>c</b>) Accuracy for landslide damage detection. (<b>d</b>) Accuracy for healthy domain (1-<math display="inline"><semantics> <mrow> <mi>F</mi> <mi>P</mi> <mi>R</mi> </mrow> </semantics></math>).</p>
Full article ">Figure 18
<p>Anomaly scores for the different OCC techniques. Color intensity correlates with anomaly score. False predictions are shown with a cross. The blue region contains the points seen during training that are predicted as healthy/non-anomalous, and red corresponds to the damage evaluation set. For states, refer to <a href="#sensors-24-07866-t001" class="html-table">Table 1</a>.</p>
Full article ">Figure 19
<p>Z24 model in ANSYS 2023 R1. The outermost piers are embedded within the ground elevations and hence are not visible. Two of the four features used, listed in <a href="#sensors-24-07866-f007" class="html-fig">Figure 7</a>, are shown in the bottom right and bottom left figures. (<b>a</b>) Second mode shape of the Z24 bridge. (<b>b</b>) Fourth mode shape of the Z24 bridge. (<b>c</b>) Transmittance of R2L relative to R1V. In green are the windows from which we extract the relative position of the peaks. These constitute the first and last features of <a href="#sensors-24-07866-f007" class="html-fig">Figure 7</a>.</p>
Full article ">
14 pages, 2267 KiB  
Article
Aero-Engine Fault Detection with an LSTM Auto-Encoder Combined with a Self-Attention Mechanism
by Wenyou Du, Jingyi Zhang, Guanglei Meng and Haoran Zhang
Machines 2024, 12(12), 879; https://doi.org/10.3390/machines12120879 - 4 Dec 2024
Viewed by 574
Abstract
The safe operation of aero-engines is crucial for ensuring flight safety, and effective fault detection methods are fundamental to achieving this objective. In this paper, we propose a novel approach that integrates an auto-encoder with long short-term memory (LSTM) networks and a self-attention [...] Read more.
The safe operation of aero-engines is crucial for ensuring flight safety, and effective fault detection methods are fundamental to achieving this objective. In this paper, we propose a novel approach that integrates an auto-encoder with long short-term memory (LSTM) networks and a self-attention mechanism for the anomaly detection of aero-engine time-series data. The dataset utilized in this study was simulated from real data and injected with fault information. A fault detection model is developed utilizing normal data samples for training and faulty data samples for testing. The LSTM auto-encoder processes the time-series data through an encoder–decoder architecture, extracting latent representations and reconstructing the original inputs. Furthermore, the self-attention mechanism captures long-range dependencies and significant features within the sequences, thereby enhancing the detection accuracy of the model. Comparative analyses with the traditional LSTM auto-encoder, as well as one-class support vector machines (OC-SVM) and isolation forests (IF), reveal that the experimental results substantiate the feasibility and effectiveness of the proposed method, highlighting its potential value in engineering applications. Full article
Show Figures

Figure 1

Figure 1
<p>Auto-encoder.</p>
Full article ">Figure 2
<p>LSTM structure.</p>
Full article ">Figure 3
<p>Self-attention structure.</p>
Full article ">Figure 4
<p>SLAE fault detection process.</p>
Full article ">Figure 5
<p>Fault detection flow chart.</p>
Full article ">Figure 6
<p>Fault 1 raw data.</p>
Full article ">Figure 7
<p>Fault 2 raw data.</p>
Full article ">Figure 8
<p>Fault 3 raw data.</p>
Full article ">Figure 9
<p>Fault 1 dection. (<b>a</b>) SLAE fault dection; (<b>b</b>) LSTM fault dection; (<b>c</b>) OC-SVM fault dection; (<b>d</b>) IF fault dection.</p>
Full article ">Figure 10
<p>Fault 2 dection. (<b>a</b>) SLAE fault dection; (<b>b</b>) LSTM fault dection; (<b>c</b>) OC-SVM fault dection; (<b>d</b>) IF fault dection.</p>
Full article ">Figure 11
<p>Fault 3 dection. (<b>a</b>) SLAE fault dection; (<b>b</b>) LSTM fault dection; (<b>c</b>) OC-SVM fault dection; (<b>d</b>) IF fault dection.</p>
Full article ">
25 pages, 4000 KiB  
Article
CASSAD: Chroma-Augmented Semi-Supervised Anomaly Detection for Conveyor Belt Idlers
by Fahad Alharbi, Suhuai Luo, Abdullah Alsaedi, Sipei Zhao and Guang Yang
Sensors 2024, 24(23), 7569; https://doi.org/10.3390/s24237569 - 27 Nov 2024
Viewed by 561
Abstract
Idlers are essential to conveyor systems, as well as supporting and guiding belts to ensure production efficiency. Proper idler maintenance prevents failures, reduces downtime, cuts costs, and improves reliability. Most studies on idler fault detection rely on supervised methods, which depend on large [...] Read more.
Idlers are essential to conveyor systems, as well as supporting and guiding belts to ensure production efficiency. Proper idler maintenance prevents failures, reduces downtime, cuts costs, and improves reliability. Most studies on idler fault detection rely on supervised methods, which depend on large labelled datasets for training. However, acquiring such labelled data is often challenging in industrial environments due to the rarity of faults and the labour-intensive nature of the labelling process. To address this, we propose the chroma-augmented semi-supervised anomaly detection (CASSAD) method, designed to perform effectively with limited labelled data. At the core of CASSAD is the one-class SVM (OC-SVM), a model specifically developed for anomaly detection in cases where labelled anomalies are scarce. We also compare CASSAD’s performance with other common models like the local outlier factor (LOF) and isolation forest (iForest), evaluating each with the area under the curve (AUC) to assess their ability to distinguish between normal and anomalous data. CASSAD introduces chroma features, such as chroma energy normalised statistics (CENS), the constant-Q transform (CQT), and the chroma short-time Fourier transform (STFT), enhanced through filtering to capture rich harmonic information from idler sounds. To reduce feature complexity, we utilize the mean and standard deviation (std) across chroma features. The dataset is further augmented using additive white Gaussian noise (AWGN). Testing on an industrial dataset of idler sounds, CASSAD achieved an AUC of 96% and an accuracy of 91%, surpassing a baseline autoencoder and other traditional models. These results demonstrate the model’s robustness in detecting anomalies with minimal dependence on labelled data, offering a practical solution for industries with limited labelled datasets. Full article
Show Figures

Figure 1

Figure 1
<p>Autoencoder structure.</p>
Full article ">Figure 2
<p>Overview of the proposed chroma-augmented semi-supervised anomaly detection (CASSAD) model.</p>
Full article ">Figure 3
<p>Analysis of different chroma features across four stages: original, harmonic, filtered, and smoothed, showing (<b>a</b>) STFT, (<b>b</b>) CQT, and (<b>c</b>) CENS.</p>
Full article ">Figure 4
<p>Distribution of mean and standard deviation (STD) of chroma features (CQT, CENS, and STFT) for normal and abnormal signals.</p>
Full article ">Figure 5
<p>ROC curves with the best AUC for LOF, isolation forest, and one-class SVM models.</p>
Full article ">Figure 6
<p>Analysis results from training the autoencoder on chroma features. (<b>Top Row</b>): Original and reconstructed spectrograms (<b>Left</b>) and the loss distribution for chroma_cens_features using the mean absolute error (MAE) function (<b>Right</b>). (<b>Bottom row</b>): Training and validation loss curves over 30 epochs showing minimal gap and no significant overfitting or underfitting.</p>
Full article ">Figure 7
<p>Isolation forest detection results for idler components (mean vs. standard deviation with/without PCA). This figure highlights the model’s performance across different chroma features and aggregation methods.</p>
Full article ">Figure 8
<p>LOF detection results for idler components (mean vs. standard deviation with/without PCA). This figure highlights the model’s performance across different chroma features and aggregation methods.</p>
Full article ">Figure 9
<p>Visualisation of the best results for LOF + PCA, isolation forest, and the proposed CASSAD model. (<b>Top row</b>): Confusion matrices showing classification results. (<b>Second row</b>): ROC curves, with the proposed CASSAD model reaching the highest AUC of 0.96. (<b>Third row</b>): Precision–recall curves, where the proposed CASSAD model without PCA shows the best balance. (<b>Bottom row</b>): t-SNE plots illustrating data separability, with the proposed CASSAD model achieving the clearest distinction.</p>
Full article ">Figure 10
<p>Comparison of AUC and accuracy metrics for the proposed model using all chroma features, both with and without noise filtering. The figure highlights the performance improvements achieved by the proposed model (CASSAD).</p>
Full article ">Figure 11
<p>Consumption time comparison across models and features.</p>
Full article ">
17 pages, 6729 KiB  
Article
Anomaly Detection Method for Harmonic Reducers with Only Healthy Data
by Yuqing Li, Linghui Zhu, Minqiang Xu and Yunzhao Jia
Sensors 2024, 24(23), 7435; https://doi.org/10.3390/s24237435 - 21 Nov 2024
Viewed by 501
Abstract
A harmonic reducer is an important component of industrial robots. In practical applications, it is difficult to obtain enough anomaly data from error cases for the supervised training of models. Whether the information contained in regular features is sensitive to anomaly detection is [...] Read more.
A harmonic reducer is an important component of industrial robots. In practical applications, it is difficult to obtain enough anomaly data from error cases for the supervised training of models. Whether the information contained in regular features is sensitive to anomaly detection is unknown. In this paper, we propose an anomaly detection frame for a harmonic reducer with only healthy data. We considered an auto-encoder trained using only healthy features, such as feature mapping, in which the difference between the output and the input constitutes a new high-dimensional feature space that retained information relevant only to anomalies. Compared to the original feature space, this space was more sensitive to abnormal data. The mapped features were then fed into the OCSVM to preserve the feature details of the abnormal information. The effectiveness of this method was validated by multiple sets of data collecting from harmonic reducers. Three different residual calculations and four different AE models were used, showing that the method outperforms an AE or an OCSVM alone. It is also verified that the method outperforms other typical anomaly detection methods. Full article
(This article belongs to the Section Fault Diagnosis & Sensors)
Show Figures

Figure 1

Figure 1
<p>Unit of LSTM.</p>
Full article ">Figure 2
<p>Structure of AE-OCSVM model.</p>
Full article ">Figure 3
<p>Harmonic reducer test bench.</p>
Full article ">Figure 4
<p>Ref. [<a href="#B7-sensors-24-07435" class="html-bibr">7</a>] faulty part: (<b>a</b>) circular spline gear crack fault; (<b>b</b>) flex–spline gear wear fault; (<b>c</b>) flex–spline gear crack fault; (<b>d</b>) flexible bearing outer race semi-crack fault; (<b>e</b>) flexible bearing inner race wear; (<b>f</b>) cross roller bearing outer race fault; (<b>g</b>) cross roller bearing inner race fault.</p>
Full article ">Figure 5
<p>Vibration signal after standardization: (<b>a</b>) normal; (<b>b</b>–<b>h</b>): F1–F7.</p>
Full article ">Figure 6
<p>Structure of AE.</p>
Full article ">Figure 7
<p>Results with different res.</p>
Full article ">Figure 8
<p>Part of the MSE of different models. (<b>a</b>–<b>f</b>) represents six different working conditions. The first row of images shows the MSE of each test data. The second row of images shows the probability distribution of the MSE. Blue represents the probability distribution of the training data, red represents the normal data in the test data, and yellow represents the abnormal data in the test data. The purple vertical line represents the threshold of the model.</p>
Full article ">Figure 9
<p>Features of different models after PCA. (<b>a</b>) original features; (<b>b</b>) SpAE-OCSVM; (<b>c</b>) LSTMAE-OCSVM; (<b>d</b>) StAE-OCSVM; (<b>e</b>) VAE-OCSVM.</p>
Full article ">Figure 10
<p>Results of AE-OCSVM, AE, and OCSVM.</p>
Full article ">Figure 11
<p>AUC of different methods.</p>
Full article ">
14 pages, 3986 KiB  
Article
Anomaly Detection Utilizing One-Class Classification—A Machine Learning Approach for the Analysis of Plant Fast Fluorescence Kinetics
by Nam Trung Tran
Stresses 2024, 4(4), 773-786; https://doi.org/10.3390/stresses4040051 - 18 Nov 2024
Viewed by 643
Abstract
The analysis of fast fluorescence kinetics, specifically through the JIP test, is a valuable tool for identifying and characterizing plant stress. However, interpreting OJIP data requires a comprehensive understanding of their underlying theory. This study proposes a Machine Learning-based approach using a One-Class [...] Read more.
The analysis of fast fluorescence kinetics, specifically through the JIP test, is a valuable tool for identifying and characterizing plant stress. However, interpreting OJIP data requires a comprehensive understanding of their underlying theory. This study proposes a Machine Learning-based approach using a One-Class Support Vector Machine anomaly detection model to effectively categorize OJIP measurements into “normal”, representing healthy plants, and “anomalies”. This approach was validated using a previously published dataset. A subgroup of the identified “anomalies” was clearly linked to stress-induced reductions in photosynthesis. Furthermore, the percentage of these “anomalies” showed a meaningful correlation with both the progression and severity of stress. The results highlight the still largely unexploited potential of Machine Learning in OJIP analysis. Full article
(This article belongs to the Collection Feature Papers in Plant and Photoautotrophic Stresses)
Show Figures

Figure 1

Figure 1
<p>Identification of 24 anomaly measurements (red) as those with significantly lower F<sub>V</sub>/F<sub>M</sub> values compared to the initial measurements. Outlier threshold was set at the latter’s mean minus three times the standard deviation.</p>
Full article ">Figure 2
<p>Fine-tuning of the classification model. Models were trained on the training dataset, with <span class="html-italic">nu</span> values ranging from 0.01 to 0.5, and tested on the fine-tuning dataset. Green: Type I error (“normal” misidentified as “anomalies”). Red: Type II error (“anomalies” misidentified as “normal”). The fractions show the number of misidentifications over the total number of predictions. The nu value of 0.05 (filled triangles) was used for all subsequent predictions.</p>
Full article ">Figure 3
<p>Comparison of the photosynthetic performance between the “anomaly” (red) and “normal” (blue) groups using standard JIP test. For a more detailed understanding of the presented OJIP parameters, please see Stirbet and Govindjee, 2011 [<a href="#B11-stresses-04-00051" class="html-bibr">11</a>]. The values are normalized, with values from “normal” samples set to 1. The small concentric circles on the right (with the red numbers) show the scales.</p>
Full article ">Figure 4
<p>Comparison between “anomalies” (A) and “normal” (N) measurements in three field experiments and in the greenhouse experiment. Two commonly used OJIP metrics are used for comparison: F<sub>V</sub>/F<sub>M</sub>—the maximum photochemical quantum yield of PS II; PI<sub>ABS</sub>—the performance index on energy absorption basis. Stars (*) denote statistically significant differences (<span class="html-italic">p</span> &lt; 0.05) between “anomalies” and “normal” measurements in each experiment.</p>
Full article ">Figure 5
<p>UMAP visualization of all measurements.</p>
Full article ">Figure 6
<p>The photosynthetic performance of the “normal” (blue), “anomaly” type 1 (red), and type 2 (green) groups was compared using the standard JIP test. For a more detailed understanding of the presented OJIP parameters, please see Stirbet and Govindjee, 2011 [<a href="#B11-stresses-04-00051" class="html-bibr">11</a>]. Box plots were used to display the comparison of two commonly used OJIP metrics: F<sub>V</sub>/F<sub>M</sub>, which represents the maximum photochemical quantum yield of PS II; and PI<sub>ABS</sub>, which represents the performance index on an energy absorption basis. The values are normalized, with values from “normal” samples set to 1. The letters (a–c) indicate groups that are statistically significantly different from each other. Samples with the same letters are not statistically different.</p>
Full article ">Figure 7
<p>The percentage of detected “anomaly” types 1 and 2 on each measurement day across the four experiments.</p>
Full article ">Figure 8
<p>Nine features extracted from OJIP curve for classification: baseline fluorescence intensity (F<sub>O</sub>); peak fluorescence intensity (F<sub>M</sub>); fluorescence intensities at five specific time marks (50 µs, 100 µs, 300 µs, 2 ms, and 30 ms—F1, F2, F3, F4, and F5); the time at which the maximum fluorescence value, F<sub>M</sub>, was reached (Tfm); and the area above the fluorescence curve between F<sub>O</sub> and F<sub>M</sub> (Area).</p>
Full article ">
26 pages, 1244 KiB  
Article
Privacy-Preserving Detection of Tampered Radio-Frequency Transmissions Utilizing Federated Learning in LoRa Networks
by Nurettin Selcuk Senol, Mohamed Baza, Amar Rasheed and Maazen Alsabaan
Sensors 2024, 24(22), 7336; https://doi.org/10.3390/s24227336 - 17 Nov 2024
Viewed by 752
Abstract
LoRa networks, widely adopted for low-power, long-range communication in IoT applications, face critical security concerns as radio-frequency transmissions are increasingly vulnerable to tampering. This paper addresses the dual challenges of privacy-preserving detection of tampered transmissions and the identification of unknown attacks in LoRa-based [...] Read more.
LoRa networks, widely adopted for low-power, long-range communication in IoT applications, face critical security concerns as radio-frequency transmissions are increasingly vulnerable to tampering. This paper addresses the dual challenges of privacy-preserving detection of tampered transmissions and the identification of unknown attacks in LoRa-based IoT networks. Leveraging Federated Learning (FL), our approach enables the detection of tampered RF transmissions while safeguarding sensitive IoT data, as FL allows model training on distributed devices without sharing raw data. We evaluated the performance of multiple FL-enabled anomaly-detection algorithms, including Convolutional Autoencoder Federated Learning (CAE-FL), Isolation Forest Federated Learning (IF-FL), One-Class Support Vector Machine Federated Learning (OCSVM-FL), Local Outlier Factor Federated Learning (LOF-FL), and K-Means Federated Learning (K-Means-FL). Using metrics such as accuracy, precision, recall, and F1-score, CAE-FL emerged as the top performer, achieving 97.27% accuracy and a balanced precision, recall, and F1-score of 0.97, with IF-FL close behind at 96.84% accuracy. Competitive performance from OCSVM-FL and LOF-FL, along with the comparable results of K-Means-FL, highlighted the robustness of clustering-based detection methods in this context. Visual analyses using confusion matrices and ROC curves provided further insights into each model’s effectiveness in detecting tampered signals. This research underscores the capability of federated learning to enhance privacy and security in anomaly detection for LoRa networks, even against unknown attacks, marking a significant advancement in securing IoT communications in sensitive applications. Full article
Show Figures

Figure 1

Figure 1
<p>Federated learning architecture.</p>
Full article ">Figure 2
<p>One class SVM algorithm with FL client one results.</p>
Full article ">Figure 3
<p>One class SVM algorithm with FL client two results.</p>
Full article ">Figure 4
<p>One class SVM algorithm with FL client three results.</p>
Full article ">Figure 5
<p>One class SVM algorithm with FL client four results.</p>
Full article ">Figure 6
<p>One class SVM algorithm with FL client five results.</p>
Full article ">Figure 7
<p>One class SVM algorithm with FL client five results.</p>
Full article ">Figure 8
<p>One class SVM algorithm with FL client five results.</p>
Full article ">Figure 9
<p>Isolation Forest with FL data across clients.</p>
Full article ">Figure 10
<p>LOF client data distribution graph.</p>
Full article ">Figure 11
<p>K-Means algorithm with FL client results.</p>
Full article ">Figure 12
<p>Average time calculation between algorithms.</p>
Full article ">Figure 13
<p>AUC curves for the algorithms (<b>a</b>) OCSVM-FL, (<b>b</b>) CAE-FL, (<b>c</b>) IF-FL, (<b>d</b>) LOF-FL, and (<b>e</b>) K-Means-FL.</p>
Full article ">Figure 14
<p>The performance of five anomaly-detection methods in classifying normal and anomalous data is displayed in confusion matrices. These algorithms are (<b>a</b>) OCSVM-FL, (<b>b</b>) CAE-FL, (<b>c</b>) IF-FL, (<b>d</b>) LOF-FL, and (<b>e</b>) K-Means-FL. Lighter hues indicate misclassifications (false positives and false negatives), while darker hues indicate accurate classifications (true positives and true negatives).</p>
Full article ">
20 pages, 3150 KiB  
Article
Early Fault Detection and Operator-Based MIMO Fault-Tolerant Temperature Control of Microreactor
by Yuma Morita and Mingcong Deng
Appl. Sci. 2024, 14(21), 9907; https://doi.org/10.3390/app14219907 - 29 Oct 2024
Viewed by 708
Abstract
A microreactor is a chemical reaction device that mixes liquids in a very narrow channel and continuously generates reactions. They are attracting attention as next-generation chemical reaction devices because of their ability to achieve small-scale and highly efficient reactions compared to the conventional [...] Read more.
A microreactor is a chemical reaction device that mixes liquids in a very narrow channel and continuously generates reactions. They are attracting attention as next-generation chemical reaction devices because of their ability to achieve small-scale and highly efficient reactions compared to the conventional badge method. However, the challenge is to design a control system that is tolerant of faults in some of the enormous number of sensors in order to achieve parallel production by numbering up. In a previous study, a simultaneous control system for two different temperatures was proposed in an experimental system that imitated the microreactor cooled by Peltier devices. In addition, a fault-tolerant control system for one area has also been proposed. However, the fault-tolerant control system could not be applied to the control system of two temperatures in the previous study. In this paper, we extend it to a two-input, two-output fault-tolerant control system. We also use a fault detection system that combines ChangeFinder, a time-series data analysis method, and One-Class SVM, an unsupervised learning method. Finally, the effectiveness of the proposed method is confirmed by experiments. Full article
(This article belongs to the Special Issue Artificial Intelligence in Fault Diagnosis and Signal Processing)
Show Figures

Figure 1

Figure 1
<p>Experimental system: (<b>a</b>) external view. (<b>b</b>) Schematic diagram of the microreactor system.</p>
Full article ">Figure 2
<p>Assumed temperature sensor fault conditions: (<b>a</b>) deviation of normal and faulty sensor readings from the correct values. (<b>b</b>) Temperature sensor fault condition.</p>
Full article ">Figure 3
<p>Area (Part) definition of microreactor system.</p>
Full article ">Figure 4
<p>Control system based on operator theory.</p>
Full article ">Figure 5
<p>Fault-tolerant control system.</p>
Full article ">Figure 6
<p>Control system to remove coupling effects.</p>
Full article ">Figure 7
<p>Overall control system.</p>
Full article ">Figure 8
<p>Representation of temperature sensor faults in control systems.</p>
Full article ">Figure 9
<p>Training data learning.</p>
Full article ">Figure 10
<p>Hyperparameter optimization process using the real-coded genetic algorithm.</p>
Full article ">Figure 11
<p>Simulation results under condition where Part <math display="inline"><semantics> <msub> <mi mathvariant="normal">W</mi> <mn>1</mn> </msub> </semantics></math> increased <math display="inline"><semantics> <mrow> <mn>3.2</mn> <mspace width="3.33333pt"/> <mmultiscripts> <mi mathvariant="normal">C</mi> <none/> <none/> <mprescripts/> <none/> <mo>°</mo> </mmultiscripts> </mrow> </semantics></math> from <math display="inline"><semantics> <mrow> <mn>500</mn> <mspace width="3.33333pt"/> <mi mathvariant="normal">s</mi> </mrow> </semantics></math> and Part <math display="inline"><semantics> <msub> <mi mathvariant="normal">W</mi> <mn>3</mn> </msub> </semantics></math> increased <math display="inline"><semantics> <mrow> <mn>3.6</mn> <mspace width="3.33333pt"/> <mmultiscripts> <mi mathvariant="normal">C</mi> <none/> <none/> <mprescripts/> <none/> <mo>°</mo> </mmultiscripts> </mrow> </semantics></math> from <math display="inline"><semantics> <mrow> <mn>500</mn> <mspace width="3.33333pt"/> <mi mathvariant="normal">s</mi> </mrow> </semantics></math>: (<b>a</b>) temperature. (<b>b</b>) Measurements of faulty sensors. (<b>c</b>) Control input.</p>
Full article ">Figure 12
<p>Simulation results under condition where Part <math display="inline"><semantics> <msub> <mi mathvariant="normal">W</mi> <mn>1</mn> </msub> </semantics></math> increased <math display="inline"><semantics> <mrow> <mn>3.2</mn> <mspace width="3.33333pt"/> <mmultiscripts> <mi mathvariant="normal">C</mi> <none/> <none/> <mprescripts/> <none/> <mo>°</mo> </mmultiscripts> </mrow> </semantics></math> from <math display="inline"><semantics> <mrow> <mn>500</mn> <mspace width="3.33333pt"/> <mi mathvariant="normal">s</mi> </mrow> </semantics></math> and Part <math display="inline"><semantics> <msub> <mi mathvariant="normal">W</mi> <mn>3</mn> </msub> </semantics></math> increased <math display="inline"><semantics> <mrow> <mn>3.6</mn> <mspace width="3.33333pt"/> <mmultiscripts> <mi mathvariant="normal">C</mi> <none/> <none/> <mprescripts/> <none/> <mo>°</mo> </mmultiscripts> </mrow> </semantics></math> from <math display="inline"><semantics> <mrow> <mn>500</mn> <mspace width="3.33333pt"/> <mi mathvariant="normal">s</mi> </mrow> </semantics></math>: (<b>a</b>) change-point scores. (<b>b</b>) Classification of fault detection.</p>
Full article ">Figure 13
<p>Simulation results under condition where Part <math display="inline"><semantics> <msub> <mi mathvariant="normal">W</mi> <mn>3</mn> </msub> </semantics></math> increased <math display="inline"><semantics> <mrow> <mn>0.2</mn> <mspace width="3.33333pt"/> <mmultiscripts> <mi mathvariant="normal">C</mi> <none/> <none/> <mprescripts/> <none/> <mo>°</mo> </mmultiscripts> </mrow> </semantics></math> from <math display="inline"><semantics> <mrow> <mn>500</mn> <mspace width="3.33333pt"/> <mi mathvariant="normal">s</mi> </mrow> </semantics></math>: (<b>a</b>) temperature. (<b>b</b>) Measurements of faulty sensors. (<b>c</b>) Control input.</p>
Full article ">Figure 14
<p>Simulation results under condition where Part <math display="inline"><semantics> <msub> <mi mathvariant="normal">W</mi> <mn>3</mn> </msub> </semantics></math> increased <math display="inline"><semantics> <mrow> <mn>0.2</mn> <mspace width="3.33333pt"/> <mmultiscripts> <mi mathvariant="normal">C</mi> <none/> <none/> <mprescripts/> <none/> <mo>°</mo> </mmultiscripts> </mrow> </semantics></math> from <math display="inline"><semantics> <mrow> <mn>500</mn> <mspace width="3.33333pt"/> <mi mathvariant="normal">s</mi> </mrow> </semantics></math>: (<b>a</b>) change-point scores. (<b>b</b>) Classification of fault detection.</p>
Full article ">Figure 15
<p>Experimental results under condition where Part <math display="inline"><semantics> <msub> <mi mathvariant="normal">W</mi> <mn>1</mn> </msub> </semantics></math> increased <math display="inline"><semantics> <mrow> <mn>3.0</mn> <mspace width="3.33333pt"/> <mmultiscripts> <mi mathvariant="normal">C</mi> <none/> <none/> <mprescripts/> <none/> <mo>°</mo> </mmultiscripts> </mrow> </semantics></math> from <math display="inline"><semantics> <mrow> <mn>601</mn> <mspace width="3.33333pt"/> <mi mathvariant="normal">s</mi> </mrow> </semantics></math> and Part <math display="inline"><semantics> <msub> <mi mathvariant="normal">W</mi> <mn>3</mn> </msub> </semantics></math> increased <math display="inline"><semantics> <mrow> <mn>3.5</mn> <mspace width="3.33333pt"/> <mmultiscripts> <mi mathvariant="normal">C</mi> <none/> <none/> <mprescripts/> <none/> <mo>°</mo> </mmultiscripts> </mrow> </semantics></math> from <math display="inline"><semantics> <mrow> <mn>601</mn> <mspace width="3.33333pt"/> <mi mathvariant="normal">s</mi> </mrow> </semantics></math>: (<b>a</b>) temperature. (<b>b</b>) Measurements of faulty sensors. (<b>c</b>) Control input.</p>
Full article ">Figure 16
<p>Experimental results under condition where Part <math display="inline"><semantics> <msub> <mi mathvariant="normal">W</mi> <mn>1</mn> </msub> </semantics></math> increased <math display="inline"><semantics> <mrow> <mn>3.0</mn> <mspace width="3.33333pt"/> <mmultiscripts> <mi mathvariant="normal">C</mi> <none/> <none/> <mprescripts/> <none/> <mo>°</mo> </mmultiscripts> </mrow> </semantics></math> from <math display="inline"><semantics> <mrow> <mn>601</mn> <mspace width="3.33333pt"/> <mi mathvariant="normal">s</mi> </mrow> </semantics></math> and Part <math display="inline"><semantics> <msub> <mi mathvariant="normal">W</mi> <mn>3</mn> </msub> </semantics></math> increased <math display="inline"><semantics> <mrow> <mn>3.5</mn> <mspace width="3.33333pt"/> <mmultiscripts> <mi mathvariant="normal">C</mi> <none/> <none/> <mprescripts/> <none/> <mo>°</mo> </mmultiscripts> </mrow> </semantics></math> from <math display="inline"><semantics> <mrow> <mn>601</mn> <mspace width="3.33333pt"/> <mi mathvariant="normal">s</mi> </mrow> </semantics></math>: (<b>a</b>) change-point scores. (<b>b</b>) Classification of fault detection.</p>
Full article ">
18 pages, 1300 KiB  
Article
XAI-Based Accurate Anomaly Detector That Is Robust Against Black-Box Evasion Attacks for the Smart Grid
by Islam Elgarhy, Mahmoud M. Badr, Mohamed Mahmoud, Maazen Alsabaan, Tariq Alshawi and Muteb Alsaqhan
Appl. Sci. 2024, 14(21), 9897; https://doi.org/10.3390/app14219897 - 29 Oct 2024
Viewed by 1183
Abstract
In the realm of smart grids, machine learning (ML) detectors—both binary (or supervised) and anomaly (or unsupervised)—have proven effective in detecting electricity theft (ET). However, binary detectors are designed for specific attacks, making their performance unpredictable against new attacks. Anomaly detectors, conversely, are [...] Read more.
In the realm of smart grids, machine learning (ML) detectors—both binary (or supervised) and anomaly (or unsupervised)—have proven effective in detecting electricity theft (ET). However, binary detectors are designed for specific attacks, making their performance unpredictable against new attacks. Anomaly detectors, conversely, are trained on benign data and identify deviations from benign patterns as anomalies, but their performance is highly sensitive to the selected threshold values. Additionally, ML detectors are vulnerable to evasion attacks, where attackers make minimal changes to malicious samples to evade detection. To address these limitations, we introduce a hybrid anomaly detector that combines a Deep Auto-Encoder (DAE) with a One-Class Support Vector Machine (OCSVM). This detector not only enhances classification performance but also mitigates the threshold sensitivity of the DAE. Furthermore, we evaluate the vulnerability of this detector to benchmark evasion attacks. Lastly, we propose an accurate and robust cluster-based DAE+OCSVM ET anomaly detector, trained using Explainable Artificial Intelligence (XAI) explanations generated by the Shapley Additive Explanations (SHAP) method on consumption readings. Our experimental results demonstrate that the proposed XAI-based detector achieves superior classification performance and exhibits enhanced robustness against various evasion attacks, including gradient-based and optimization-based methods, under a black-box threat model. Full article
(This article belongs to the Special Issue IoT in Smart Cities and Homes, 2nd Edition)
Show Figures

Figure 1

Figure 1
<p>Black-box threat model.</p>
Full article ">Figure 2
<p>The proposed XAI-based robust and accurate ET anomaly detector.</p>
Full article ">Figure 3
<p>SHAP values: top-20 features (readings).</p>
Full article ">Figure 4
<p>PCA applied to global consumption readings.</p>
Full article ">Figure 5
<p>PCA applied to cluster-based consumption readings (e.g., Cluster 1).</p>
Full article ">Figure 6
<p>PCA applied to the cluster-based SHAP explanations of consumption readings (e.g., Cluster 1).</p>
Full article ">Figure 7
<p>Comparison of the robustness of DAE+OCSVM with and without the proposed defense against CNN-based evasion attacks in terms of DR.</p>
Full article ">Figure 8
<p>Comparison of the robustness of DAE+OCSVM with and without the proposed defense against FFNN-based evasion attacks in terms of DR.</p>
Full article ">Figure 9
<p>PR and ROC curves of DAE+OCSVM with and without the proposed defense (XAI and clustering).</p>
Full article ">
20 pages, 1240 KiB  
Review
Handling the Imbalanced Problem in Agri-Food Data Analysis
by Adeyemi O. Adegbenjo and Michael O. Ngadi
Foods 2024, 13(20), 3300; https://doi.org/10.3390/foods13203300 - 17 Oct 2024
Viewed by 977
Abstract
Imbalanced data situations exist in most fields of endeavor. The problem has been identified as a major bottleneck in machine learning/data mining and is becoming a serious issue of concern in food processing applications. Inappropriate analysis of agricultural and food processing data was [...] Read more.
Imbalanced data situations exist in most fields of endeavor. The problem has been identified as a major bottleneck in machine learning/data mining and is becoming a serious issue of concern in food processing applications. Inappropriate analysis of agricultural and food processing data was identified as limiting the robustness of predictive models built from agri-food applications. As a result of rare cases occurring infrequently, classification rules that detect small groups are scarce, so samples belonging to small classes are largely misclassified. Most existing machine learning algorithms including the K-means, decision trees, and support vector machines (SVMs) are not optimal in handling imbalanced data. Consequently, models developed from the analysis of such data are very prone to rejection and non-adoptability in real industrial and commercial settings. This paper showcases the reality of the imbalanced data problem in agri-food applications and therefore proposes some state-of-the-art artificial intelligence algorithm approaches for handling the problem using methods including data resampling, one-class learning, ensemble methods, feature selection, and deep learning techniques. This paper further evaluates existing and newer metrics that are well suited for handling imbalanced data. Rightly analyzing imbalanced data from food processing application research works will improve the accuracy of results and model developments. This will consequently enhance the acceptability and adoptability of innovations/inventions. Full article
(This article belongs to the Special Issue Impacts of Innovative Processing Technologies on Food Quality)
Show Figures

Figure 1

Figure 1
<p>Receiver operating characteristic (ROC) curves for different classifiers: A—good model, B and C—poor models.</p>
Full article ">Figure 2
<p>Typical precision-recall curve for best threshold identification.</p>
Full article ">Figure 3
<p>Typical precision-recall curve for optimal model identification (PPV-positive predictive value (precision), SEN- sensitivity (recall), MD1-MD15: Model1-Model15).</p>
Full article ">
25 pages, 2699 KiB  
Article
Accurate Power Consumption Predictor and One-Class Electricity Theft Detector for Smart Grid “Change-and-Transmit” Advanced Metering Infrastructure
by Atef Bondok, Omar Abdelsalam, Mahmoud Badr, Mohamed Mahmoud, Maazen Alsabaan, Muteb Alsaqhan and Mohamed I. Ibrahem
Appl. Sci. 2024, 14(20), 9308; https://doi.org/10.3390/app14209308 - 12 Oct 2024
Viewed by 887
Abstract
The advanced metering infrastructure (AMI) of the smart grid plays a critical role in energy management and billing by enabling the periodic transmission of consumers’ power consumption readings. To optimize data collection efficiency, AMI employs a “change and transmit” (CAT) approach. This approach [...] Read more.
The advanced metering infrastructure (AMI) of the smart grid plays a critical role in energy management and billing by enabling the periodic transmission of consumers’ power consumption readings. To optimize data collection efficiency, AMI employs a “change and transmit” (CAT) approach. This approach ensures that readings are only transmitted when there is enough change in consumption, thereby reducing data traffic. Despite the benefits of this approach, it faces security challenges where malicious consumers can manipulate their readings to launch cyberattacks for electricity theft, allowing them to illegally reduce their bills. While this challenge has been addressed for supervised learning CAT settings, it remains insufficiently addressed in unsupervised learning settings. Moreover, due to the distortion introduced in the power consumption readings due to using the CAT approach, the accurate prediction of future consumption for energy management is a challenge. In this paper, we propose a two-stage approach to predict future readings and detect electricity theft in the smart grid while optimizing data collection using the CAT approach. For the first stage, we developed a predictor that is trained exclusively on benign CAT power consumption readings, and the output of the predictor is the actual readings. To enhance the prediction accuracy, we propose a cluster-based predictor that groups consumers into clusters with similar consumption patterns, and a dedicated predictor is trained for each cluster. For the second stage, we trained an autoencoder and a one-class support vector machine (SVM) on the benign reconstruction errors of the predictor to classify instances of electricity theft. We conducted comprehensive experiments to assess the effectiveness of our proposed approach. The experimental results indicate that the prediction error is very small and the accuracy of detection of the electricity theft attacks is high. Full article
(This article belongs to the Section Transportation and Future Mobility)
Show Figures

Figure 1

Figure 1
<p>The main entities and the flow of data in a smart grid AMI network.</p>
Full article ">Figure 2
<p>Flowchart of K-means clustering.</p>
Full article ">Figure 3
<p>One-class support vector machine (OC-SVM).</p>
Full article ">Figure 4
<p>Our proposed predictor architecture.</p>
Full article ">Figure 5
<p>Our proposed clustering approach.</p>
Full article ">Figure 6
<p>Selecting the number of clusters.</p>
Full article ">Figure 7
<p>Our proposed one-class classifier.</p>
Full article ">Figure 8
<p>MAAPE vs. used transmission approach for LSTM of global-based predictor.</p>
Full article ">Figure 9
<p>MAAPE vs. used transmission approach for transformer of global-based predictor.</p>
Full article ">Figure 10
<p>MAAPE vs. used transmission approach for LSTM + transformer of global-based predictor.</p>
Full article ">Figure 11
<p><span class="html-italic">MAAPE</span> vs. used transmission approach for LSTM of cluster-based predictor.</p>
Full article ">Figure 12
<p><math display="inline"><semantics> <mrow> <mi>M</mi> <mi>A</mi> <mi>A</mi> <mi>P</mi> <mi>E</mi> </mrow> </semantics></math> vs. used transmission approach for transformer of cluster-based predictor.</p>
Full article ">Figure 13
<p><math display="inline"><semantics> <mrow> <mi>M</mi> <mi>A</mi> <mi>A</mi> <mi>P</mi> <mi>E</mi> </mrow> </semantics></math> vs. used transmission approach for LSTM + transformer of cluster-based predictor.</p>
Full article ">Figure 14
<p>Loss of autoencoder vs. number of epochs.</p>
Full article ">
16 pages, 12099 KiB  
Article
Application of the Semi-Supervised Learning Approach for Pavement Defect Detection
by Peng Cui, Nurjihan Ala Bidzikrillah, Jiancong Xu and Yazhou Qin
Sensors 2024, 24(18), 6130; https://doi.org/10.3390/s24186130 - 23 Sep 2024
Cited by 1 | Viewed by 1129
Abstract
Road surface quality is essential for driver comfort and safety, making it crucial to monitor pavement conditions and detect defects in real time. However, the diversity of defects and the complexity of ambient conditions make it challenging to develop an effective and robust [...] Read more.
Road surface quality is essential for driver comfort and safety, making it crucial to monitor pavement conditions and detect defects in real time. However, the diversity of defects and the complexity of ambient conditions make it challenging to develop an effective and robust classification and detection algorithm. In this study, we adopted a semi-supervised learning approach to train ResNet-18 for image feature retrieval and then classification and detection of pavement defects. The resulting feature embedding vectors from image patches were retrieved, concatenated, and randomly sampled to model a multivariate normal distribution based on the only one-class training pavement image dataset. The calibration pavement image dataset was used to determine the defect score threshold based on the receiver operating characteristic curve, with the Mahalanobis distance employed as a metric to evaluate differences between normal and defect pavement images. Finally, a heatmap derived from the defect score map for the testing dataset was overlaid on the original pavement images to provide insight into the network’s decisions and guide measures to improve its performance. The results demonstrate that the model’s classification accuracy improved from 0.868 to 0.887 using the expanded and augmented pavement image data based on the analysis of heatmaps. Full article
Show Figures

Figure 1

Figure 1
<p>Diagram of the explainability of the intelligent models and their predictive power (courtesy of online).</p>
Full article ">Figure 2
<p>The flowchart of the study.</p>
Full article ">Figure 3
<p>Schematic diagram of the building block of ResNet-18.</p>
Full article ">Figure 4
<p>The architecture of the ResNet-18 network and the retrieval of feature embeddings (dash-dot lines denote change dimension; two convolutional layers are omitted in blocks 3 and 4 due to space limitation. Conv in <a href="#sensors-24-06130-f004" class="html-fig">Figure 4</a> stands for convolutional layer, 3×3 is the size of filters, 64 is the number of filters, and /2 denotes a stride of 2).</p>
Full article ">Figure 5
<p>Pavement image data collected from various roads: (<b>a</b>) Qingnian Road; (<b>b</b>) Riverside Road; and (<b>c</b>) Tongjing Dadao.</p>
Full article ">Figure 6
<p>Some samples of the pavement images (top: images without defects; bottom: images with defects).</p>
Full article ">Figure 7
<p>Histogram of the mean defect score for the calibration pavement dataset.</p>
Full article ">Figure 8
<p>The ROC curve for the calibration pavement dataset.</p>
Full article ">Figure 9
<p>The confusion matrix for the testing pavement dataset.</p>
Full article ">Figure 10
<p>The original image and heatmap image for the true positive pavement images: (<b>a</b>) the original pavement image and (<b>b</b>) the heatmap image of the true positive results.</p>
Full article ">Figure 11
<p>The original image and heatmap image for the true negative results: (<b>a</b>) the original pavement images and (<b>b</b>) the heatmap image of the true negative results.</p>
Full article ">Figure 12
<p>The original images and heatmaps for the false positive results: (<b>a</b>) the original images and (<b>b</b>) heatmaps for the false positive results.</p>
Full article ">Figure 13
<p>The original images and heatmaps for the false negative results: (<b>a</b>) the original pavement images and (<b>b</b>) the heatmaps for the false negative results.</p>
Full article ">Figure 13 Cont.
<p>The original images and heatmaps for the false negative results: (<b>a</b>) the original pavement images and (<b>b</b>) the heatmaps for the false negative results.</p>
Full article ">Figure 14
<p>The confusion matrix with the expanded image dataset.</p>
Full article ">
Back to TopTop