[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (20,234)

Search Parameters:
Keywords = deep-learning models

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
20 pages, 2044 KiB  
Article
Dynamic Task Planning for Multi-Arm Harvesting Robots Under Multiple Constraints Using Deep Reinforcement Learning
by Feng Xie , Zhengwei Guo, Tao Li, Qingchun Feng and Chunjiang Zhao
Horticulturae 2025, 11(1), 88; https://doi.org/10.3390/horticulturae11010088 - 14 Jan 2025
Abstract
Global fruit production costs are increasing amid intensified labor shortages, driving heightened interest in robotic harvesting technologies. Although multi-arm coordination in harvesting robots is considered a highly promising solution to this issue, it introduces technical challenges in achieving effective coordination. These challenges include [...] Read more.
Global fruit production costs are increasing amid intensified labor shortages, driving heightened interest in robotic harvesting technologies. Although multi-arm coordination in harvesting robots is considered a highly promising solution to this issue, it introduces technical challenges in achieving effective coordination. These challenges include mutual interference among multi-arm mechanical structures, task allocation across multiple arms, and dynamic operating conditions. This imposes higher demands on task coordination for multi-arm harvesting robots, requiring collision-free collaboration, optimization of task sequences, and dynamic re-planning. In this work, we propose a framework that models the task planning problem of multi-arm operation as a Markov game. First, considering multi-arm cooperative movement and picking sequence optimization, we employ a two-agent Markov game framework to model the multi-arm harvesting robot task planning problem. Second, we introduce a self-attention mechanism and a centralized training and execution strategy in the design and training of our deep reinforcement learning (DRL) model, thereby enhancing the model’s adaptability in dynamic and uncertain environments and improving decision accuracy. Finally, we conduct extensive numerical simulations in static environments; when the harvesting targets are set to 25 and 50, the execution time is reduced by 10.7% and 3.1%, respectively, compared to traditional methods. Additionally, in dynamic environments, both operational efficiency and robustness are superior to traditional approaches. The results underscore the potential of our approach to revolutionize multi-arm harvesting robotics by providing a more adaptive and efficient task planning solution. We will research improving the positioning accuracy of fruits in the future, which will make it possible to apply this framework to real robots. Full article
(This article belongs to the Section Fruit Production Systems)
39 pages, 5494 KiB  
Article
Learning Rate Tuner with Relative Adaptation (LRT-RA): Road to Sustainable Computing
by Saptarshi Biswas, Sumagna Dey and Subhrapratim Nath
AppliedMath 2025, 5(1), 8; https://doi.org/10.3390/appliedmath5010008 - 14 Jan 2025
Abstract
Optimizing learning rates (LRs) in deep learning (DL) has long been challenging. Previous solutions, such as learning rate scheduling (LRS) and adaptive learning rate (ALR) algorithms like RMSProp and Adam, added complexity by introducing new hyperparameters, thereby increasing the cost of model training [...] Read more.
Optimizing learning rates (LRs) in deep learning (DL) has long been challenging. Previous solutions, such as learning rate scheduling (LRS) and adaptive learning rate (ALR) algorithms like RMSProp and Adam, added complexity by introducing new hyperparameters, thereby increasing the cost of model training through expensive cross-validation experiments. These methods mainly focus on local gradient patterns, which may not be effective in scenarios with multiple local optima near the global optimum. A new technique called Learning Rate Tuner with Relative Adaptation (LRT-RA) is introduced to tackle these issues. This approach dynamically adjusts LRs during training by analyzing the global loss curve, eliminating the need for costly initial LR estimation through cross-validation. This method reduces training expenses and carbon footprint and enhances training efficiency. It demonstrates promising results in preventing premature convergence, exhibiting inherent optimization behavior, and elucidating the correlation between dataset distribution and optimal LR selection. The proposed method achieves 84.96% accuracy on the CIFAR-10 dataset while reducing the power usage to 0.07 kWh, CO2 emissions to 0.05, and both SO2 and NOx emissions to 0.00003 pounds, during the whole training and testing process. Full article
(This article belongs to the Special Issue Optimization and Machine Learning)
Show Figures

Figure 1

Figure 1
<p>Theoretical trend curve for possible combinations to check for in cross-validation experiments.</p>
Full article ">Figure 2
<p>Cyclic Learning Rate (triangular) [<a href="#B28-appliedmath-05-00008" class="html-bibr">28</a>].</p>
Full article ">Figure 3
<p>Accuracy plots for different experiments.</p>
Full article ">Figure 4
<p>Linear regression on Boston House Pricing dataset.</p>
Full article ">Figure 5
<p>Linear regression on Red Wine Quality Prediction dataset.</p>
Full article ">Figure 6
<p>Behavior of learning rate.</p>
Full article ">Figure 6 Cont.
<p>Behavior of learning rate.</p>
Full article ">Figure 7
<p>Learning rate plots for different settings of <math display="inline"><semantics> <mrow> <mi>γ</mi> <mo>,</mo> <msub> <mi>ϵ</mi> <mn>1</mn> </msub> </mrow> </semantics></math>, and <math display="inline"><semantics> <msub> <mi>ϵ</mi> <mn>2</mn> </msub> </semantics></math>.</p>
Full article ">Figure 7 Cont.
<p>Learning rate plots for different settings of <math display="inline"><semantics> <mrow> <mi>γ</mi> <mo>,</mo> <msub> <mi>ϵ</mi> <mn>1</mn> </msub> </mrow> </semantics></math>, and <math display="inline"><semantics> <msub> <mi>ϵ</mi> <mn>2</mn> </msub> </semantics></math>.</p>
Full article ">Figure 8
<p>Analysis of cross-validation experiment. (<b>a</b>) Analysis of power usage and <math display="inline"><semantics> <mrow> <mi>C</mi> <msub> <mi>O</mi> <mn>2</mn> </msub> </mrow> </semantics></math> emission. (<b>b</b>) Analysis of <math display="inline"><semantics> <mrow> <mi>S</mi> <msub> <mi>O</mi> <mn>2</mn> </msub> </mrow> </semantics></math> and <math display="inline"><semantics> <mrow> <mi>N</mi> <msub> <mi>O</mi> <mi>x</mi> </msub> </mrow> </semantics></math> emission. (<b>c</b>) Effect of Hyperparameter reduction for grid search on Adam. (<b>d</b>) Effect of hyperparameter reduction for grid search on Cyclic Learning Rate Scheduler (exp_range). (<b>e</b>) Effect of hyperparameter reduction for grid search on SGD with Warm Restart.</p>
Full article ">Figure 8 Cont.
<p>Analysis of cross-validation experiment. (<b>a</b>) Analysis of power usage and <math display="inline"><semantics> <mrow> <mi>C</mi> <msub> <mi>O</mi> <mn>2</mn> </msub> </mrow> </semantics></math> emission. (<b>b</b>) Analysis of <math display="inline"><semantics> <mrow> <mi>S</mi> <msub> <mi>O</mi> <mn>2</mn> </msub> </mrow> </semantics></math> and <math display="inline"><semantics> <mrow> <mi>N</mi> <msub> <mi>O</mi> <mi>x</mi> </msub> </mrow> </semantics></math> emission. (<b>c</b>) Effect of Hyperparameter reduction for grid search on Adam. (<b>d</b>) Effect of hyperparameter reduction for grid search on Cyclic Learning Rate Scheduler (exp_range). (<b>e</b>) Effect of hyperparameter reduction for grid search on SGD with Warm Restart.</p>
Full article ">
18 pages, 1353 KiB  
Article
Enhancing Privacy While Preserving Context in Text Transformations by Large Language Models
by Tymon Lesław Żarski and Artur Janicki
Information 2025, 16(1), 49; https://doi.org/10.3390/info16010049 - 14 Jan 2025
Abstract
Data security is a critical concern for Internet users, primarily as more people rely on social networks and online tools daily. Despite the convenience, many users are unaware of the risks posed to their sensitive and personal data. This study addresses this issue [...] Read more.
Data security is a critical concern for Internet users, primarily as more people rely on social networks and online tools daily. Despite the convenience, many users are unaware of the risks posed to their sensitive and personal data. This study addresses this issue by presenting a comprehensive solution to prevent personal data leakage using online tools. We developed a conceptual solution that enhances user privacy by identifying and anonymizing named entity classes representing sensitive data while maintaining the original context by swapping source entities for functional data. Our approach utilizes natural language processing methods, combining machine learning tools such as MITIE and spaCy with rule-based text analysis. We employed regular expressions and large language models to anonymize text, preserving its context for further processing or enabling restoration to the original form after transformations. The results demonstrate the effectiveness of our custom-trained models, achieving an F1 score of 0.8292. Additionally, the proposed algorithms successfully preserved context in approximately 93.23% of test cases, indicating a promising solution for secure data handling in online environments. Full article
Show Figures

Figure 1

Figure 1
<p>Processing flow and visualization of system components.</p>
Full article ">Figure 2
<p>Display of parallel coordinates with resulting F1 score of the run.</p>
Full article ">Figure 3
<p>Detection page with selected MITIE NER model.</p>
Full article ">Figure 4
<p>Detection page displaying (<b>a</b>) the statistics and (<b>b</b>) results of NER of the provided text, (<b>c</b>) the results of the anonymization process models, and (<b>d</b>) summarized text with received previously anonymized personal information.</p>
Full article ">Figure 5
<p>Confusionmatrix of the SVM MITIE model.</p>
Full article ">Figure 6
<p>Confusion matrix of the CNN spaCy model.</p>
Full article ">
37 pages, 2498 KiB  
Review
Application of Convolutional Neural Networks and Recurrent Neural Networks in Food Safety
by Haohan Ding, Haoke Hou, Long Wang, Xiaohui Cui, Wei Yu and David I. Wilson
Foods 2025, 14(2), 247; https://doi.org/10.3390/foods14020247 - 14 Jan 2025
Abstract
This review explores the application of convolutional neural networks (CNNs) and recurrent neural networks (RNNs) in food safety detection and risk prediction. This paper highlights the advantages of CNNs in image processing and feature recognition, as well as the powerful capabilities of RNNs [...] Read more.
This review explores the application of convolutional neural networks (CNNs) and recurrent neural networks (RNNs) in food safety detection and risk prediction. This paper highlights the advantages of CNNs in image processing and feature recognition, as well as the powerful capabilities of RNNs (especially their variant LSTM) in time series data modeling. This paper also makes a comparative analysis in many aspects: Firstly, the advantages and disadvantages of traditional food safety detection and risk prediction methods are compared with deep learning technologies such as CNNs and RNNs. Secondly, the similarities and differences between CNNs and fully connected neural networks in processing image data are analyzed. Furthermore, the advantages and disadvantages of RNNs and traditional statistical modeling methods in processing time series data are discussed. Finally, the application directions of CNNs in food safety detection and RNNs in food safety risk prediction are compared. This paper also discusses combining these deep learning models with technologies such as the Internet of Things (IoT), blockchain, and federated learning to improve the accuracy and efficiency of food safety detection and risk warning. Finally, this paper mentions the limitations of RNNs and CNNs in the field of food safety, as well as the challenges in the interpretability of the model, and suggests the use of interpretable artificial intelligence (XAI) technology to improve the transparency of the model. Full article
Show Figures

Figure 1

Figure 1
<p>Application of CNNs and RNNs in food safety.</p>
Full article ">Figure 2
<p>Basic architecture of CNNs. “•••” denotes omitted convolutional layers, arrows indicate data flow and processing steps.</p>
Full article ">Figure 3
<p>Basic architecture of an RNN. Arrows indicate data flow and processing steps, subscripts denote different nodes, different colors of “•••” indicate that intermediate nodes are omitted at different time steps or in different layers.</p>
Full article ">Figure 4
<p>Internal structure of an LSTM.</p>
Full article ">Figure 5
<p>Workflow of food safety testing using CNN. “•••” indicate more application scenarios for food safety testing.</p>
Full article ">Figure 6
<p>Future outlook, limitations, and challenges.</p>
Full article ">
32 pages, 6487 KiB  
Article
FS-DDPG: Optimal Control of a Fan Coil Unit System Based on Safe Reinforcement Learning
by Chenyang Li, Qiming Fu, Jianping Chen, You Lu, Yunzhe Wang and Hongjie Wu
Buildings 2025, 15(2), 226; https://doi.org/10.3390/buildings15020226 - 14 Jan 2025
Abstract
To optimize the control of fan coil unit (FCU) systems under model-free conditions, researchers have integrated reinforcement learning (RL) into the control processes of system pumps and fans. However, traditional RL methods can lead to significant fluctuations in the flow of pumps and [...] Read more.
To optimize the control of fan coil unit (FCU) systems under model-free conditions, researchers have integrated reinforcement learning (RL) into the control processes of system pumps and fans. However, traditional RL methods can lead to significant fluctuations in the flow of pumps and fans, posing a safety risk. To address this issue, we propose a novel FCU control method, Fluctuation Suppression–Deep Deterministic Policy Gradient (FS-DDPG). The key innovation lies in applying a constrained Markov decision process to model the FCU control problem, where a penalty term for process constraints is incorporated into the reward function, and constraint tightening is introduced to limit the action space. In addition, to validate the performance of the proposed method, we established a variable operating conditions FCU simulation platform based on the parameters of the actual FCU system and ten years of historical weather data. The platform’s correctness and effectiveness were verified from three aspects: heat transfer, the air side and the water side, under different dry and wet operating conditions. The experimental results show that compared with DDPG, FS-DDPG avoids 98.20% of the pump flow and 95.82% of the fan flow fluctuations, ensuring the safety of the equipment. Compared with DDPG and RBC, FS-DDPG achieves 11.9% and 51.76% energy saving rates, respectively, and also shows better performance in terms of operational performance and satisfaction. In the future, we will further improve the scalability and apply the method to more complex FCU systems in variable environments. Full article
(This article belongs to the Section Building Energy, Physics, Environment, and Systems)
Show Figures

Figure 1

Figure 1
<p>The traditional reinforcement learning method controls part of the water flow and air flow in an FCU system.</p>
Full article ">Figure 2
<p>Fan coil system working diagram.</p>
Full article ">Figure 3
<p>Control flow.</p>
Full article ">Figure 4
<p>Room model.</p>
Full article ">Figure 5
<p>Cooling load data for room 1-N-6 in 2021.</p>
Full article ">Figure 6
<p>Input and output simulation model of FCU.</p>
Full article ">Figure 7
<p>Markov decision model of fan coil system.</p>
Full article ">Figure 8
<p>FS-DDPG control method based on SRL.</p>
Full article ">Figure 9
<p>RBC sequence decision control logic.</p>
Full article ">Figure 10
<p>Cumulative reward.</p>
Full article ">Figure 11
<p>Average annual power consumption comparison.</p>
Full article ">Figure 12
<p>Average annual relative error distribution of water flow by FS-DDPG, DDPG and MBC.</p>
Full article ">Figure 13
<p>Average annual relative error distribution of air flow by FS-DDPG, DDPG and MBC.</p>
Full article ">Figure 14
<p>Annual relative error of water flow.</p>
Full article ">Figure 15
<p>Annual relative error of air flow.</p>
Full article ">Figure 16
<p>Cooling load deviation under the FS-DDPG control method.</p>
Full article ">Figure A1
<p>Water/air flow power diagram.</p>
Full article ">
21 pages, 2525 KiB  
Article
A Data-Driven Deep Learning Framework for Prediction of Traffic Crashes at Road Intersections
by Mengxiang Wang, Wang-Chien Lee, Na Liu, Qiang Fu, Fujun Wan and Ge Yu
Appl. Sci. 2025, 15(2), 752; https://doi.org/10.3390/app15020752 - 14 Jan 2025
Abstract
Traffic crash prediction (TCP) is a fundamental problem for intelligent transportation systems in smart cities. Improving the accuracy of traffic crash prediction is important for road safety and effective traffic management. Owing to recent advances in artificial neural networks, several new deep-learning models [...] Read more.
Traffic crash prediction (TCP) is a fundamental problem for intelligent transportation systems in smart cities. Improving the accuracy of traffic crash prediction is important for road safety and effective traffic management. Owing to recent advances in artificial neural networks, several new deep-learning models have been proposed for TCP. However, these works mainly focus on accidents in regions, which are typically pre-determined using a grid map. We argue that TCP for roads, especially for crashes at or near road intersections which account for more than 50% of the fatal or injury crashes based on the Federal Highway Administration, has a significant practical and research value and thus deserves more research. In this paper, we formulate TCP at Road Intersections as a classification problem and propose a three-phase data-driven deep learning model, called Road Intersection Traffic Crash Prediction (RoadInTCP), to predict traffic crashes at intersections by exploiting publicly available heterogeneous big data. In Phase I we extract discriminative latent features called topological-relational features (tr-features), of intersections using a neural network model by exploiting topological information of the road network and various relationships amongst nearby intersections. In Phase II, in addition to tr-features which capture some inherent properties of the road network, we also explore additional thematic information in terms of environmental, traffic, weather, risk, and calendar features associated with intersections. In order to incorporate the potential correlation in nearby intersections, we utilize a Graph Convolution Network (GCN) to aggregate features from neighboring intersections based on a message-passing paradigm for TCP. While Phase II serves well as a TCP model, we further explore the signals embedded in the sequential feature changes over time for TCP in Phase III, by exploring RNN or 1DCNN which have known success on sequential data. Additionally, to address the serious issues of imbalanced classes in TCP and large-scale heterogeneous big data, we propose an effective data sampling approach in data preparation to facilitate model training. We evaluate the proposed RoadInTCP model via extensive experiments on a real-world New York City traffic dataset. The experimental results show that the proposed RoadInTCP robustly outperforms existing methods. Full article
Show Figures

Figure 1

Figure 1
<p>A portion distribution of traffic crashes on NYC road networks in 2016. The solid point means a vehicle crash occurred in the location and the hollow circle means the multiple vehicle crashes occurred in the same location.</p>
Full article ">Figure 2
<p>The Framework of RoadInTCP.</p>
Full article ">Figure 3
<p>The Relationship between Traffic Crashes and Weather Types at Road Intersections from 2014 to 2016.</p>
Full article ">Figure 4
<p>(<b>a</b>) The relationship between the number of traffic crashes and one-way road segments. (<b>b</b>) The relationship between the number of traffic crashes and traffic lanes.</p>
Full article ">Figure 5
<p>The process of RoadInTCP in Phase II.</p>
Full article ">Figure 6
<p>(<b>a</b>) Weekly traffic crash curves of a specific intersection. (<b>b</b>) Hourly traffic crash curves of a specific intersection.</p>
Full article ">Figure 7
<p>An example of input time series to RoadInTCP in Phase III.</p>
Full article ">Figure 8
<p>Parameter analysis for RoadInTCP model in Phase I of traffic signal classification.</p>
Full article ">Figure 9
<p>Parameter analysis for RoadInTCP model in Phase I of avenue classification.</p>
Full article ">
22 pages, 2713 KiB  
Article
An Efficient 3D Convolutional Neural Network for Dose Prediction in Cancer Radiotherapy from CT Images
by Lam Thanh Hien, Pham Trung Hieu and Do Nang Toan
Diagnostics 2025, 15(2), 177; https://doi.org/10.3390/diagnostics15020177 - 14 Jan 2025
Abstract
Introduction: Cancer is a highly lethal disease with a significantly high mortality rate. One of the most commonly used methods for treatment is radiation therapy. However, cancer treatment using radiotherapy is a time-consuming process that requires significant manual work from planners and [...] Read more.
Introduction: Cancer is a highly lethal disease with a significantly high mortality rate. One of the most commonly used methods for treatment is radiation therapy. However, cancer treatment using radiotherapy is a time-consuming process that requires significant manual work from planners and doctors. In radiation therapy treatment planning, determining the dose distribution for each of the regions of the patient’s body is one of the most difficult and important tasks. Nowadays, artificial intelligence has shown promising results in improving the quality of disease treatment, particularly in cancer radiation therapy. Objectives: The main objective of this study is to build a high-performance deep learning model for predicting radiation therapy doses for cancer and to develop software to easily manipulate and use this model. Materials and Methods: In this paper, we propose a custom 3D convolutional neural network model with a U-Net-based architecture to automatically predict radiation doses during cancer radiation therapy from CT images. To ensure that the predicted doses do not have negative values, which are not valid for radiation doses, a rectified linear unit (ReLU) function is applied to the output to convert negative values to zero. Additionally, a proposed loss function based on a dose–volume histogram is used to train the model, ensuring that the predicted dose concentrations are highly meaningful in terms of radiation therapy. The model is developed using the OpenKBP challenge dataset, which consists of 200, 100, and 40 head and neck cancer patients for training, testing, and validation, respectively. Before the training phase, preprocessing and augmentation techniques, such as standardization, translation, and flipping, are applied to the training set. During the training phase, a cosine annealing scheduler is applied to update the learning rate. Results and Conclusions: Our model achieved strong performance, with a good DVH score (1.444 Gy) on the test dataset, compared to previous studies and state-of-the-art models. In addition, we developed software to display the dose maps predicted by the proposed model for each 2D slice in order to facilitate usage and observation. These results may help doctors in treating cancer with radiation therapy in terms of both time and effectiveness. Full article
(This article belongs to the Special Issue Deep Learning Techniques for Medical Image Analysis)
Show Figures

Figure 1

Figure 1
<p>Illustration of a 2D slice image of a patient. The first image is a CT image, the second image contains information about the PTV areas, and the last image is the corresponding radiation therapy dose.</p>
Full article ">Figure 2
<p>The architecture of our proposed model.</p>
Full article ">Figure 3
<p>Overview of cascade learning in deep learning.</p>
Full article ">Figure 4
<p>Residual block.</p>
Full article ">Figure 5
<p>Flowchart representing the training, testing, and implementation phases.</p>
Full article ">Figure 6
<p>The total loss of our model on the training and validation datasets.</p>
Full article ">Figure 7
<p>The loss of Model A on the training and validation datasets.</p>
Full article ">Figure 8
<p>The loss of Model B on the training and validation datasets.</p>
Full article ">Figure 9
<p>The difference between the predicted and ground-truth DVH values of our model on the test set.</p>
Full article ">Figure 10
<p>Comparison of the predicted (dashed lines) and ground-truth (solid lines) dose–volume histograms for three patients: 274, 279, and 313.</p>
Full article ">Figure 11
<p>Three-dimensional dose distributions for three patients: 274, 279, and 313.</p>
Full article ">Figure 12
<p>The interface of the software for predicting the radiation dose.</p>
Full article ">
40 pages, 7115 KiB  
Article
Emotion Recognition from EEG Signals Using Advanced Transformations and Deep Learning
by Jonathan Axel Cruz-Vazquez, Jesús Yaljá Montiel-Pérez, Rodolfo Romero-Herrera and Elsa Rubio-Espino
Mathematics 2025, 13(2), 254; https://doi.org/10.3390/math13020254 - 14 Jan 2025
Abstract
Affective computing aims to develop systems capable of effectively interacting with people through emotion recognition. Neuroscience and psychology have established models that classify universal human emotions, providing a foundational framework for developing emotion recognition systems. Brain activity related to emotional states can be [...] Read more.
Affective computing aims to develop systems capable of effectively interacting with people through emotion recognition. Neuroscience and psychology have established models that classify universal human emotions, providing a foundational framework for developing emotion recognition systems. Brain activity related to emotional states can be captured through electroencephalography (EEG), enabling the creation of models that classify emotions even in uncontrolled environments. In this study, we propose an emotion recognition model based on EEG signals using deep learning techniques on a proprietary database. To improve the separability of emotions, we explored various data transformation techniques, including Fourier Neural Networks and quantum rotations. The convolutional neural network model, combined with quantum rotations, achieved a 95% accuracy in emotion classification, particularly in distinguishing sad emotions. The integration of these transformations can further enhance overall emotion recognition performance. Full article
(This article belongs to the Special Issue Deep Neural Networks: Theory, Algorithms and Applications)
Show Figures

Figure 1

Figure 1
<p>Emotion classification processing workflow with EEG signals.</p>
Full article ">Figure 2
<p>Recording protocol and stimulus exposure at different times of the day.</p>
Full article ">Figure 3
<p>Recording room for the experiment.</p>
Full article ">Figure 4
<p>(<b>a</b>) Diagram of electrode positions on the scalp according to the 10–20 system; (<b>b</b>) representation of brain regions corresponding to the electrodes.</p>
Full article ">Figure 5
<p>Graph of raw EEG signals from 14 channels over time.</p>
Full article ">Figure 6
<p>Notch filter application: (<b>a</b>) before applying the notch filter, (<b>b</b>) after applying the notch filter.</p>
Full article ">Figure 7
<p>Spatial distribution of independent components (ICA).</p>
Full article ">Figure 8
<p>Detailed analysis of ICA004 component: (<b>a</b>) topographic map, (<b>b</b>) segment image and ERP/ERF, (<b>c</b>) frequency spectrum, (<b>d</b>) dropped segments.</p>
Full article ">Figure 9
<p>Compares EEG signals before and after cleaning artifacts using ICA.</p>
Full article ">Figure 10
<p>Recording of EEG signals from 14 channels during ERP segmentation.</p>
Full article ">Figure 11
<p>Average ERP signals across the 14 EEG channels.</p>
Full article ">Figure 12
<p>Topographic maps of temporal evolution of an ERP.</p>
Full article ">Figure 13
<p>Representation of the 14 EEG channels and topographic maps of an ERP.</p>
Full article ">Figure 14
<p>Scatter plots of extracted features.</p>
Full article ">Figure 15
<p>Correlation matrix of EEG time and frequency domain features.</p>
Full article ">Figure 16
<p>Distribution of features transformed with the Fourier Neural Network.</p>
Full article ">Figure 17
<p>Quantum rotations: (<b>a</b>) emotions before applying quantum rotations, (<b>b</b>) emotions after applying quantum rotations.</p>
Full article ">Figure 18
<p>Quantum rotated features for different emotional states: (<b>a</b>) quantum-rotated features for happy, (<b>b</b>) quantum-rotated features for sad, (<b>c</b>) quantum-rotated features for neutral.</p>
Full article ">Figure 19
<p>Performance of dense network with Fourier features: (<b>a</b>) confusion matrix, (<b>b</b>) precision curves for training and validation, (<b>c</b>) loss curves for training and validation.</p>
Full article ">Figure 20
<p>Performance of dense network with quantum-rotated features: (<b>a</b>) confusion matrix, (<b>b</b>) precision curves for training and validation, (<b>c</b>) loss curves for training and validation.</p>
Full article ">Figure 21
<p>Performance of convolutional neural network (CNN) with Fourier features: (<b>a</b>) confusion matrix, (<b>b</b>) precision curves for training and validation, (<b>c</b>) loss curves for training and validation.</p>
Full article ">Figure 22
<p>Performance of convolutional neural network (CNN) with quantum-rotated features: (<b>a</b>) confusion matrix, (<b>b</b>) precision curves for training and validation, (<b>c</b>) loss curves for training and validation.</p>
Full article ">
31 pages, 2149 KiB  
Article
Enhanced Deep Autoencoder-Based Reinforcement Learning Model with Improved Flamingo Search Policy Selection for Attack Classification
by Dharani Kanta Roy and Hemanta Kumar Kalita
J. Cybersecur. Priv. 2025, 5(1), 3; https://doi.org/10.3390/jcp5010003 - 14 Jan 2025
Abstract
Intrusion detection has been a vast-surveyed topic for many decades as network attacks are tremendously growing. This has heightened the need for security in networks as web-based communication systems are advanced nowadays. The proposed work introduces an intelligent semi-supervised intrusion detection system based [...] Read more.
Intrusion detection has been a vast-surveyed topic for many decades as network attacks are tremendously growing. This has heightened the need for security in networks as web-based communication systems are advanced nowadays. The proposed work introduces an intelligent semi-supervised intrusion detection system based on different algorithms to classify the network attacks accurately. Initially, the pre-processing is accomplished using null value dropping and standard scaler normalization. After pre-processing, an enhanced Deep Reinforcement Learning (EDRL) model is employed to extract high-level representations and learn complex patterns from data by means of interaction with the environment. The enhancement of deep reinforcement learning is made by associating a deep autoencoder (AE) and an improved flamingo search algorithm (IFSA) to approximate the Q-function and optimal policy selection. After feature representations, a support vector machine (SVM) classifier, which discriminates the input into normal and attack instances, is employed for classification. The presented model is simulated in the Python platform and evaluated using the UNSW-NB15, CICIDS2017, and NSL-KDD datasets. The overall classification accuracy is 99.6%, 99.93%, and 99.42% using UNSW-NB15, CICIDS2017, and NSL-KDD datasets, which is higher than the existing detection frameworks. Full article
Show Figures

Figure 1

Figure 1
<p>Flow diagram of the proposed attack classification and response scheme.</p>
Full article ">Figure 2
<p>Architecture of enhanced deep reinforcement learning.</p>
Full article ">Figure 3
<p>Semi-supervised workflow.</p>
Full article ">Figure 4
<p>Deep AE model used for approximating the Q-function.</p>
Full article ">Figure 5
<p>Model of SVM used for attack classification.</p>
Full article ">Figure 6
<p>Confusion matrix of the proposed work.</p>
Full article ">Figure 7
<p>Precision comparison of the proposed and existing works.</p>
Full article ">Figure 8
<p>Recall the comparison of the proposed and existing works.</p>
Full article ">Figure 9
<p>F1-score comparison of the proposed and existing works.</p>
Full article ">Figure 10
<p>Accuracy comparison of the proposed and existing works.</p>
Full article ">Figure 11
<p>FPR comparison of the proposed and existing works.</p>
Full article ">Figure 12
<p>Kappa score comparison of the proposed and existing works.</p>
Full article ">Figure 13
<p>MCC comparison of the proposed and existing works.</p>
Full article ">Figure 14
<p>RoC curve analysis of the proposed work.</p>
Full article ">
18 pages, 1035 KiB  
Article
Bidirectional f-Divergence-Based Deep Generative Method for Imputing Missing Values in Time-Series Data
by Wen-Shan Liu, Tong Si, Aldas Kriauciunas, Marcus Snell and Haijun Gong
Stats 2025, 8(1), 7; https://doi.org/10.3390/stats8010007 - 14 Jan 2025
Abstract
Imputing missing values in high-dimensional time-series data remains a significant challenge in statistics and machine learning. Although various methods have been proposed in recent years, many struggle with limitations and reduced accuracy, particularly when the missing rate is high. In this work, we [...] Read more.
Imputing missing values in high-dimensional time-series data remains a significant challenge in statistics and machine learning. Although various methods have been proposed in recent years, many struggle with limitations and reduced accuracy, particularly when the missing rate is high. In this work, we present a novel f-divergence-based bidirectional generative adversarial imputation network, tf-BiGAIN, designed to address these challenges in time-series data imputation. Unlike traditional imputation methods, tf-BiGAIN employs a generative model to synthesize missing values without relying on distributional assumptions. The imputation process is achieved by training two neural networks, implemented using bidirectional modified gated recurrent units, with f-divergence serving as the objective function to guide optimization. Compared to existing deep learning-based methods, tf-BiGAIN introduces two key innovations. First, the use of f-divergence provides a flexible and adaptable framework for optimizing the model across diverse imputation tasks, enhancing its versatility. Second, the use of bidirectional gated recurrent units allows the model to leverage both forward and backward temporal information. This bidirectional approach enables the model to effectively capture dependencies from both past and future observations, enhancing its imputation accuracy and robustness. We applied tf-BiGAIN to analyze two real-world time-series datasets, demonstrating its superior performance in imputing missing values and outperforming existing methods in terms of accuracy and robustness. Full article
Show Figures

Figure 1

Figure 1
<p>An illustration of the bidirectional f-divergence based generative adversarial imputation network architecture. The Generator takes as input the incomplete time-series data, a corresponding mask matrix, and a time-lag matrix, along with a random matrix, to generate synthetic imputed data. The Discriminator distinguishes between real and imputed values generated by the Generator. Both the Generator and Discriminator are implemented using bidirectional GRUI and trained using f-divergence loss functions.</p>
Full article ">Figure 2
<p>Structure of gated recurrent unit (left) and modified gated recurrent unit (right).</p>
Full article ">Figure 3
<p>AUC scores for mortality prediction in PhysioNet data using an RNN classifier trained on unidirectional and bidirectional GRUI-based generative adversarial imputation networks with various f-divergence functions as the adversarial loss.</p>
Full article ">Figure 4
<p>AUC scores for air pollution prediction in KDD air quality data using an RNN classifier trained on unidirectional GRUI-based generative adversarial imputation networks with various f-divergence functions as the adversarial loss across varying missing rates.</p>
Full article ">Figure 5
<p>AUC scores for air pollution prediction in KDD air quality data using an RNN classifier trained on bidirectional GRUI-based generative adversarial imputation networks with various f-divergence functions as the adversarial loss across varying missing rates.</p>
Full article ">
18 pages, 6983 KiB  
Article
Multiscale Convolution-Based Efficient Channel Estimation Techniques for OFDM Systems
by Nahyeon Kwon, Bora Yoon and Junghyun Kim
Electronics 2025, 14(2), 307; https://doi.org/10.3390/electronics14020307 - 14 Jan 2025
Viewed by 121
Abstract
With the advancement of wireless communication technology, the significance of efficient and accurate channel estimation methods has grown substantially. Recently, deep learning-based methods are being adopted to estimate channels with higher precision than traditional methods, even in the absence of prior channel statistics. [...] Read more.
With the advancement of wireless communication technology, the significance of efficient and accurate channel estimation methods has grown substantially. Recently, deep learning-based methods are being adopted to estimate channels with higher precision than traditional methods, even in the absence of prior channel statistics. In this paper, we propose two deep learning-based channel estimation models, CAMPNet and MSResNet, which are designed to consider channel characteristics from a multiscale perspective. The convolutional attention and multiscale parallel network (CAMPNet) accentuates critical channel characteristics by utilizing parallel multiscale features and convolutional attention, while the multiscale residual network (MSResNet) integrates information across various scales through cross-connected multiscale convolutional structures. Both models are designed to perform robustly in environments with complex frequency domain information and various Doppler shifts. Experimental results demonstrate that CAMPNet and MSResNet achieve superior performance compared to existing channel estimation methods within various channel models. Notably, the proposed models show exceptional performance in high signal-to-noise ratio (SNR) environments, achieving up to a 48.98% reduction in mean squared error(MSE) compared to existing methods at an SNR of 25dB. In experiments evaluating the generalization capabilities of the proposed models, they show greater stability and robustness compared to existing methods. These results suggest that deep learning-based channel estimation models have the potential to overcome the limitations of existing methods, offering high performance and efficiency in real-world communication environments. Full article
(This article belongs to the Section Circuit and Signal Processing)
Show Figures

Figure 1

Figure 1
<p>Flowchart of pilot-based channel estimation in SISO communication system.</p>
Full article ">Figure 2
<p>The structure of ReEsNet [<a href="#B17-electronics-14-00307" class="html-bibr">17</a>] and Interpolation-ResNet [<a href="#B18-electronics-14-00307" class="html-bibr">18</a>]. (<b>a</b>) ReEsNet; (<b>b</b>) Interpolation-ResNet.</p>
Full article ">Figure 3
<p>The structure of proposed CAMPNet.</p>
Full article ">Figure 4
<p>The structure of proposed MSResNet.</p>
Full article ">Figure 5
<p>Binary interpolation.</p>
Full article ">Figure 6
<p>Comparison of MSE performance under EPA and ETU channel model across different SNR values for various channel estimation models; ReEsNet [<a href="#B17-electronics-14-00307" class="html-bibr">17</a>], Interpolation-ResNet [<a href="#B18-electronics-14-00307" class="html-bibr">18</a>], AttenReEsNet [<a href="#B33-electronics-14-00307" class="html-bibr">33</a>], proposed CAMPNet, and proposed MSResNet. (<b>a</b>) EPA channel model. (<b>b</b>) ETU channel model.</p>
Full article ">Figure 7
<p>Generalization capability comparison of various channel estimation models; ReEsNet [<a href="#B17-electronics-14-00307" class="html-bibr">17</a>], Interpolation-ResNet [<a href="#B18-electronics-14-00307" class="html-bibr">18</a>], AttenReEsNet [<a href="#B33-electronics-14-00307" class="html-bibr">33</a>], proposed CAMPNet, and proposed MSResNet. (<b>a</b>) MSE curves of models trained on the EPA channel and tested on the ETU channel. (<b>b</b>) MSE curves of models trained on the ETU channel and tested on the EPA channel. Our proposed models, CAMPNet and MSResNet, trained on the ETA channel do not adapt well to the ETU channel, whereas those trained on the ETU channel demonstrate excellent performance on the ETA channel.</p>
Full article ">Figure 8
<p>Performances of the estimators for the different Doppler shifts for various channel estimation models; ReEsNet [<a href="#B17-electronics-14-00307" class="html-bibr">17</a>], Interpolation-ResNet [<a href="#B18-electronics-14-00307" class="html-bibr">18</a>], AttenReEsNet [<a href="#B33-electronics-14-00307" class="html-bibr">33</a>], proposed CAMPNet, and proposed MSResNet. The CAMPNet and MSResNet adapt better to Doppler variations compared to ReEsNet and Interpolation-ResNet while demonstrating performance similar to that of AttenReEsNet.</p>
Full article ">
21 pages, 282 KiB  
Article
Foodborne Event Detection Based on Social Media Mining: A Systematic Review
by Silvano Salaris, Honoria Ocagli, Alessandra Casamento, Corrado Lanera and Dario Gregori
Foods 2025, 14(2), 239; https://doi.org/10.3390/foods14020239 - 14 Jan 2025
Viewed by 258
Abstract
Foodborne illnesses represent a significant global health challenge, causing substantial morbidity and mortality. Conventional surveillance methods, such as laboratory-based reporting and physician notifications, often fail to enable early detection, prompting the exploration of innovative solutions. Social media platforms, combined with machine learning (ML), [...] Read more.
Foodborne illnesses represent a significant global health challenge, causing substantial morbidity and mortality. Conventional surveillance methods, such as laboratory-based reporting and physician notifications, often fail to enable early detection, prompting the exploration of innovative solutions. Social media platforms, combined with machine learning (ML), offer new opportunities for real-time monitoring and outbreak analysis. This systematic review evaluated the role of social networks in detecting and managing foodborne illnesses, particularly through the use of ML techniques to identify unreported events and enhance outbreak response. This review analyzed studies published up to December 2024 that utilized social media data and data mining to predict and prevent foodborne diseases. A comprehensive search was conducted across PubMed, EMBASE, CINAHL, Arxiv, Scopus, and Web of Science databases, excluding clinical trials, case reports, and reviews. Two independent reviewers screened studies using Covidence, with a third resolving conflicts. Study variables included social media platforms, ML techniques (shallow and deep learning), and model performance, with a risk of bias assessed using the PROBAST tool. The results highlighted Twitter and Yelp as primary data sources, with shallow learning models dominating the field. Many studies were identified as having high or unclear risk of bias. This review underscored the potential of social media and ML in foodborne disease surveillance and emphasizes the need for standardized methodologies and further exploration of deep learning models. Full article
(This article belongs to the Section Food Microbiology)
19 pages, 983 KiB  
Article
Ultra-Short-Term Photovoltaic Power Prediction Based on BiLSTM with Wavelet Decomposition and Dual Attention Mechanism
by Mingyang Liu, Xiaohuan Wang and Zhiwen Zhong
Electronics 2025, 14(2), 306; https://doi.org/10.3390/electronics14020306 - 14 Jan 2025
Viewed by 184
Abstract
Photovoltaic power generation relies on sunlight conditions, and traditional prediction models find it difficult to capture the deep features of power data, resulting in low prediction accuracy. In addition, there are problems such as outliers and missing values in the data collected on [...] Read more.
Photovoltaic power generation relies on sunlight conditions, and traditional prediction models find it difficult to capture the deep features of power data, resulting in low prediction accuracy. In addition, there are problems such as outliers and missing values in the data collected on site. This article proposes an ultra-short-term photovoltaic power generation prediction model based on wavelet decomposition, a dual attention mechanism, and a bidirectional long short-term memory network (W-DA-BiLSTM), aiming to address the limitations of existing deep learning models in processing nonlinear data and automatic feature extraction and optimize for the common problems of outliers and missing values in on-site data collection. This model uses the quartile range method for outlier detection and multiple interpolation methods for missing value completion. In the prediction section, wavelet decomposition is used to effectively handle the volatility and nonlinear characteristics of photovoltaic power generation data, while the bidirectional long short-term memory network (LSTM) structure and dual attention mechanism enhance the model’s comprehensive learning ability for time series data. The experimental results show that compared with the SOTA method, the model proposed in this paper has higher accuracy and efficiency in predicting photovoltaic power generation and can effectively address common random fluctuations and nonlinear problems in photovoltaic power generation. Full article
Show Figures

Figure 1

Figure 1
<p>Irradiance quantile–quantile chart.</p>
Full article ">Figure 2
<p>Wavelet transform process.</p>
Full article ">Figure 3
<p>Principle diagram of attention mechanism.</p>
Full article ">Figure 4
<p>BiLSTM schematic diagram.</p>
Full article ">Figure 5
<p>Structure diagram of W-DA-BiLSTM photovoltaic power generation prediction model.</p>
Full article ">Figure 6
<p>Comparison of forecast results in different seasons.</p>
Full article ">Figure 7
<p>Scatter plot of photovoltaic power generation and irradiance before processing.</p>
Full article ">Figure 8
<p>Scatter plot of photovoltaic power generation and irradiance after processing.</p>
Full article ">Figure 9
<p>Wavelet decomposition of raw signal and trend signal.</p>
Full article ">Figure 10
<p>The wavelet decomposes the detailed signal.</p>
Full article ">Figure 11
<p>Weight distribution of feature attention mechanism.</p>
Full article ">Figure 12
<p>Weight distribution of time attention mechanism.</p>
Full article ">
25 pages, 3173 KiB  
Article
A Novel CNN-Based Framework for Alzheimer’s Disease Detection Using EEG Spectrogram Representations
by Konstantinos Stefanou, Katerina D. Tzimourta, Christos Bellos, Georgios Stergios, Konstantinos Markoglou, Emmanouil Gionanidis, Markos G. Tsipouras, Nikolaos Giannakeas, Alexandros T. Tzallas and Andreas Miltiadous
J. Pers. Med. 2025, 15(1), 27; https://doi.org/10.3390/jpm15010027 - 14 Jan 2025
Viewed by 176
Abstract
Background: Alzheimer’s disease (AD) is a progressive neurodegenerative disorder that poses critical challenges in global healthcare due to its increasing prevalence and severity. Diagnosing AD and other dementias, such as frontotemporal dementia (FTD), is slow and resource-intensive, underscoring the need for automated approaches. [...] Read more.
Background: Alzheimer’s disease (AD) is a progressive neurodegenerative disorder that poses critical challenges in global healthcare due to its increasing prevalence and severity. Diagnosing AD and other dementias, such as frontotemporal dementia (FTD), is slow and resource-intensive, underscoring the need for automated approaches. Methods: To address this gap, this study proposes a novel deep learning methodology for EEG classification of AD, FTD, and control (CN) signals. The approach incorporates advanced preprocessing techniques and CNN classification of FFT-based spectrograms and is evaluated using the leave-N-subjects-out validation, ensuring robust cross-subject generalizability. Results: The results indicate that the proposed methodology outperforms state-of-the-art machine learning and EEG-specific neural network models, achieving an accuracy of 79.45% for AD/CN classification and 80.69% for AD+FTD/CN classification. Conclusions: These results highlight the potential of EEG-based deep learning models for early dementia screening, enabling more efficient, scalable, and accessible diagnostic tools. Full article
(This article belongs to the Special Issue Personalized Treatment of Neurological Diseases)
Show Figures

Graphical abstract

Graphical abstract
Full article ">Figure 1
<p>A detailed illustration of the methodology and system architecture.</p>
Full article ">Figure 2
<p>Scalp heatmaps of the average PSD of the 3 different classes, across the 5 frequency bands of interest.</p>
Full article ">Figure 3
<p>The charts depict the FFT transformation of one epoch for a participant from each of the three classes. The data are presented in a 3D representation to enhance visualization and provide a clearer view of the frequency distribution across channels (the original data were displayed as flat RGB images).</p>
Full article ">Figure 4
<p>The train accuracy (<b>left</b>) and train loss (<b>right</b>) in the AD-CN problem, with respect to the number of epochs. An optimal performance is obtained at 100 epochs.</p>
Full article ">Figure 5
<p>Averaged FFT heatmaps for each group in the 0.5–45 Hz frequency range. Each row represents a channel and each column represents a frequency value. The units are uV<sup>2</sup>/Hz.</p>
Full article ">Figure 6
<p>Averaged FFT heatmaps for each group in the 0.5–25 Hz frequency range. Each row represents a channel and each column represents a frequency value. The units are uV<sup>2</sup>/Hz.</p>
Full article ">
21 pages, 957 KiB  
Article
Human Trajectory Imputation Model: A Hybrid Deep Learning Approach for Pedestrian Trajectory Imputation
by Deb Kanti Barua, Mithun Halder, Shayanta Shopnil and Md. Motaharul Islam
Appl. Sci. 2025, 15(2), 745; https://doi.org/10.3390/app15020745 - 14 Jan 2025
Viewed by 169
Abstract
Pedestrian trajectories are crucial for self-driving cars to plan their paths effectively. The sensors implanted in these self-driving vehicles, despite being state-of-the-art ones, often face inaccuracies in the perception of surrounding environments due to technical challenges in adverse weather conditions, interference from other [...] Read more.
Pedestrian trajectories are crucial for self-driving cars to plan their paths effectively. The sensors implanted in these self-driving vehicles, despite being state-of-the-art ones, often face inaccuracies in the perception of surrounding environments due to technical challenges in adverse weather conditions, interference from other vehicles’ sensors and electronic devices, and signal reception failure, leading to incompleteness in the trajectory data. But for real-time decision making for autonomous driving, trajectory imputation is no less crucial. Previous attempts to address this issue, such as statistical inference and machine learning approaches, have shown promise. Yet, the landscape of deep learning is rapidly evolving, with new and more robust models emerging. In this research, we have proposed an encoder–decoder architecture, the Human Trajectory Imputation Model, coined HTIM, to tackle these challenges. This architecture aims to fill in the missing parts of pedestrian trajectories. The model is evaluated using the Intersection drone the inD dataset, containing trajectory data at suitable altitudes, preserving naturalistic pedestrian behavior with varied dataset sizes. To assess the effectiveness of our model, we utilize L1, MSE, and quantile and ADE loss. Our experiments demonstrate that HTIM outperforms the majority of the state-of-the-art methods in this field, thus indicating its superior performance in imputing pedestrian trajectories. Full article
(This article belongs to the Special Issue Advances in Autonomous Driving and Smart Transportation)
Show Figures

Figure 1

Figure 1
<p>Humans are unpredictable, a critical security concern for autonomous vehicles. Image courtesy of <a href="http://coursera.org" target="_blank">coursera.org</a> accessed on 25 June 2024.</p>
Full article ">Figure 2
<p>The figure illustrates trajectory segmentation for the imputation process. The blue segments indicate the portions of the trajectory provided as input to the model, while the pink segments represent the randomly masked portions of the trajectory, which the model is tasked with imputing. This segmentation highlights the distinction between observed and unobserved data in the trajectory.</p>
Full article ">Figure 3
<p>Encoder-Decoder Architecture with Teacher Forcing: The red-labeled points represent the true outputs used during training.</p>
Full article ">Figure 4
<p>Comparison of generated and ground truth trajectories with velocity values.</p>
Full article ">Figure 5
<p>L1 loss variation across training epochs: tracking the model’s performance.</p>
Full article ">Figure 6
<p>L1 loss variation across validation epochs: tracking the model’s performance.</p>
Full article ">Figure 7
<p>MSE loss variation across training epochs: tracking the model’s performance.</p>
Full article ">Figure 8
<p>MSE loss variation across validation epochs: tracking the model’s performance.</p>
Full article ">Figure 9
<p>ADE test loss: evaluation metric for model performance on test dataset.</p>
Full article ">Figure 10
<p>Quantile loss variation across training epochs: tracking the model’s performance.</p>
Full article ">Figure 11
<p>Quantile loss variation across validation epochs: tracking the model’s performance.</p>
Full article ">
Back to TopTop