[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (4,928)

Search Parameters:
Keywords = diversity learning

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
27 pages, 3027 KiB  
Article
DeepONet-Inspired Architecture for Efficient Financial Time Series Prediction
by Zeeshan Ahmad, Shudi Bao and Meng Chen
Mathematics 2024, 12(24), 3950; https://doi.org/10.3390/math12243950 (registering DOI) - 16 Dec 2024
Abstract
Financial time series prediction is a fundamental problem in investment and risk management. Deep learning models, such as multilayer perceptrons, Convolutional Neural Networks (CNNs), and Long Short-Term Memory (LSTM), have been widely used in modeling time series data by incorporating historical information. Among [...] Read more.
Financial time series prediction is a fundamental problem in investment and risk management. Deep learning models, such as multilayer perceptrons, Convolutional Neural Networks (CNNs), and Long Short-Term Memory (LSTM), have been widely used in modeling time series data by incorporating historical information. Among them, LSTM has shown excellent performance in capturing long-term temporal dependencies in time-series data, owing to its enhanced internal memory mechanism. In spite of the success of these models, it is observed that in the presence of sharp changing points, these models fail to perform. To address this problem, we propose, in this article, an innovative financial time series prediction method inspired by the Deep Operator Network (DeepONet) architecture, which uses a combination of transformer architecture and a one-dimensional CNN network for processing feature-based information, followed by an LSTM based network for processing temporal information. It is therefore named the CNN–LSTM–Transformer (CLT) model. It not only incorporates external information to identify latent patterns within the financial data but also excels in capturing their temporal dynamics. The CLT model adapts to evolving market conditions by leveraging diverse deep-learning techniques. This dynamic adaptation of the CLT model plays a pivotal role in navigating abrupt changes in the financial markets. Furthermore, the CLT model improves the long-term prediction accuracy and stability compared with state-of-the-art existing deep learning models and also mitigates adverse effects of market volatility. The experimental results show the feasibility and superiority of the proposed CLT model in terms of prediction accuracy and robustness as compared to existing prediction models. Moreover, we posit that the innovation encapsulated in the proposed DeepONet-inspired CLT model also holds promise for applications beyond the confines of finance, such as remote sensing, data mining, natural language processing, and so on. Full article
(This article belongs to the Section Financial Mathematics)
Show Figures

Figure 1

Figure 1
<p>Analysis flowchart of the stock price prediction.</p>
Full article ">Figure 2
<p>Schematic diagram of the DeepONet.</p>
Full article ">Figure 3
<p>Proposed DeepONet-inspired architecture of the CLT model.</p>
Full article ">Figure 4
<p>Data sampling method.</p>
Full article ">Figure 5
<p>Loss curve on training and validation sets.</p>
Full article ">Figure 6
<p>Comparison of the prediction accuracy with varying number of nodes in a hidden layer.</p>
Full article ">Figure 7
<p>Comparison of the prediction accuracy with varying number of hidden layers.</p>
Full article ">Figure 8
<p>Comparison of the results produced by various models for the AEX.</p>
Full article ">Figure 9
<p>Comparison of the results produced by various models for the ATX.</p>
Full article ">Figure 10
<p>Model performance variation.</p>
Full article ">Figure A1
<p>Comparison of the results produced by various models for the following: (<b>a</b>) FCHI; (<b>b</b>) FTSE; (<b>c</b>) HSI; (<b>d</b>) JKSE; (<b>e</b>) KLSE; (<b>f</b>) OEX.</p>
Full article ">
11 pages, 2439 KiB  
Article
AISMPred: A Machine Learning Approach for Predicting Anti-Inflammatory Small Molecules
by Subathra Selvam, Priya Dharshini Balaji, Honglae Sohn and Thirumurthy Madhavan
Pharmaceuticals 2024, 17(12), 1693; https://doi.org/10.3390/ph17121693 (registering DOI) - 15 Dec 2024
Abstract
Background/Objectives: Inflammation serves as a vital response to diverse harmful stimuli like infections, toxins, or tissue injuries, aiding in the elimination of pathogens and tissue repair. However, persistent inflammation can lead to chronic diseases. Peptide therapeutics have gained attention for their specificity in [...] Read more.
Background/Objectives: Inflammation serves as a vital response to diverse harmful stimuli like infections, toxins, or tissue injuries, aiding in the elimination of pathogens and tissue repair. However, persistent inflammation can lead to chronic diseases. Peptide therapeutics have gained attention for their specificity in targeting cells, yet their development remains costly and time-consuming. Therefore, small molecules, with their stability, low immunogenicity, and oral bioavailability, have become a focal point for predicting anti-inflammatory small molecules (AISMs). Methods: In this study, we introduce a computational method called AISMPred, designed to classify AISMs and non-AISMs. To develop this approach, we constructed a dataset comprising 1750 AISMs and non-AISMs, each annotated with IC50 values sourced from the PubChem BioAssay database. We computed two distinct types of molecular descriptors using PaDEL and Mordred tools. Subsequently, these descriptors were concatenated to form a hybrid feature set. The SVC-L1 regularization method was implemented for the optimum feature selection to develop robust Machine learning (ML) models. Five different conventional ML classifiers were employed, such as RF, ET, KNN, LR, and Ensemble methods. Results: A total of 15 ML models were developed using 2D, FP, and Hybrid feature sets, with the ET model with hybrid features achieving the highest accuracy of 92% and an AUC of 0.97 on the independent test dataset. Conclusions: This study provides an effective method for screening AISMs, potentially impacting drug discovery and design. Full article
Show Figures

Figure 1

Figure 1
<p>The chemical space of the compounds in the training set compared with that in the test set. (<b>a</b>) 2D descriptors, (<b>b</b>) fingerprints, (<b>c</b>) hybrid (2D + FP).</p>
Full article ">Figure 2
<p>(<b>a</b>) Comparison of receiver operating characteristic curves of the four models on external data using Hybrid dataset. The curve closer to the upper left corner showed better overall discrimination ability. (<b>b</b>) Comparison of precision-recall curves of the four models on external data. The curve closer to the upper right corner also showed the ability to combine precision with sensitivity. (AP: average precision, AUC: area under the receiver operating characteristic curve, ROC: receiver operating characteristic).</p>
Full article ">Figure 3
<p>Feature importance plot for the selected ML-based ExtraTree model using hybrid feature set.</p>
Full article ">Figure 4
<p>Computational framework of AISMPred. It includes data collection, feature selection, model construction, and performance comparison.</p>
Full article ">
14 pages, 2007 KiB  
Article
Hardware-Assisted Low-Latency NPU Virtualization Method for Multi-Sensor AI Systems
by Jong-Hwan Jean and Dong-Sun Kim
Sensors 2024, 24(24), 8012; https://doi.org/10.3390/s24248012 (registering DOI) - 15 Dec 2024
Abstract
Recently, AI systems such as autonomous driving and smart homes have become integral to daily life. Intelligent multi-sensors, once limited to single data types, now process complex text and image data, demanding faster and more accurate processing. While integrating NPUs and sensors has [...] Read more.
Recently, AI systems such as autonomous driving and smart homes have become integral to daily life. Intelligent multi-sensors, once limited to single data types, now process complex text and image data, demanding faster and more accurate processing. While integrating NPUs and sensors has improved processing speed and accuracy, challenges like low resource utilization and long memory latency remain. This study proposes a method to reduce processing time and improve resource utilization by virtualizing NPUs to simultaneously handle multiple deep-learning models, leveraging a hardware scheduler and data prefetching techniques. Experiments with 30,000 SA resources showed that the hardware scheduler reduced memory cycles by over 10% across all models, with reductions of 30% for NCF and 70% for DLRM. The hardware scheduler effectively minimized memory latency and idle NPU resources in resource-constrained environments with frequent context switching. This approach is particularly valuable for real-time applications like autonomous driving, enabling smooth transitions between tasks such as object detection and route planning. It also enhances multitasking in smart homes by reducing latency when managing diverse data streams. The proposed system is well suited for resource-constrained environments that demand efficient multitasking and low-latency processing. Full article
(This article belongs to the Section Intelligent Sensors)
Show Figures

Figure 1

Figure 1
<p>Examples of real-life applications of multi-sensor AI in various fields.</p>
Full article ">Figure 2
<p>NPU virtualization operation flow. The symbol ‘#’ represents the number of cores.</p>
Full article ">Figure 3
<p>TPU v4 architecture.</p>
Full article ">Figure 4
<p>Hardware-assisted NPU virtualization system.</p>
Full article ">Figure 5
<p>Comparison before and after the hardware scheduler when the burst size was changed.</p>
Full article ">Figure 6
<p>Comparison before and after hardware scheduler application when the number of available SA changed.</p>
Full article ">Figure 7
<p>Difference in memory cycles with and without a hardware scheduler.</p>
Full article ">
28 pages, 51329 KiB  
Article
MHTAPred-SS: A Highly Targeted Autoencoder-Driven Deep Multi-Task Learning Framework for Accurate Protein Secondary Structure Prediction
by Runqiu Feng, Xun Wang, Zhijun Xia, Tongyu Han, Hanyu Wang and Wenqian Yu
Int. J. Mol. Sci. 2024, 25(24), 13444; https://doi.org/10.3390/ijms252413444 (registering DOI) - 15 Dec 2024
Abstract
Accurate protein secondary structure prediction (PSSP) plays a crucial role in biopharmaceutics and disease diagnosis. Current prediction methods are mainly based on multiple sequence alignment (MSA) encoding and collaborative operations of diverse networks. However, existing encoding approaches lead to poor feature space utilization, [...] Read more.
Accurate protein secondary structure prediction (PSSP) plays a crucial role in biopharmaceutics and disease diagnosis. Current prediction methods are mainly based on multiple sequence alignment (MSA) encoding and collaborative operations of diverse networks. However, existing encoding approaches lead to poor feature space utilization, and encoding quality decreases with fewer homologous proteins. Moreover, the performance of simple stacked networks is greatly limited by feature extraction capabilities and learning strategies. To this end, we propose MHTAPred-SS, a novel PSSP framework based on the fusion of six features, including the embedding feature derived from a pre-trained protein language model. First, we propose a highly targeted autoencoder (HTA) as the driver to encode sequences in a homologous protein-independent manner. Second, under the guidance of biological knowledge, we design a protein secondary structure prediction model based on the multi-task learning strategy (PSSP-MTL). Experimental results on six independent test sets show that MHTAPred-SS achieves state-of-the-art performance, with values of 88.14%, 84.89%, 78.74% and 77.15% for Q3, SOV3, Q8 and SOV8 metrics on the TEST2016 dataset, respectively. Additionally, we demonstrate that MHTAPred-SS has significant advantages in single-category and boundary secondary structure prediction, and can finely capture the distribution of secondary structure segments, thereby contributing to subsequent tasks. Full article
(This article belongs to the Special Issue Structural and Functional Analysis of Amino Acids and Proteins)
Show Figures

Figure 1

Figure 1
<p>The secondary structure segment length pattern. The horizontal axis represents the interval to which the secondary structure segment length belongs, and the vertical axis represents the number of secondary structure segments.</p>
Full article ">Figure 2
<p>The correlation between secondary structure and RSA. Each horizontal axis represents the eight secondary structure categories, and each vertical axis represents the proportion of RSA ≤ 0.15 and RSA &gt; 0.15. The gray dashed lines indicate that there are no amino acid residues belonging to this secondary structure category in the dataset.</p>
Full article ">Figure 3
<p>Examples of boundary amino acid residues.</p>
Full article ">Figure 4
<p>The prediction performance of models under BiLSTM with different numbers of layers and hidden units.</p>
Full article ">Figure 5
<p>The prediction performance of models when using different residual convolution scales. The expanded chart is the eight-state prediction experimental results based on models of different scales on Validation set2.</p>
Full article ">Figure 6
<p>The prediction performance of models under different weight assignments.</p>
Full article ">Figure 7
<p>The prediction performance under different learning strategies. “Difference” represents the difference between the prediction performance of the MTL model and the STL model.</p>
Full article ">Figure 8
<p>Normalized confusion matrices of three-state (<b>a</b>) and eight-state (<b>b</b>) prediction on TEST2016.</p>
Full article ">Figure 9
<p>Visualization of secondary structure prediction results from different methods. The dashed box shows the biological experimental results.</p>
Full article ">Figure 10
<p>Visualization of secondary structure prediction results for difficult proteins. The dashed box shows the biological experimental results.</p>
Full article ">Figure 11
<p>Prediction results for orphan proteins. The green segments indicate the correct predictions, and the red segments indicate the wrong predictions.</p>
Full article ">Figure 12
<p>The workflow of MHTAPred-SS. Our proposed MHTAPred-SS consists of four key components: (<b>1</b>) data acquisition: two sets of datasets are obtained for model training, validation and testing; (<b>2</b>) multi-feature fusion: six different features are obtained using five methods; (<b>3</b>) PSSP-MTL model: the PSSP-MTL model consists of three modules and (<b>4</b>) output predictor: the output predictor simultaneously outputs predicted results of secondary structure and RSA.</p>
Full article ">Figure 13
<p>The operating mechanism of HTA. HTA is divided into two parts: encoder (DY-CNN module) and decoder (BiLSTM module), which reconstruct the primary structure information of each protein.</p>
Full article ">Figure 14
<p>The weighted fusion principle of features output by expert networks. “Weight Assignment” aims to assign weights to the output features of each expert network, and “Weighted addition” means aggregating the output features of each expert network according to the assigned weights to obtain the features extracted for each task.</p>
Full article ">Figure 15
<p>The architecture of the TCN unit.</p>
Full article ">
15 pages, 739 KiB  
Article
Weighted Linear Discriminant Analysis: An Effective Feature Extraction Method for Multi-Class Imbalanced Datasets
by Yuhan Liu and Shuangle Guo
Symmetry 2024, 16(12), 1656; https://doi.org/10.3390/sym16121656 (registering DOI) - 14 Dec 2024
Viewed by 251
Abstract
In high-dimensional machine learning tasks, supervised feature extraction is essential for improving model performance, with Linear Discriminant Analysis (LDA) being a common approach. However, LDA tends to deliver suboptimal performance when dealing with class imbalance. To address this issue, we propose a novel [...] Read more.
In high-dimensional machine learning tasks, supervised feature extraction is essential for improving model performance, with Linear Discriminant Analysis (LDA) being a common approach. However, LDA tends to deliver suboptimal performance when dealing with class imbalance. To address this issue, we propose a novel feature extraction model, Weighted Linear Discriminant Analysis (WLDA), which integrates cost-sensitive techniques into the traditional LDA framework. By assigning weights inversely proportional to class sample sizes, WLDA achieves effective feature extraction under imbalanced sample conditions. We introduce an efficient solution algorithm for the proposed model and provide a thorough complexity analysis. Experimental results demonstrate the superior performance of WLDA in handling imbalanced datasets, confirming its potential as a robust tool for high-dimensional data scenarios. Overall, WLDA not only improves feature extraction for imbalanced datasets but also enhances classification accuracy across diverse applications. Full article
(This article belongs to the Special Issue Symmetry/Asymmetry in Neural Networks and Applications)
Show Figures

Figure 1

Figure 1
<p>Distribution of values pie chart for (<b>a</b>) Zoo, (<b>b</b>) Credit Card, (<b>c</b>) Dermatology, (<b>d</b>) Gas Sensor Array Drift, (<b>e</b>) Urban_Land_Cover and (<b>f</b>) USPS.</p>
Full article ">Figure 2
<p>Comparative analysis of feature extraction algorithms on imbalanced datasets with SVM classifiers.</p>
Full article ">Figure 3
<p>Comparative analysis of feature extraction algorithms on imbalanced datasets with logistic regression classifiers.</p>
Full article ">Figure 4
<p>Comparative analysis of feature extraction algorithms on imbalanced datasets with decision tree classifiers.</p>
Full article ">
34 pages, 9176 KiB  
Article
A Multi-Strategy Improved Honey Badger Algorithm for Engineering Design Problems
by Tao Han, Tingting Li, Quanzeng Liu, Yourui Huang and Hongping Song
Algorithms 2024, 17(12), 573; https://doi.org/10.3390/a17120573 - 13 Dec 2024
Viewed by 194
Abstract
A multi-strategy improved honey badger algorithm (MIHBA) is proposed to address the problem that the honey badger algorithm may fall into local optimum and premature convergence when dealing with complex optimization problems. By introducing Halton sequences to initialize the population, the diversity of [...] Read more.
A multi-strategy improved honey badger algorithm (MIHBA) is proposed to address the problem that the honey badger algorithm may fall into local optimum and premature convergence when dealing with complex optimization problems. By introducing Halton sequences to initialize the population, the diversity of the population is enhanced, and premature convergence is effectively avoided. The dynamic density factor of water waves is added to improve the search efficiency of the algorithm in the solution space. Lens opposition learning based on the principle of lens imaging is also introduced to enhance the ability of the algorithm to get rid of local optimums. MIHBA achieves the best ranking in 23 test functions and 4 engineering design problems. The improvement of this paper improves the convergence speed and accuracy of the algorithm, enhances the adaptability and solving ability of the algorithm to complex functions, and provides new ideas for solving complex engineering design problems. Full article
15 pages, 2934 KiB  
Article
Evaluation of Pothole Detection Performance Using Deep Learning Models Under Low-Light Conditions
by Yuliia Zanevych, Vasyl Yovbak, Oleh Basystiuk, Nataliya Shakhovska, Solomiia Fedushko and Sotirios Argyroudis
Sustainability 2024, 16(24), 10964; https://doi.org/10.3390/su162410964 - 13 Dec 2024
Viewed by 335
Abstract
In our interconnected society, prioritizing the resilience and sustainability of road infrastructure has never been more critical, especially in light of growing environmental and climatic challenges. By harnessing data from various sources, we can proactively enhance our ability to detect road damage. This [...] Read more.
In our interconnected society, prioritizing the resilience and sustainability of road infrastructure has never been more critical, especially in light of growing environmental and climatic challenges. By harnessing data from various sources, we can proactively enhance our ability to detect road damage. This approach will enable us to make well-informed decisions for timely maintenance and implement effective mitigation strategies, ultimately leading to safer and more durable road systems. This paper presents a new method for detecting road potholes during low-light conditions, particularly at night when influenced by street and traffic lighting. We examined and assessed various advanced machine learning and computer vision models, placing a strong emphasis on deep learning algorithms such as YOLO, as well as the combination of Grad-CAM++ with feature pyramid networks for feature extraction. Our approach utilized innovative data augmentation techniques, which enhanced the diversity and robustness of the training dataset, ultimately leading to significant improvements in model performance. The study results reveal that the proposed YOLOv11+FPN+Grad-CAM model achieved a mean average precision (mAP) score of 0.72 for the 50–95 IoU thresholds, outperforming other tested models, including YOLOv8 Medium with a score of 0.611. The proposed model also demonstrated notable improvements in key metrics, with mAP50 and mAP75 values of 0.88 and 0.791, reflecting enhancements of 1.5% and 5.7%, respectively, compared to YOLOv11. These results highlight the model’s superior performance in detecting potholes under low-light conditions. By leveraging a specialized dataset for nighttime scenarios, the approach offers significant advancements in hazard detection, paving the way for more effective and timely driver alerts and ultimately contributing to improved road safety. This paper makes several key contributions, including implementing advanced data augmentation methods and a thorough comparative analysis of various YOLO-based models. Future plans involve developing a real-time driver warning application, introducing enhanced evaluation metrics, and demonstrating the model’s adaptability in diverse environmental conditions, such as snow and rain. The contributions significantly advance the field of road maintenance and safety by offering a robust and scalable solution for pothole detection, particularly in developing countries. Full article
Show Figures

Figure 1

Figure 1
<p>Graph of analysis of literature sources related to YOLO models generated by the Litmaps tool.</p>
Full article ">Figure 2
<p>The research process.</p>
Full article ">Figure 3
<p>Comparison of results for pothole recognition models: (<b>a</b>) Yolo-based models; (<b>b</b>) Yolo + FPN-based models.</p>
Full article ">Figure 4
<p>This is an illustration of the framework. (<b>a</b>) FPN backbone, (<b>b</b>) bottom-up path augmentation, (<b>c</b>) adaptive feature pooling, (<b>d</b>) box branch, (<b>e</b>) fully-connected fusion.</p>
Full article ">Figure 5
<p>Feature pyramid network architecture.</p>
Full article ">Figure 6
<p>The research pipeline.</p>
Full article ">Figure 7
<p>Comparison of model performance: (blue) Based on mean average precision IoU 50; (orange) mean average precision IoU 75; (green) mean average precision IoU 50–95.</p>
Full article ">Figure 8
<p>Comparison of night model performance.</p>
Full article ">
19 pages, 414 KiB  
Article
Fair and Transparent Student Admission Prediction Using Machine Learning Models
by George Raftopoulos, Gregory Davrazos and Sotiris Kotsiantis
Algorithms 2024, 17(12), 572; https://doi.org/10.3390/a17120572 - 13 Dec 2024
Viewed by 190
Abstract
Student admission prediction is a crucial aspect of academic planning, offering insights into enrollment trends, resource allocation, and institutional growth. However, traditional methods often lack the ability to address fairness and transparency, leading to potential biases and inequities in the decision-making process. This [...] Read more.
Student admission prediction is a crucial aspect of academic planning, offering insights into enrollment trends, resource allocation, and institutional growth. However, traditional methods often lack the ability to address fairness and transparency, leading to potential biases and inequities in the decision-making process. This paper explores the development and evaluation of machine learning models designed to predict student admissions while prioritizing fairness and interpretability. We employ a diverse set of algorithms, including Logistic Regression, Decision Trees, and ensemble methods, to forecast admission outcomes based on academic, demographic, and extracurricular features. Experimental results on real-world datasets highlight the effectiveness of the proposed models in achieving competitive predictive performance while adhering to fairness metrics such as demographic parity and equalized odds. Our findings demonstrate that machine learning can not only enhance the accuracy of admission predictions but also support equitable access to education by promoting transparency and accountability in automated systems. Full article
(This article belongs to the Special Issue Artificial Intelligence Algorithms and Generative AI in Education)
Show Figures

Graphical abstract

Graphical abstract
Full article ">
16 pages, 2713 KiB  
Review
Machine Learning-Driven Innovations in Microfluidics
by Jinseok Park, Yang Woo Kim and Hee-Jae Jeon
Biosensors 2024, 14(12), 613; https://doi.org/10.3390/bios14120613 - 13 Dec 2024
Viewed by 443
Abstract
Microfluidic devices have revolutionized biosensing by enabling precise manipulation of minute fluid volumes across diverse applications. This review investigates the incorporation of machine learning (ML) into the design, fabrication, and application of microfluidic biosensors, emphasizing how ML algorithms enhance performance by improving design [...] Read more.
Microfluidic devices have revolutionized biosensing by enabling precise manipulation of minute fluid volumes across diverse applications. This review investigates the incorporation of machine learning (ML) into the design, fabrication, and application of microfluidic biosensors, emphasizing how ML algorithms enhance performance by improving design accuracy, operational efficiency, and the management of complex diagnostic datasets. Integrating microfluidics with ML has fostered intelligent systems capable of automating experimental workflows, enabling real-time data analysis, and supporting informed decision-making. Recent advances in health diagnostics, environmental monitoring, and synthetic biology driven by ML are critically examined. This review highlights the transformative potential of ML-enhanced microfluidic systems, offering insights into the future trajectory of this rapidly evolving field. Full article
Show Figures

Figure 1

Figure 1
<p>A timeline depicting the integration of microfluidics and machine learning biosensing applications. (<b>A</b>) The timeline of technological advancements in microfluidics and machine learning, highlighting key milestones and their integration. (<b>B</b>) Specific breakthroughs and applications in intelligent microfluidics.</p>
Full article ">Figure 2
<p>Leveraging machine learning for microfluidic device design optimization. (<b>A</b>) A schematic of the convolutional neural network (CNN) used in this work. Adapted with permission [<a href="#B65-biosensors-14-00613" class="html-bibr">65</a>], copyright 2017, <span class="html-italic">Scientific Reports</span>. (<b>B</b>) The workflow of the developed design automation tool for flow-focusing droplet generators, called DAFD. Adapted with permission [<a href="#B64-biosensors-14-00613" class="html-bibr">64</a>], copyright 2021, <span class="html-italic">Nature Communications</span>. (<b>C</b>) A workflow chart and a description of methods used in the building of the machine learning-guided design of an experiment based on the decision tree method. Adapted with permission [<a href="#B66-biosensors-14-00613" class="html-bibr">66</a>], copyright 2023, <span class="html-italic">Chemical Engineering Research and Design</span>.</p>
Full article ">Figure 3
<p>Machine learning-enhanced control and design of droplets in microfluidic systems. (<b>A</b>) Prediction of droplet flow dynamics using deep neural networks. Adapted with permission [<a href="#B65-biosensors-14-00613" class="html-bibr">65</a>], copyright 2019, <span class="html-italic">Scientific Reports</span>. (<b>B</b>) Real-time droplet classification for multijet monitoring. Adapted with permission [<a href="#B64-biosensors-14-00613" class="html-bibr">64</a>], copyright 2022, <span class="html-italic">ACS Applied Materials &amp; Interfaces</span>. (<b>C</b>) Design automation of single- and double-emulsion droplets through machine learning. Adapted with permission [<a href="#B66-biosensors-14-00613" class="html-bibr">66</a>], copyright 2024, <span class="html-italic">Nature Communications</span>.</p>
Full article ">Figure 4
<p>Machine learning-enabled real-time monitoring and analysis within microfluidic platforms. (<b>A</b>) Microscopy-based label-free imaging flow cytometry with real-time image processing. Adapted with permission [<a href="#B90-biosensors-14-00613" class="html-bibr">90</a>], copyright 2017, <span class="html-italic">Scientific Reports</span>. (<b>B</b>) Viral aerosol detection employing microfluidics and machine learning for rapid classification. Adapted with permission [<a href="#B91-biosensors-14-00613" class="html-bibr">91</a>], copyright 2024, <span class="html-italic">ACS Nano</span>. (<b>C</b>) Skin-interfaced microfluidic patch with machine learning-based image processing for sweat biomarker analysis. Adapted with permission [<a href="#B92-biosensors-14-00613" class="html-bibr">92</a>], copyright 2022, <span class="html-italic">Advanced Materials Technologies</span>.</p>
Full article ">
19 pages, 3440 KiB  
Article
A Hybrid Strategy-Improved SSA-CNN-LSTM Model for Metro Passenger Flow Forecasting
by Jing Liu, Qingling He, Zhikun Yue and Yulong Pei
Mathematics 2024, 12(24), 3929; https://doi.org/10.3390/math12243929 - 13 Dec 2024
Viewed by 313
Abstract
To address the issues of slow convergence and large errors in existing metaheuristic algorithms when optimizing neural network-based subway passenger flow prediction, we propose the following improvements. First, we replace the random initialization method of the population in the SSA with Circle mapping [...] Read more.
To address the issues of slow convergence and large errors in existing metaheuristic algorithms when optimizing neural network-based subway passenger flow prediction, we propose the following improvements. First, we replace the random initialization method of the population in the SSA with Circle mapping to enhance its diversity and quality. Second, we introduce a hybrid mechanism combining dimensional small-hole imaging backward learning and Cauchy mutation, which improves the diversity of the individual sparrow selection of optimal positions and helps overcome the algorithm’s tendency to become trapped in local optima and premature convergence. Finally, we enhance the individual sparrow position update process by integrating a cosine strategy with an inertia weight adjustment, which improves the algorithm’s global search ability, effectively balancing global search and local exploitation, and reducing the risk of local optima and insufficient convergence precision. Based on the analysis of the correlation between different types of subway station passenger flows and weather factors, the ISSA is used to optimize the hyperparameters of the CNN-LSTM model to construct a subway passenger flow prediction model based on ISSA-CNN-LSTM. Simulation experiments were conducted using card swipe data from Harbin Metro Line 1. The results show that the ISSA provides a more accurate optimization with the average values and standard deviations of the 12 benchmark test function simulations being closer to the optimal values. The ISSA-CNN-LSTM model outperforms the SSA-CNN-LSTM, PSO-ELMAN, GA-BP, CNN-LSTM, and LSTM models in terms of error evaluation metrics such as MAE, RMSE, and MAPE, with improvements ranging from 189.8% to 374.6%, 190.9% to 389.5%, and 3.3% to 6.7%, respectively. Moreover, the ISSA-CNN-LSTM model exhibits the smallest variation in prediction errors across different types of subway stations. The ISSA demonstrates superior parameter optimization accuracy and convergence speed compared to the SSA. The ISSA-CNN-LSTM model is suitable for the precise prediction of passenger flow at different types of subway stations, providing theoretical and data support for subway station passenger density and trend forecasting, passenger organization and management, risk emergency response, and the improvement of service quality and operational safety. Full article
Show Figures

Figure 1

Figure 1
<p>Correlation between influencing factors of metro station passenger flow.</p>
Full article ">Figure 2
<p>CNN-LSTM network structure.</p>
Full article ">Figure 3
<p>ISSA-CNN-LSTM prediction process.</p>
Full article ">Figure 4
<p>Statistical results of simulations for benchmark functions. (The boxplot in the text represents the results of 30 simulation runs for each benchmark test function, corresponding to the optimization solutions of various heuristic algorithms. It is primarily used to compare the accuracy and efficiency of the optimization solutions provided by each algorithm. The figure labels below each subplot correspond to the respective benchmark test functions (<b>f1</b>–<b>f12</b>)).</p>
Full article ">Figure 5
<p>Convergence curves of benchmark functions. (The convergence curve represents the variation in the mean fitness values of the results from 30 simulation runs for each benchmark test function, corresponding to the optimization solutions of various heuristic algorithms, plotted against the number of algorithm iterations. The figure labels below each subplot correspond to the respective benchmark test functions (<b>f1–f12</b>)).</p>
Full article ">Figure 6
<p>The correlation between actual values and predicted values is depicted.</p>
Full article ">Figure 7
<p>Comparative error analysis.</p>
Full article ">Figure 8
<p>Error measurement plots for different types of stations and models.</p>
Full article ">Figure 8 Cont.
<p>Error measurement plots for different types of stations and models.</p>
Full article ">
16 pages, 230 KiB  
Article
Exploring Autonomy in the AI Wilderness: Learner Challenges and Choices
by Antonie Alm
Educ. Sci. 2024, 14(12), 1369; https://doi.org/10.3390/educsci14121369 - 13 Dec 2024
Viewed by 300
Abstract
The emergence of Generative Artificial Intelligence (GenAI) raises critical questions about learner autonomy and agency. This exploratory case study examines how four university-level German language learners with diverse backgrounds developed autonomy in their learning process through engagement with AI tools. The study was [...] Read more.
The emergence of Generative Artificial Intelligence (GenAI) raises critical questions about learner autonomy and agency. This exploratory case study examines how four university-level German language learners with diverse backgrounds developed autonomy in their learning process through engagement with AI tools. The study was conducted in early 2023 when most learners were first discovering ChatGPT’s potential for language learning. Data were collected through reflective journals, digital portfolios, and interviews during a semester-long course that scaffolded self-directed learning with AI integration. The findings reveal emerging patterns of shared agency between learners and AI tools. Learners developed distinct strategies for AI integration based on their language learning backgrounds, with heritage speakers focusing on accuracy improvement while classroom learners emphasized communication practice. Cross-case analyses identified key dimensions of autonomy development: a critical evaluation of AI output, evolving learner–AI relationships, maintaining and developing a second language (L2) voice, and the strategic integration of AI tools while preserving learner agency. These patterns suggest that autonomy in AI-mediated environments manifests through learners’ capacity to engage productively with AI while maintaining critical awareness and personal agency in their learning process. Full article
13 pages, 3607 KiB  
Article
Multi-Modal Machine Learning Approach for COVID-19 Detection Using Biomarkers and X-Ray Imaging
by Kagan Tur
Diagnostics 2024, 14(24), 2800; https://doi.org/10.3390/diagnostics14242800 - 13 Dec 2024
Viewed by 351
Abstract
Background: Accurate and rapid detection of COVID-19 remains critical for clinical management, especially in resource-limited settings. Current diagnostic methods face challenges in terms of speed and reliability, creating a need for complementary AI-based models that integrate diverse data sources. Objectives: This [...] Read more.
Background: Accurate and rapid detection of COVID-19 remains critical for clinical management, especially in resource-limited settings. Current diagnostic methods face challenges in terms of speed and reliability, creating a need for complementary AI-based models that integrate diverse data sources. Objectives: This study aimed to develop and evaluate a multi-modal machine learning model that combines clinical biomarkers and chest X-ray images to enhance diagnostic accuracy and provide interpretable insights. Methods: We used a dataset of 250 patients (180 COVID-19 positive and 70 negative cases) collected from clinical settings. Biomarkers such as CRP, ferritin, NLR, and albumin were included alongside chest X-ray images. Random Forest and Gradient Boosting models were used for biomarkers, and ResNet and VGG CNN architectures were applied to imaging data. A late-fusion strategy integrated predictions from these modalities. Stratified k-fold cross-validation ensured robust evaluation while preventing data leakage. Model performance was assessed using AUC-ROC, F1-score, Specificity, Negative Predictive Value (NPV), and Matthews Correlation Coefficient (MCC), with confidence intervals calculated via bootstrap resampling. Results: The Gradient Boosting + VGG fusion model achieved the highest performance, with an AUC-ROC of 0.94, F1-score of 0.93, Specificity of 93%, NPV of 96%, and MCC of 0.91. SHAP and LIME interpretability analyses identified CRP, ferritin, and specific lung regions as key contributors to predictions. Discussion: The proposed multi-modal approach significantly enhances diagnostic accuracy compared to single-modality models. Its interpretability aligns with clinical understanding, supporting its potential for real-world application. Full article
(This article belongs to the Special Issue Medical Data Processing and Analysis—2nd Edition)
Show Figures

Figure 1

Figure 1
<p>AUC-ROC curves for models.</p>
Full article ">Figure 2
<p>Confusion matrix for Gradient Boosting + VGG Model.</p>
Full article ">Figure 3
<p>Distribution of key biomarkers by COVID-19 outcome (<b>top left</b>: CRP, <b>top right</b>: Ferritin, <b>bottom left</b>: NLR, <b>bottom right</b>: Albumin).</p>
Full article ">Figure 4
<p>SHAP summary for key biomarkers.</p>
Full article ">Figure 5
<p>LIME explanation of key X-ray regions.</p>
Full article ">
14 pages, 11336 KiB  
Article
Prediction of Full-Load Electrical Power Output of Combined Cycle Power Plant Using a Super Learner Ensemble
by Yujeong Song, Jisu Park, Myoung-Seok Suh and Chansoo Kim
Appl. Sci. 2024, 14(24), 11638; https://doi.org/10.3390/app142411638 - 12 Dec 2024
Viewed by 455
Abstract
Combined Cycle Power Plants (CCPPs) generate electrical power through gas turbines and use the exhaust heat from those turbines to power steam turbines, resulting in 50% more power output compared to traditional simple cycle power plants. Predicting the full-load electrical power output ( [...] Read more.
Combined Cycle Power Plants (CCPPs) generate electrical power through gas turbines and use the exhaust heat from those turbines to power steam turbines, resulting in 50% more power output compared to traditional simple cycle power plants. Predicting the full-load electrical power output (PE) of a CCPP is crucial for efficient operation and sustainable development. Previous studies have used machine learning models, such as the Bagging and Boosting models to predict PE. In this study, we propose employing Super Learner (SL), an ensemble machine learning algorithm, to enhance the accuracy and robustness of predictions. SL utilizes cross-validation to estimate the performance of diverse machine learning models and generates an optimal weighted average based on their respective predictions. It may provide information on the relative contributions of each base learner to the overall prediction skill. For constructing the SL, we consider six individual and ensemble machine learning models as base learners and assess their performances compared to the SL. The dataset used in this study was collected over six years from an operational CCPP. It contains one output variable and four input variables: ambient temperature, atmospheric pressure, relative humidity, and vacuum. The results show that the Boosting algorithms significantly influence the performance of the SL in comparison to the other base learners. The SL outperforms the six individual and ensemble machine learning models used as base learners. It indicates that the SL improves the generalization performance of predictions by combining the predictions of various machine learning models. Full article
Show Figures

Figure 1

Figure 1
<p>The correlation matrix showing the relationships between input and output variables.</p>
Full article ">Figure 2
<p>Box plot for input variables.</p>
Full article ">Figure 3
<p>The coefficient-stacked bar displays the optimal weight combinations calculated from the SL for each of the 10-fold cross-validations. Coefficients below 0.02 are not displayed in the text.</p>
Full article ">Figure 4
<p>Kernel density estimation of the predicted values for each model and its corresponding actual values.</p>
Full article ">Figure 5
<p>Comparisons of the prediction performances for each 10 validation folds and averaged performance in terms of RMSE and MAE.</p>
Full article ">Figure 6
<p>Comparison of the prediction performances of SL and base learners.</p>
Full article ">Figure 7
<p>The average performances of 2-fold cross-validation on 5 files for each model.</p>
Full article ">Figure 8
<p>The plot of the real value and predicted value with SL for 200 points.</p>
Full article ">
20 pages, 4577 KiB  
Article
FedLSTM: A Federated Learning Framework for Sensor Fault Detection in Wireless Sensor Networks
by Rehan Khan, Umer Saeed and Insoo Koo
Electronics 2024, 13(24), 4907; https://doi.org/10.3390/electronics13244907 - 12 Dec 2024
Viewed by 321
Abstract
The rapid growth of Internet of Things (IoT) devices has significantly increased reliance on sensor-generated data, which are essential to a wide range of systems and services. Wireless sensor networks (WSNs), crucial to this ecosystem, are often deployed in diverse and challenging environments, [...] Read more.
The rapid growth of Internet of Things (IoT) devices has significantly increased reliance on sensor-generated data, which are essential to a wide range of systems and services. Wireless sensor networks (WSNs), crucial to this ecosystem, are often deployed in diverse and challenging environments, making them susceptible to faults such as software bugs, communication breakdowns, and hardware malfunctions. These issues can compromise data accuracy, stability, and reliability, ultimately jeopardizing system security. While advanced sensor fault detection methods in WSNs leverage a machine learning approach to achieve high accuracy, they typically rely on centralized learning, and face scalability and privacy challenges, especially when transferring large volumes of data. In our experimental setup, we employ a decentralized approach using federated learning with long short-term memory (FedLSTM) for sensor fault detection in WSNs, thereby preserving client privacy. This study utilizes temperature data enhanced with synthetic sensor data to simulate various common sensor faults: bias, drift, spike, erratic, stuck, and data-loss. We evaluate the performance of FedLSTM against the centralized approach based on accuracy, precision, sensitivity, and F1-score. Additionally, we analyze the impacts of varying the client participation rates and the number of local training epochs. In federated learning environments, comparative analysis with established models like the one-dimensional convolutional neural network and multilayer perceptron demonstrate the promising results of FedLSTM in maintaining client privacy while reducing communication overheads and the server load. Full article
(This article belongs to the Special Issue Advances in Cyber-Security and Machine Learning)
Show Figures

Figure 1

Figure 1
<p>IoT applications connected to a central base station in various smart environments.</p>
Full article ">Figure 2
<p>Communication setups in IoT-based wireless sensor networks: (<b>a</b>) single-hop and (<b>b</b>) multi-hop scenario.</p>
Full article ">Figure 3
<p>Representative plots of the various faults monitored by the proposed FedLSTM for distributed sensor fault detection and employed across multiple clients (sensors).</p>
Full article ">Figure 4
<p>Framework of the proposed FedLSTM for distributed sensor fault detection. Each client trains its local model and collaborates with a central server to build the global model.</p>
Full article ">Figure 5
<p>The proposed workflow for sensor fault diagnosis in WSNs depicts the following stages: data acquisition from multiple sensors, data preprocessing, including generating synthetic data for various common sensor faults, data partitioning for distributed storage, local model training, where clients train models locally using their respective datasets, and global model aggregation using FL for fault detection and classification.</p>
Full article ">Figure 6
<p>Comparison of FedLSTM and the centralized model for sensor fault detection in WSNs.</p>
Full article ">Figure 7
<p>Performance of the FedLSTM model in terms of (<b>a</b>) accuracy and (<b>b</b>) loss over 50 communication rounds, in single-hop, multi-hop, and combined (single-hop and multi-hop) scenarios.</p>
Full article ">Figure 8
<p>Confusion matrices for (<b>a</b>) FedLSTM and (<b>b</b>) the centralized model from multiclass sensor-fault detection.</p>
Full article ">Figure 9
<p>Impact from varying the number of local epochs <math display="inline"><semantics> <mi>ϵ</mi> </semantics></math> on (<b>a</b>) accuracy and (<b>b</b>) loss convergence of the FedLSTM model across 50 communication rounds.</p>
Full article ">Figure 10
<p>(<b>a</b>) Accuracy and (<b>b</b>) loss from FedLSTM, the 1D-CNN, and MLP.</p>
Full article ">
19 pages, 1847 KiB  
Article
Gender Differences in the Use of Generative Artificial Intelligence Chatbots in Higher Education: Characteristics and Consequences
by Anja Møgelvang, Camilla Bjelland, Simone Grassini and Kristine Ludvigsen
Educ. Sci. 2024, 14(12), 1363; https://doi.org/10.3390/educsci14121363 - 12 Dec 2024
Viewed by 580
Abstract
Student gender differences in technology acceptance and use have persisted for years, giving rise to equity concerns in higher education (HE). To explore if such differences extend to generative artificial intelligence (genAI) chatbot use, we surveyed a large Norwegian HE student sample ( [...] Read more.
Student gender differences in technology acceptance and use have persisted for years, giving rise to equity concerns in higher education (HE). To explore if such differences extend to generative artificial intelligence (genAI) chatbot use, we surveyed a large Norwegian HE student sample (n = 2692) using a fully mixed concurrent equal status design. Our findings show that men exhibit more frequent engagement with genAI chatbots across a broader spectrum of applications. Further, men demonstrate a heightened interest in genAI chatbots as tools and in their relevance to future career prospects. Women primarily utilize genAI chatbots in text-related tasks and express greater concerns regarding critical and independent thinking. Women also exhibit a stronger need to learn how to determine when it is wise to use and how to trust genAI chatbots. Consequences are discussed for the individual, society, and HE institutions in terms of social reproduction, diversity competence, and equitable teaching practices. Full article
Show Figures

Figure 1

Figure 1
<p>The 5-step analysis of the qualitative data.</p>
Full article ">Figure 2
<p>GenAI chatbot use frequency split by gender (<span class="html-italic">n</span> = 2380).</p>
Full article ">Figure 3
<p>Word trees for “think”.</p>
Full article ">Figure 4
<p>Word trees for “workplace”.</p>
Full article ">
Back to TopTop