[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 

Topic Editors

Artificial Intelligence College, Future Technology College, Nanjing University of Information Science and Technology, Nanjing, China
Prof. Dr. Junzo Watada
Graduate School of Information, Production and Systems, Waseda University, Tokyo, Japan
VSB, Technical University of Ostrava, 708 00 Ostrava, Czech Republic
Dr. Pei Hu
School of Computer and Software, Nanyang Institute of Technology, Nanyang, China

Applications of Machine Learning in Large-Scale Optimization and High-Dimensional Learning

Abstract submission deadline
28 February 2025
Manuscript submission deadline
30 April 2025
Viewed by
6348

Topic Information

Dear Colleagues,

Machine Learning (ML) has found a wide range of applications in large-scale optimization and high-dimensional learning problems. Below are some notable areas where ML is applied:

  1. Large-Scale Optimization: ML techniques are used to tackle complex optimization problems in various domains. These include optimizing supply chain logistics, scheduling tasks in industrial processes, and parameter tuning in machine learning algorithms;
  2. Multi-Objective Optimization: ML is well suited for multi-objective optimization problems, where there are multiple conflicting objectives to be optimized simultaneously. These scenarios are common in fields such as engineering, finance, and resource allocation;
  3. High-Dimensional Data Analysis: ML aids in discovering patterns in high-dimensional data. These patterns can be used in various applications, such as customer segmentation in marketing, anomaly detection, and image segmentation.

Prof. Dr. Jeng-Shyang Pan
Prof. Dr. Junzo Watada
Prof. Dr. Vaclav Snasel
Dr. Pei Hu
Topic Editors

Keywords

  • machine learning
  • large-scale optimization
  • multi-objective optimization
  • high-dimensional data analysis
  • artificial intelligence

Participating Journals

Journal Name Impact Factor CiteScore Launched Year First Decision (median) APC
AI
ai
3.1 7.2 2020 17.6 Days CHF 1600 Submit
Buildings
buildings
3.1 3.4 2011 17.2 Days CHF 2600 Submit
Computers
computers
2.6 5.4 2012 17.2 Days CHF 1800 Submit
Drones
drones
4.4 5.6 2017 21.7 Days CHF 2600 Submit
Entropy
entropy
2.1 4.9 1999 22.4 Days CHF 2600 Submit
Symmetry
symmetry
2.2 5.4 2009 16.8 Days CHF 2400 Submit

Preprints.org is a multidiscipline platform providing preprint service that is dedicated to sharing your research from the start and empowering your research journey.

MDPI Topics is cooperating with Preprints.org and has built a direct connection between MDPI journals and Preprints.org. Authors are encouraged to enjoy the benefits by posting a preprint at Preprints.org prior to publication:

  1. Immediately share your ideas ahead of publication and establish your research priority;
  2. Protect your idea from being stolen with this time-stamped preprint article;
  3. Enhance the exposure and impact of your research;
  4. Receive feedback from your peers in advance;
  5. Have it indexed in Web of Science (Preprint Citation Index), Google Scholar, Crossref, SHARE, PrePubMed, Scilit and Europe PMC.

Published Papers (4 papers)

Order results
Result details
Journals
Select all
Export citation of selected articles as:
23 pages, 6011 KiB  
Article
Optimizing Steering Angle Prediction in Self-Driving Vehicles Using Evolutionary Convolutional Neural Networks
by Bashar Khawaldeh, Antonio M. Mora and Hossam Faris
AI 2024, 5(4), 2147-2169; https://doi.org/10.3390/ai5040105 - 30 Oct 2024
Viewed by 1009
Abstract
The global community is awaiting the advent of a self-driving vehicle that is safe, reliable, and capable of navigating a diverse range of road conditions and terrains. This requires a lot of research, study, and optimization. Thus, this work focused on implementing, training, [...] Read more.
The global community is awaiting the advent of a self-driving vehicle that is safe, reliable, and capable of navigating a diverse range of road conditions and terrains. This requires a lot of research, study, and optimization. Thus, this work focused on implementing, training, and optimizing a convolutional neural network (CNN) model, aiming to predict the steering angle during driving (one of the main issues). The considered dataset comprises images collected inside a car-driving simulator and further processed for augmentation and removal of unimportant details. In addition, an innovative data-balancing process was previously performed. A CNN model was trained with the dataset, conducting a comparison between several different standard optimizers. Moreover, evolutionary optimization was applied to optimize the model’s weights as well as the optimizers themselves. Several experiments were performed considering different approaches of genetic algorithms (GAs) along with other optimizers from the state of the art. The obtained results demonstrate that the GA is an effective optimization tool for this problem. Full article
Show Figures

Figure 1

Figure 1
<p>Udacity simulator.</p>
Full article ">Figure 2
<p>Udacity’s Lake and Jungle Tracks.</p>
Full article ">Figure 3
<p>Example of images of three camera positions.</p>
Full article ">Figure 4
<p>Distribution of dataset before and after data undersampling.</p>
Full article ">Figure 5
<p>Original training and validation dataset distribution.</p>
Full article ">Figure 6
<p>Novel training and validation dataset distribution.</p>
Full article ">Figure 7
<p>Augmentation examples.</p>
Full article ">Figure 8
<p>Original image and image preprocessed by cropping, RGB to YUV, Gaussian blur filter, and resizing.</p>
Full article ">Figure 9
<p>Convolutional neural network (CNN) for steering angle prediction based on NVIDIA architecture.</p>
Full article ">Figure 10
<p>Genetic algorithm procedure.</p>
Full article ">Figure 11
<p>Example of crossover process.</p>
Full article ">Figure 12
<p>Loss (MSE) boxplots for the 30 trained models for each optimizer, based on the training dataset (left boxplot) and validation dataset (right boxplot).</p>
Full article ">Figure 13
<p>Average loss (MSE) of 10 runs for each of the different configurations for the GA optimizer over the CNN model.</p>
Full article ">
24 pages, 6064 KiB  
Article
Efficient First-Order Algorithms for Large-Scale, Non-Smooth Maximum Entropy Models with Application to Wildfire Science
by Gabriel Provencher Langlois, Jatan Buch and Jérôme Darbon
Entropy 2024, 26(8), 691; https://doi.org/10.3390/e26080691 - 15 Aug 2024
Viewed by 759
Abstract
Maximum entropy (MaxEnt) models are a class of statistical models that use the maximum entropy principle to estimate probability distributions from data. Due to the size of modern data sets, MaxEnt models need efficient optimization algorithms to scale well for big data applications. [...] Read more.
Maximum entropy (MaxEnt) models are a class of statistical models that use the maximum entropy principle to estimate probability distributions from data. Due to the size of modern data sets, MaxEnt models need efficient optimization algorithms to scale well for big data applications. State-of-the-art algorithms for MaxEnt models, however, were not originally designed to handle big data sets; these algorithms either rely on technical devices that may yield unreliable numerical results, scale poorly, or require smoothness assumptions that many practical MaxEnt models lack. In this paper, we present novel optimization algorithms that overcome the shortcomings of state-of-the-art algorithms for training large-scale, non-smooth MaxEnt models. Our proposed first-order algorithms leverage the Kullback–Leibler divergence to train large-scale and non-smooth MaxEnt models efficiently. For MaxEnt models with discrete probability distribution of n elements built from samples, each containing m features, the stepsize parameter estimation and iterations in our algorithms scale on the order of O(mn) operations and can be trivially parallelized. Moreover, the strong 1 convexity of the Kullback–Leibler divergence allows for larger stepsize parameters, thereby speeding up the convergence rate of our algorithms. To illustrate the efficiency of our novel algorithms, we consider the problem of estimating probabilities of fire occurrences as a function of ecological features in the Western US MTBS-Interagency wildfire data set. Our numerical results show that our algorithms outperform the state of the art by one order of magnitude and yield results that agree with physical models of wildfire occurrence and previous statistical analyses of wildfire drivers. Full article
Show Figures

Figure 1

Figure 1
<p>Wildfire activity in the western United States from 1984 to 2020. (<b>Left</b>) Fire locations of all fires (black dots) in the Western US MTBS-Interagency (WUMI) data set; also shown are three ecological divisions characterized by their primary vegetation type—forests (green), deserts (yellow), and plains (gray). (<b>Right</b>) Prior distribution indicating mean fire probability across all calendar months.</p>
Full article ">Figure 2
<p>Spatial probability plot for different hyperparameter values with elastic net penalty parameter <math display="inline"><semantics> <mrow> <mi>α</mi> <mo>=</mo> <mo>{</mo> <mn>0.95</mn> <mo>,</mo> <mn>0.40</mn> <mo>,</mo> <mn>0.15</mn> <mo>,</mo> <mn>0.05</mn> <mo>}</mo> </mrow> </semantics></math>.</p>
Full article ">Figure 3
<p>Same as <a href="#entropy-26-00691-f002" class="html-fig">Figure 2</a> but for (<b>left</b>) the non-overlapping group lasso with <math display="inline"><semantics> <mrow> <mi>α</mi> <mo>=</mo> <mn>1</mn> </mrow> </semantics></math> and (<b>right</b>) the <math display="inline"><semantics> <msub> <mi>l</mi> <mo>∞</mo> </msub> </semantics></math> MaxEnt models, respectively.</p>
Full article ">Figure 4
<p>Number of non-zero coefficients along the regularization path plots for elastic net penalty parameter <math display="inline"><semantics> <mrow> <mi>α</mi> <mo>=</mo> <mo>{</mo> <mn>0.95</mn> <mo>,</mo> <mn>0.40</mn> <mo>,</mo> <mn>0.15</mn> <mo>,</mo> <mn>0.05</mn> <mo>}</mo> </mrow> </semantics></math>. The dashed vertical lines highlight the <math display="inline"><semantics> <mrow> <mi>t</mi> <mo>/</mo> <msub> <mi>t</mi> <mi>max</mi> </msub> </mrow> </semantics></math> value at which the first feature of the group indicated by inset text is selected.</p>
Full article ">
26 pages, 8560 KiB  
Article
Power Transmission Lines Foreign Object Intrusion Detection Method for Drone Aerial Images Based on Improved YOLOv8 Network
by Hongbin Sun, Qiuchen Shen, Hongchang Ke, Zhenyu Duan and Xi Tang
Drones 2024, 8(8), 346; https://doi.org/10.3390/drones8080346 - 25 Jul 2024
Cited by 2 | Viewed by 1269
Abstract
With the continuous growth of electricity demand, the safety and stability of transmission lines have become increasingly important. To ensure the reliability of power supply, it is essential to promptly detect and address foreign object intrusions on transmission lines, such as tree branches, [...] Read more.
With the continuous growth of electricity demand, the safety and stability of transmission lines have become increasingly important. To ensure the reliability of power supply, it is essential to promptly detect and address foreign object intrusions on transmission lines, such as tree branches, kites, and balloons. Addressing the issues where foreign objects can cause power outages and severe safety accidents, as well as the inefficiency, time consumption, and labor-intensiveness of traditional manual inspection methods, especially in large-scale power transmission lines, we propose an enhanced YOLOv8-based model for detecting foreign objects. This model incorporates the Swin Transformer, AFPN (Asymptotic Feature Pyramid Network), and a novel loss function, Focal SIoU, to improve both the accuracy and real-time detection of hazards. The integration of the Swin Transformer into the YOLOv8 backbone network significantly improves feature extraction capabilities. The AFPN enhances the multi-scale feature fusion process, effectively integrating information from different levels and improving detection accuracy, especially for small and occluded objects. The introduction of the Focal SIoU loss function optimizes the model’s training process, enhancing its ability to handle hard-to-classify samples and uncertain predictions. This method achieves efficient automatic detection of foreign objects by comprehensively utilizing multi-level feature information and optimized label matching strategies. The dataset used in this study consists of images of foreign objects on power transmission lines provided by a power supply company in Jilin, China. These images were captured by drones, offering a comprehensive view of the transmission lines and enabling the collection of detailed data on various foreign objects. Experimental results show that the improved YOLOv8 network has high accuracy and recall rates in detecting foreign objects such as balloons, kites, and bird nests, while also possessing good real-time processing capabilities. Full article
Show Figures

Figure 1

Figure 1
<p>Samples of foreign objects from different perspectives.</p>
Full article ">Figure 2
<p>The YOLOv8 network architecture.</p>
Full article ">Figure 3
<p>The training flowchart of the YOLOv8 network is shown in.</p>
Full article ">Figure 4
<p>(<b>a</b>) The structure of Swin Transformer (Swin-T). (<b>b</b>) Two successive Swin Transformer blocks.</p>
Full article ">Figure 5
<p>Feature Pyramid Network. In this figure, feature maps are indicated by blue outlines and thicker outlines denote semantically stronger features.</p>
Full article ">Figure 6
<p>The architecture of the proposed Asymptotic Feature Pyramid Network (AFPN). In the initial stage, two low-level features are fused first, followed by the integration of high-level features, and finally, the top-level features are fused.</p>
Full article ">Figure 7
<p>Performance of different models under foreign object intrusion.</p>
Full article ">Figure 8
<p>FPS values of different models.</p>
Full article ">Figure 9
<p>Params values of different models.</p>
Full article ">Figure 10
<p>mAP values of different models.</p>
Full article ">Figure 11
<p>AP performance of different models on the dataset.</p>
Full article ">Figure 12
<p>mAP performance of different models on the dataset.</p>
Full article ">Figure 13
<p>Evaluation metrics of different models.</p>
Full article ">Figure 14
<p>Loss values for different loss functions.</p>
Full article ">Figure 15
<p>mAP for different loss functions.</p>
Full article ">Figure 16
<p>Precision for different loss functions.</p>
Full article ">Figure 17
<p>Compared to other advanced algorithms, our model has a high accuracy in detecting foreign objects.</p>
Full article ">Figure 17 Cont.
<p>Compared to other advanced algorithms, our model has a high accuracy in detecting foreign objects.</p>
Full article ">
22 pages, 1589 KiB  
Article
Knowledge Distillation in Image Classification: The Impact of Datasets
by Ange Gabriel Belinga, Cédric Stéphane Tekouabou Koumetio, Mohamed El Haziti and Mohammed El Hassouni
Computers 2024, 13(8), 184; https://doi.org/10.3390/computers13080184 - 24 Jul 2024
Viewed by 1641
Abstract
As the demand for efficient and lightweight models in image classification grows, knowledge distillation has emerged as a promising technique to transfer expertise from complex teacher models to simpler student models. However, the efficacy of knowledge distillation is intricately linked to the choice [...] Read more.
As the demand for efficient and lightweight models in image classification grows, knowledge distillation has emerged as a promising technique to transfer expertise from complex teacher models to simpler student models. However, the efficacy of knowledge distillation is intricately linked to the choice of datasets used during training. Datasets are pivotal in shaping a model’s learning process, influencing its ability to generalize and discriminate between diverse patterns. While considerable research has independently explored knowledge distillation and image classification, a comprehensive understanding of how different datasets impact knowledge distillation remains a critical gap. This study systematically investigates the impact of diverse datasets on knowledge distillation in image classification. By varying dataset characteristics such as size, domain specificity, and inherent biases, we aim to unravel the nuanced relationship between datasets and the efficacy of knowledge transfer. Our experiments employ a range of datasets to comprehensively explore their impact on the performance gains achieved through knowledge distillation. This study contributes valuable guidance for researchers and practitioners seeking to optimize image classification models through kno-featured applications. By elucidating the intricate interplay between dataset characteristics and knowledge distillation outcomes, our findings empower the community to make informed decisions when selecting datasets, ultimately advancing the field toward more robust and efficient model development. Full article
Show Figures

Figure 1

Figure 1
<p>Flowchart of the proposed approach to highlight the impact of the dataset on knowledge distillation in DNN.</p>
Full article ">Figure 2
<p>Residual blocks of ResNet architecture [<a href="#B46-computers-13-00184" class="html-bibr">46</a>].</p>
Full article ">Figure 3
<p>Response-based knowledge distillation [<a href="#B48-computers-13-00184" class="html-bibr">48</a>].</p>
Full article ">Figure 4
<p>Intermediate knowledge distillation [<a href="#B48-computers-13-00184" class="html-bibr">48</a>].</p>
Full article ">Figure 5
<p>Variation in the validation accuracy by epochs for (<b>a</b>) the teacher model (ResNet50) and (<b>b</b>) the student model (ResNet18).</p>
Full article ">Figure 6
<p>Test accuracy for the teacher and instance student model from scratch.</p>
Full article ">Figure 7
<p>Test accuracy for the teacher and instance student models RKD.</p>
Full article ">Figure 8
<p>Difference between the student model from scratch and the student IKD accuracy.</p>
Full article ">Figure 9
<p>Difference between the teacher and instance student model distilled from RKD (<b>a</b>) and instance student model distilled from IKD (<b>b</b>).</p>
Full article ">Figure 10
<p>Student performance gain after distillation.</p>
Full article ">Figure 11
<p>All instances of student performance compared to teacher performance. bar visualisation (<b>a</b>) and curve visualisation (<b>b</b>).</p>
Full article ">Figure 12
<p>Impact of dataset.</p>
Full article ">Figure A1
<p>Accuracy and loss over epochs during the training phase of the student model in the MNIST dataset. (<b>a</b>) Training student from scratch, (<b>b</b>) RKD student training, and (<b>c</b>) IKD student training.</p>
Full article ">Figure A2
<p>Student metrics after the training phase of the student model in MNIST dataset. (<b>a</b>) Training student from scratch, (<b>b</b>) RKD student training, and (<b>c</b>) IKD student training.</p>
Full article ">Figure A3
<p>Accuracy and loss over epochs during the training phase of student model in USPS dataset. (<b>a</b>) Training student from scratch, (<b>b</b>) RKD student training, and (<b>c</b>) IKD student training.</p>
Full article ">Figure A4
<p>Student metrics after the training phase of the student model in the USPS dataset. (<b>a</b>) Training student from scratch, (<b>b</b>) RKD student training, and (<b>c</b>) IKD student training.</p>
Full article ">Figure A5
<p>Accuracy and loss over epochs during the training phase of the student model in FashionMNIST dataset. (<b>a</b>) Training student from scratch, (<b>b</b>) RKD student training, and (<b>c</b>) IKD student training.</p>
Full article ">Figure A6
<p>Student metrics after the training phase of the student model in the FashionMNIST dataset. (<b>a</b>) Training student from scratch, (<b>b</b>) RKD student training, and (<b>c</b>) IKD student training.</p>
Full article ">Figure A7
<p>Accuracy and loss over epochs during the training phase of the student model in the CIFAR10 dataset. (<b>a</b>) Training student from scratch, (<b>b</b>) RKD student training, and (<b>c</b>) IKD student training.</p>
Full article ">Figure A8
<p>Student metrics after the training phase of the student model in the CIFAR10 dataset. (<b>a</b>) Training student from scratch, (<b>b</b>) RKD student training, and (<b>c</b>) IKD student training.</p>
Full article ">Figure A9
<p>Accuracy and loss over epochs during the training phase of the student model in the CIFAR100 dataset. (<b>a</b>) Training student from scratch, (<b>b</b>) RKD student training, and (<b>c</b>) IKD student training.</p>
Full article ">Figure A10
<p>Student metrics after the training phase of the student model in the CIFAR100 dataset. (<b>a</b>) Training student from scratch, (<b>b</b>) RKD student training, and (<b>c</b>) IKD student training.</p>
Full article ">
Back to TopTop