[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
Next Issue
Volume 15, September
Previous Issue
Volume 15, July
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 

Information, Volume 15, Issue 8 (August 2024) – 87 articles

Cover Story (view full-size image): The paper presents Darly, a QoS-, interference- and heterogeneity-aware Deep Reinforcement Learning-based Scheduler for serverless video analytics deployments on top of distributed Edge nodes. The proposed framework incorporates a DRL agent that exploits performance counters to identify the levels of interference and the degree of heterogeneity in the underlying Edge infrastructure. It combines this information along with user-defined QoS requirements to improve resource allocations by deciding the placement, migration, or horizontal scaling of serverless functions. View this paper
  • Issues are regarded as officially published after their release is announced to the table of contents alert mailing list.
  • You may sign up for e-mail alerts to receive table of contents of newly released issues.
  • PDF is the official format for papers published in both, html and pdf forms. To view the papers in pdf format, click on the "PDF Full-text" link, and use the free Adobe Reader to open them.
Order results
Result details
Section
Select all
Export citation of selected articles as:
62 pages, 1897 KiB  
Review
Construction of Knowledge Graphs: Current State and Challenges
by Marvin Hofer, Daniel Obraczka, Alieh Saeedi, Hanna Köpcke and Erhard Rahm
Information 2024, 15(8), 509; https://doi.org/10.3390/info15080509 - 22 Aug 2024
Cited by 4 | Viewed by 5344
Abstract
With Knowledge Graphs (KGs) at the center of numerous applications such as recommender systems and question-answering, the need for generalized pipelines to construct and continuously update such KGs is increasing. While the individual steps that are necessary to create KGs from unstructured sources [...] Read more.
With Knowledge Graphs (KGs) at the center of numerous applications such as recommender systems and question-answering, the need for generalized pipelines to construct and continuously update such KGs is increasing. While the individual steps that are necessary to create KGs from unstructured sources (e.g., text) and structured data sources (e.g., databases) are mostly well researched for their one-shot execution, their adoption for incremental KG updates and the interplay of the individual steps have hardly been investigated in a systematic manner so far. In this work, we first discuss the main graph models for KGs and introduce the major requirements for future KG construction pipelines. Next, we provide an overview of the necessary steps to build high-quality KGs, including cross-cutting topics such as metadata management, ontology development, and quality assurance. We then evaluate the state of the art of KG construction with respect to the introduced requirements for specific popular KGs, as well as some recent tools and strategies for KG construction. Finally, we identify areas in need of further research and improvement. Full article
(This article belongs to the Special Issue Knowledge Graph Technology and its Applications II)
Show Figures

Figure 1

Figure 1
<p>Simplified Knowledge Graph (KG) example demonstrating integrated information from five domains, showcasing ten entities of eight types connected by twelve relationships (two distinct is-a relations). Dashed lines indicate semantic structures (ontology or graph schema), such as entity types. Inferences can be made based on the relationships and typing, revealing additional information such as the broader birthplace of Aphex Twin being Ireland and Xtal belonging to the Techno genre (Not all possible inferences are shown for clarity).</p>
Full article ">Figure 2
<p>Incremental knowledge graph construction pipeline.</p>
Full article ">Figure 3
<p>Ontology and entity merging strategies.</p>
Full article ">Figure 4
<p>Knowledge extraction steps for an example sentence linking entities and relations to the DBpedia KG. Recognized named entities are highlighted in green.</p>
Full article ">Figure 5
<p>Incremental entity-resolution workflow.</p>
Full article ">
25 pages, 5483 KiB  
Article
Automated Negotiation Agents for Modeling Single-Peaked Bidders: An Experimental Comparison
by Fatemeh Hassanvand, Faria Nassiri-Mofakham and Katsuhide Fujita
Information 2024, 15(8), 508; https://doi.org/10.3390/info15080508 - 22 Aug 2024
Viewed by 685
Abstract
During automated negotiations, intelligent software agents act based on the preferences of their proprietors, interdicting direct preference exposure. The agent can be armed with a component of an opponent’s modeling features to reduce the uncertainty in the negotiation, but how negotiating agents with [...] Read more.
During automated negotiations, intelligent software agents act based on the preferences of their proprietors, interdicting direct preference exposure. The agent can be armed with a component of an opponent’s modeling features to reduce the uncertainty in the negotiation, but how negotiating agents with a single-peaked preference direct our attention has not been considered. Here, we first investigate the proper representation of single-peaked preferences and implementation of single-peaked agents within bidder agents using different instances of general single-peaked functions. We evaluate the modeling of single-peaked preferences and bidders in automated negotiating agents. Through experiments, we reveal that most of the opponent models can model our benchmark single-peaked agents with similar efficiencies. However, the accuracies differ among the models and in different rival batches. The perceptron-based P1 model obtained the highest accuracy, and the frequency-based model Randomdance outperformed the other competitors in most other performance measures. Full article
(This article belongs to the Special Issue Intelligent Agent and Multi-Agent System)
Show Figures

Figure 1

Figure 1
<p>Representations of shortfall (curve fitted to the green squares) and excess avoidance (curve fitted to the purple triangles).</p>
Full article ">Figure 2
<p>Skew-normal utility functions applied in this paper, plotted over the Pie domain.</p>
Full article ">Figure 3
<p>Gaussian utility functions used in this paper, plotted over the Pie domain.</p>
Full article ">Figure 4
<p>General triangular utility functions used in this paper, plotted over the Pie domain.</p>
Full article ">Figure 5
<p>Configuration for evaluating the opponent model of a rival: (<b>a</b>) <span class="html-italic">B</span> benchmark agent for accuracy assessment; (<b>b</b>) <span class="html-italic">BA</span> benchmark agent for performance assessment [<a href="#B13-information-15-00508" class="html-bibr">13</a>].</p>
Full article ">Figure 6
<p>Configuration of the rivals: (<b>a</b>) <span class="html-italic">BO</span> agent for the accuracy assessment and (<b>b</b>) <span class="html-italic">BOA</span> agent for the performance assessment [<a href="#B13-information-15-00508" class="html-bibr">13</a>].</p>
Full article ">Figure 7
<p>Initial interface of GENIUS, showcasing the components panel on the left and the status panel on the right [<a href="#B27-information-15-00508" class="html-bibr">27</a>].</p>
Full article ">Figure 8
<p>Interactions between the modeling agents (Side 1) and benchmark agents (Side 2) during (<b>a</b>) the performance assessment and (<b>b</b>) the accuracy assessment [<a href="#B13-information-15-00508" class="html-bibr">13</a>].</p>
Full article ">Figure 9
<p>Accuracies of the opponent models in front of (<b>a</b>) all benchmark agents; (<b>b</b>) Gaussian agents; (<b>c</b>) skew-normal agents; (<b>d</b>) general triangular agents.</p>
Full article ">Figure 10
<p>Accuracies of the opponent models for different domain sizes in front of (<b>a</b>) all benchmark agents; (<b>b</b>) Gaussian agents; (<b>c</b>) skew-normal agents; (<b>d</b>) general triangular agents.</p>
Full article ">Figure 11
<p>Performance results of the opponent models confronting all benchmark agents (black), Gaussian agents (cyan), skew-normal agents (green), and general triangular agents (red): (<b>a</b>) Individual utility; (<b>b</b>) Social welfare; (<b>c</b>) Nash distance; (<b>d</b>) Pareto distance; (<b>e</b>) Time to agreement.</p>
Full article ">
33 pages, 6672 KiB  
Review
Advancements in Deep Learning Techniques for Time Series Forecasting in Maritime Applications: A Comprehensive Review
by Meng Wang, Xinyan Guo, Yanling She, Yang Zhou, Maohan Liang and Zhong Shuo Chen
Information 2024, 15(8), 507; https://doi.org/10.3390/info15080507 - 21 Aug 2024
Viewed by 2597
Abstract
The maritime industry is integral to global trade and heavily depends on precise forecasting to maintain efficiency, safety, and economic sustainability. Adopting deep learning for predictive analysis has markedly improved operational accuracy, cost efficiency, and decision-making. This technology facilitates advanced time series analysis, [...] Read more.
The maritime industry is integral to global trade and heavily depends on precise forecasting to maintain efficiency, safety, and economic sustainability. Adopting deep learning for predictive analysis has markedly improved operational accuracy, cost efficiency, and decision-making. This technology facilitates advanced time series analysis, vital for optimizing maritime operations. This paper reviews deep learning applications in time series analysis within the maritime industry, focusing on three areas: ship operation-related, port operation-related, and shipping market-related topics. It provides a detailed overview of the existing literature on applications such as ship trajectory prediction, ship fuel consumption prediction, port throughput prediction, and shipping market prediction. The paper comprehensively examines the primary deep learning architectures used for time series forecasting in the maritime industry, categorizing them into four principal types. It systematically analyzes the advantages of deep learning architectures across different application scenarios and explores methodologies for selecting models based on specific requirements. Additionally, it analyzes data sources from the existing literature and suggests future research directions. Full article
(This article belongs to the Special Issue Deep Learning Approach for Time Series Forecasting)
Show Figures

Figure 1

Figure 1
<p>The flow chart of data collection. Source: authors.</p>
Full article ">Figure 2
<p>The foundational architecture of the ANN. Source: authors redrawn based on [<a href="#B11-information-15-00507" class="html-bibr">11</a>].</p>
Full article ">Figure 3
<p>The foundational architecture of the WaveNet.</p>
Full article ">Figure 4
<p>Visualization of a stack of causal convolutional layers.</p>
Full article ">Figure 5
<p>Visualization of a stack of dilated causal convolutional layers.</p>
Full article ">Figure 6
<p>The simplified architecture of the ELM.</p>
Full article ">Figure 7
<p>The simplified architecture of the RVFL.</p>
Full article ">Figure 8
<p>The foundational architecture of the CNN. Source: authors redrawn based on [<a href="#B11-information-15-00507" class="html-bibr">11</a>].</p>
Full article ">Figure 9
<p>The foundational architecture of the TCN. Source: authors redrawn based on [<a href="#B37-information-15-00507" class="html-bibr">37</a>].</p>
Full article ">Figure 10
<p>The foundational architecture of the RNN. Source: authors redrawn based on [<a href="#B44-information-15-00507" class="html-bibr">44</a>].</p>
Full article ">Figure 11
<p>The foundational architecture of the LSTM. Source: authors redrawn based on [<a href="#B11-information-15-00507" class="html-bibr">11</a>].</p>
Full article ">Figure 12
<p>The internal details of the LSTM. Source: authors redrawn based on [<a href="#B54-information-15-00507" class="html-bibr">54</a>].</p>
Full article ">Figure 13
<p>The internal details of the GRU. Source: authors redrawn based on [<a href="#B57-information-15-00507" class="html-bibr">57</a>].</p>
Full article ">Figure 14
<p>The architecture of the Transformer. Source: authors redrawn based on [<a href="#B63-information-15-00507" class="html-bibr">63</a>].</p>
Full article ">Figure 15
<p>Percentages of the deep learning algorithms used in the collected literature.</p>
Full article ">Figure 16
<p>Number of annually published papers.</p>
Full article ">Figure 17
<p>The number of papers published in journals or presented at conferences. Notes: The full name of “Journal of Marine Science and…” is “Journal of Marine Science and Engineering”.</p>
Full article ">Figure 18
<p>The number of papers published in journals or presented at conferences.</p>
Full article ">Figure 19
<p>Data utilized in maritime research.</p>
Full article ">
16 pages, 2069 KiB  
Article
Trading Cloud Computing Stocks Using SMA
by Xianrong Zheng and Lingyu Li
Information 2024, 15(8), 506; https://doi.org/10.3390/info15080506 - 21 Aug 2024
Viewed by 902
Abstract
As cloud computing adoption becomes mainstream, the cloud services market offers vast profits. Moreover, serverless computing, the next stage of cloud computing, comes with huge economic potential. To capitalize on this trend, investors are interested in trading cloud stocks. As high-growth technology stocks, [...] Read more.
As cloud computing adoption becomes mainstream, the cloud services market offers vast profits. Moreover, serverless computing, the next stage of cloud computing, comes with huge economic potential. To capitalize on this trend, investors are interested in trading cloud stocks. As high-growth technology stocks, investing in cloud stocks is both rewarding and challenging. The research question here is how a trading strategy will perform on cloud stocks. As a result, this paper employs an effective method—Simple Moving Average (SMA)—to trade cloud stocks. To evaluate its performance, we conducted extensive experiments with real market data that spans over 23 years. Results show that SMA can achieve satisfying performance in terms of several measures, including MAE, RMSE, and R-squared. Full article
(This article belongs to the Special Issue Blockchain Applications for Business Process Management)
Show Figures

Figure 1

Figure 1
<p>Prediction errors of SMA on cloud stock X in 2021.</p>
Full article ">Figure 2
<p>Percentage errors of SMA on cloud stock X in 2021.</p>
Full article ">Figure 3
<p>Prediction errors of SMA on cloud stock X (1999–2021).</p>
Full article ">Figure 4
<p>Percentage errors of SMA on cloud stock X (1999–2021).</p>
Full article ">Figure 5
<p>Prediction errors of SMA on cloud stock Y in 2021.</p>
Full article ">Figure 6
<p>Percentage errors of SMA on cloud stock Y in 2021.</p>
Full article ">Figure 7
<p>Prediction errors of SMA on cloud stock Y (1999–2021).</p>
Full article ">Figure 8
<p>Percentage errors of SMA on cloud stock Y (1999–2021).</p>
Full article ">
29 pages, 1854 KiB  
Article
Information Security Awareness in the Insurance Sector: Cognitive and Internal Factors and Combined Recommendations
by Morgan Djotaroeno and Erik Beulen
Information 2024, 15(8), 505; https://doi.org/10.3390/info15080505 - 21 Aug 2024
Cited by 1 | Viewed by 1600
Abstract
Cybercrime is currently rapidly developing, requiring an increased demand for information security knowledge. Attackers are becoming more sophisticated and complex in their assault tactics. Employees are a focal point since humans remain the ‘weakest link’ and are vital to prevention. This research investigates [...] Read more.
Cybercrime is currently rapidly developing, requiring an increased demand for information security knowledge. Attackers are becoming more sophisticated and complex in their assault tactics. Employees are a focal point since humans remain the ‘weakest link’ and are vital to prevention. This research investigates what cognitive and internal factors influence information security awareness (ISA) among employees, through quantitative empirical research using a survey conducted at a Dutch financial insurance firm. The research question of “How and to what extent do cognitive and internal factors contribute to information security awareness (ISA)?” has been answered, using the theory of situation awareness as the theoretical lens. The constructs of Security Complexity, Information Security Goals (InfoSec Goals), and SETA Programs (security education, training, and awareness) significantly contribute to ISA. The most important research recommendations are to seek novel explaining variables for ISA, further investigate the roots of Security Complexity and what influences InfoSec Goals, and venture into qualitative and experimental research methodologies to seek more depth. The practical recommendations are to minimize the complexity of (1) information security topics (e.g., by contextualizing it more for specific employee groups) and (2) integrate these simplifications in various SETA methods (e.g., gamification and online training). Full article
Show Figures

Figure 1

Figure 1
<p>Research Model.</p>
Full article ">Figure 2
<p>Regression Path Analysis—* <span class="html-italic">p</span> &lt; 0.05, ** <span class="html-italic">p</span> &lt; 0.01, *** <span class="html-italic">p</span> &lt; 0.001.</p>
Full article ">Figure 3
<p>Information Security Awareness from a Situation Awareness Perspective.</p>
Full article ">
26 pages, 10462 KiB  
Article
The Optimal Choice of the Encoder–Decoder Model Components for Image Captioning
by Mateusz Bartosiewicz and Marcin Iwanowski
Information 2024, 15(8), 504; https://doi.org/10.3390/info15080504 - 21 Aug 2024
Viewed by 1021
Abstract
Image captioning aims at generating meaningful verbal descriptions of a digital image. This domain is rapidly growing due to the enormous increase in available computational resources. The most advanced methods are, however, resource-demanding. In our paper, we return to the encoder–decoder deep-learning model [...] Read more.
Image captioning aims at generating meaningful verbal descriptions of a digital image. This domain is rapidly growing due to the enormous increase in available computational resources. The most advanced methods are, however, resource-demanding. In our paper, we return to the encoder–decoder deep-learning model and investigate how replacing its components with newer equivalents improves overall effectiveness. The primary motivation of our study is to obtain the highest possible level of improvement of classic methods, which are applicable in less computational environments where most advanced models are too heavy to be efficiently applied. We investigate image feature extractors, recurrent neural networks, word embedding models, and word generation layers and discuss how each component influences the captioning model’s overall performance. Our experiments are performed on the MS COCO 2014 dataset. As a result of our research, replacing components improves the quality of generating image captions. The results will help design efficient models with optimal combinations of their components. Full article
(This article belongs to the Special Issue Information Processing in Multimedia Applications)
Show Figures

Figure 1

Figure 1
<p>Encoder–decoder merge captioning model.</p>
Full article ">Figure 2
<p>Images (with COCO IDs) with properly predicted captions (see <a href="#information-15-00504-t002" class="html-table">Table 2</a> for details). (<b>a</b>) ID = 95051; (<b>b</b>) ID = 309933; (<b>c</b>) ID = 570138; (<b>d</b>) ID = 105234; (<b>e</b>) ID = 347950; (<b>f</b>) ID = 27440.</p>
Full article ">Figure 3
<p>Images with predicted captions generated artificially: (<b>a</b>) “Dog with frisbee”, (<b>b</b>) “Dog on bike”, (<b>c</b>) “Dog on rocket”, (<b>d</b>) “Tiger on rocket”; real images with objects not present in the MS COCO dataset (<b>e</b>,<b>f</b>).</p>
Full article ">
17 pages, 368 KiB  
Article
An Approach for Maximizing Computation Bits in UAV-Assisted Wireless Powered Mobile Edge Computing Networks
by Zhenbo Liu, Yunge Duan and Shuang Fu
Information 2024, 15(8), 503; https://doi.org/10.3390/info15080503 - 21 Aug 2024
Viewed by 723
Abstract
With the development of the Internet of Things (IoT), IoT nodes with limited energy and computing capability are no longer able to address increasingly complex computational tasks. To address this issue, an Unmanned Aerial Vehicle (UAV)-assisted Wireless Power Transfer (WPT) Mobile Edge Computing [...] Read more.
With the development of the Internet of Things (IoT), IoT nodes with limited energy and computing capability are no longer able to address increasingly complex computational tasks. To address this issue, an Unmanned Aerial Vehicle (UAV)-assisted Wireless Power Transfer (WPT) Mobile Edge Computing (MEC) system is proposed in this study. By jointly optimizing variables such as energy harvesting time, user transmission power, user offloading time, CPU frequency, and UAV deployment location, the system aims to maximize the number of computation bits by the users. This optimization yields a challenging non-convex optimization problem. To address these issues, a two-stage alternating method based on the Lagrangian dual method and the Successive Convex Approximation (SCA) method is proposed to decompose the initial problem into two sub-problems. Firstly, the UAV position is fixed to obtain the optimal values of other variables, and then the UAV position is optimized based on the solved variables. Finally, this iterative process continues until the algorithm convergences, and the optimal solution for the given problem is obtained. The simulation results indicate that the proposed algorithm exhibits good convergence. Compared to other benchmark solutions, the proposed approach performs optimally in maximizing computation bits. Full article
Show Figures

Figure 1

Figure 1
<p>System model.</p>
Full article ">Figure 2
<p>Frame structure.</p>
Full article ">Figure 3
<p>The total number of computation bits for all the user versus the number of iterations.</p>
Full article ">Figure 4
<p>Optimal deployment position of UAV.</p>
Full article ">Figure 5
<p>Comparison of the number of computation bits for each user before and after optimizing the UAV position.</p>
Full article ">Figure 6
<p>The UAV transmit power versus the total number of computation bits for all users.</p>
Full article ">Figure 7
<p>The average number of computation bits per user versus the number of users.</p>
Full article ">Figure 8
<p>The total number of computation bits for all users versus UAV altitude.</p>
Full article ">Figure 9
<p>Computation bits versus the transmission power under different methods.</p>
Full article ">Figure 10
<p>Computation bits versus the length of time block under different methods.</p>
Full article ">
15 pages, 2185 KiB  
Article
Cost Estimation and Prediction for Residential Projects Based on Grey Relational Analysis–Lasso Regression–Backpropagation Neural Network
by Lijun Chen and Dejiang Wang
Information 2024, 15(8), 502; https://doi.org/10.3390/info15080502 - 21 Aug 2024
Cited by 1 | Viewed by 1041
Abstract
In the early stages of residential project investment, accurately estimating the engineering costs of residential projects is crucial for cost control and management of the project. However, the current cost estimation of residential engineering in China is primarily carried out by cost personnel [...] Read more.
In the early stages of residential project investment, accurately estimating the engineering costs of residential projects is crucial for cost control and management of the project. However, the current cost estimation of residential engineering in China is primarily carried out by cost personnel based on their own experience. This process is time-consuming and labour-intensive, and it involves subjective judgement, which can lead to significant estimation errors and fail to meet the rapidly developing market demands. Data collection for residential construction projects is challenging, with small sample sizes, numerous attributes, and complexity. This paper adopts a hybrid method combining a grey relational analysis, Lasso regression, and Backpropagation Neural Network (GAR-LASSO-BPNN). This method has significant advantages in handling high-dimensional small samples and multiple correlated variables. The grey relational analysis (GRA) is used to quantitatively identify cost-driving factors, and 14 highly correlated factors are selected as input variables. Then, regularization through Lasso regression (LASSO) is used to filter the final input variables, which are subsequently input into the Backpropagation Neural Network (BPNN) to establish the relationship between the unit cost of residential projects and 12 input variables. Compared to using LASSO and BPNN methods individually, the GAR-LASSO-BPNN hybrid prediction method performs better in terms of error evaluation metrics. The research findings can provide quantitative decision support for cost estimators in the early estimation stages of residential project investment decision-making. Full article
Show Figures

Figure 1

Figure 1
<p>The flowchart of GRA-LASSO-BPNN.</p>
Full article ">Figure 2
<p>Main workflow of web scraping technology [<a href="#B30-information-15-00502" class="html-bibr">30</a>].</p>
Full article ">Figure 3
<p>Implementation principle flowchart of grey relational analysis [<a href="#B28-information-15-00502" class="html-bibr">28</a>].</p>
Full article ">Figure 4
<p>Pearson correlation analysis results.</p>
Full article ">Figure 5
<p>LASSO variable regression coefficient path solution.</p>
Full article ">Figure 6
<p>Model performance metrics with varying numbers of independent variables. (<b>a</b>) MAPE, (<b>b</b>) R<sup>2</sup>, (<b>c</b>) CP.</p>
Full article ">Figure 6 Cont.
<p>Model performance metrics with varying numbers of independent variables. (<b>a</b>) MAPE, (<b>b</b>) R<sup>2</sup>, (<b>c</b>) CP.</p>
Full article ">Figure 7
<p>Mean Absolute Error as Alpha varies.</p>
Full article ">Figure 8
<p>Visual comparison of performance of three predictive models on test set.</p>
Full article ">
12 pages, 244 KiB  
Perspective
Reviewing the Horizon: The Future of Extended Reality and Artificial Intelligence in Neurorehabilitation for Brain Injury Recovery
by Khalida Akbar, Anna Passaro, Mariacarla Di Gioia, Elvira Martini, Mirella Dragone, Antonio Zullo and Fabrizio Stasolla
Information 2024, 15(8), 501; https://doi.org/10.3390/info15080501 - 21 Aug 2024
Cited by 1 | Viewed by 1631
Abstract
People with disorders of consciousness, either as a consequence of an acquired brain injury or a traumatic brain injury, may pose serious challenges to medical and/or rehabilitative centers with an increased burden on caregivers and families. The objectives of this study were as [...] Read more.
People with disorders of consciousness, either as a consequence of an acquired brain injury or a traumatic brain injury, may pose serious challenges to medical and/or rehabilitative centers with an increased burden on caregivers and families. The objectives of this study were as follows: to explore the use of extended reality as a critical means of rehabilitative support in people with disorders of consciousness and brain injuries; to evaluate its impact on recovery processes; to assess the improvements in the participants’ quality of life, and to reduce the burden on families and caregivers by using extended reality and artificial-intelligence-based programs. A selective review of the newest empirical studies on the use of extended reality and artificial-intelligence-based interventions in patients with brain injuries and disorders of consciousness was conducted over the last decade. The potential for bias in this selective review is acknowledged. A conceptual framework was detailed. The data showed that extended reality and artificial-intelligence-based programs successfully enhanced the adaptive responding of the participants involved, and improved their quality of life. The burden on caregivers and families was reduced accordingly. Extended reality and artificial intelligence may be viewed as crucial means of recovery in people with disorders of consciousness and brain injuries. Full article
(This article belongs to the Special Issue Extended Reality and Cybersecurity)
25 pages, 514 KiB  
Article
Bridging Linguistic Gaps: Developing a Greek Text Simplification Dataset
by Leonidas Agathos, Andreas Avgoustis, Xristiana Kryelesi, Aikaterini Makridou, Ilias Tzanis, Despoina Mouratidis, Katia Lida Kermanidis and Andreas Kanavos
Information 2024, 15(8), 500; https://doi.org/10.3390/info15080500 - 20 Aug 2024
Viewed by 772
Abstract
Text simplification is crucial in bridging the comprehension gap in today’s information-rich environment. Despite advancements in English text simplification, languages with intricate grammatical structures, such as Greek, often remain under-explored. The complexity of Greek grammar, characterized by its flexible syntactic ordering, presents unique [...] Read more.
Text simplification is crucial in bridging the comprehension gap in today’s information-rich environment. Despite advancements in English text simplification, languages with intricate grammatical structures, such as Greek, often remain under-explored. The complexity of Greek grammar, characterized by its flexible syntactic ordering, presents unique challenges that hinder comprehension for native speakers, learners, tourists, and international students. This paper introduces a comprehensive dataset for Greek text simplification, containing over 7500 sentences across diverse topics such as history, science, and culture, tailored to address these challenges. We outline the methodology for compiling this dataset, including a collection of texts from Greek Wikipedia, their annotation with simplified versions, and the establishment of robust evaluation metrics. Additionally, the paper details the implementation of quality control measures and the application of machine learning techniques to analyze text complexity. Our experimental results demonstrate the dataset’s initial effectiveness and potential in reducing linguistic barriers and enhancing communication, with initial machine learning models showing promising directions for future improvements in classifying text complexity. The development of this dataset marks a significant step toward improving accessibility and comprehension for a broad audience of Greek speakers and learners, fostering a more inclusive society. Full article
(This article belongs to the Special Issue Information Extraction and Language Discourse Processing)
Show Figures

Figure 1

Figure 1
<p>Common words between complicated and simple sentences.</p>
Full article ">Figure 2
<p>Top 25 most common words in ’simple’ category.</p>
Full article ">Figure 3
<p>Top 25 most common words in ’complicated’ category.</p>
Full article ">
20 pages, 2982 KiB  
Article
Exploring Tourist Experience through Online Reviews Using Aspect-Based Sentiment Analysis with Zero-Shot Learning for Hospitality Service Enhancement
by Ibrahim Nawawi, Kurnia Fahmy Ilmawan, Muhammad Rifqi Maarif and Muhammad Syafrudin
Information 2024, 15(8), 499; https://doi.org/10.3390/info15080499 - 20 Aug 2024
Viewed by 1444
Abstract
Hospitality services play a crucial role in shaping tourist satisfaction and revisiting intention toward destinations. Traditional feedback methods like surveys often fail to capture the nuanced and real-time experiences of tourists. Digital platforms such as TripAdvisor, Yelp, and Google Reviews provide a rich [...] Read more.
Hospitality services play a crucial role in shaping tourist satisfaction and revisiting intention toward destinations. Traditional feedback methods like surveys often fail to capture the nuanced and real-time experiences of tourists. Digital platforms such as TripAdvisor, Yelp, and Google Reviews provide a rich source of user-generated content, but the sheer volume of reviews makes manual analysis impractical. This study proposes integrating aspect-based sentiment analysis with zero-shot learning to analyze online tourist reviews effectively without requiring extensive annotated datasets. Using pretrained models like RoBERTa, the research framework involves keyword extraction, sentence segment detection, aspect construction, and sentiment polarity measurement. The dataset, sourced from TripAdvisor reviews of attractions, hotels, and restaurants in Central Java, Indonesia, underwent preprocessing to ensure suitability for analysis. The results highlight the importance of aspects such as food, accommodation, and cultural experiences in tourist satisfaction. The findings indicate a need for continuous service improvement to meet evolving tourist expectations, demonstrating the potential of advanced natural language processing techniques in enhancing hospitality services and customer satisfaction. Full article
Show Figures

Figure 1

Figure 1
<p>Step-by-step research framework (originally compiled by authors).</p>
Full article ">Figure 2
<p>General BERT architecture (adapted from Vaswani et al. [<a href="#B30-information-15-00499" class="html-bibr">30</a>]).</p>
Full article ">Figure 3
<p>Word clouds of identified keywords: (<b>a</b>) all initially identified keywords; (<b>b</b>) keywords retained after filtering for relevance and significance.</p>
Full article ">Figure 4
<p>T-SNE plot of keywords’ semantic similarities based on BERT embedding.</p>
Full article ">Figure 5
<p>The sentiment analysis results for each aspect based on frequency.</p>
Full article ">Figure 6
<p>The sentiment analysis results for each aspect based on proportion.</p>
Full article ">Figure 7
<p>The distribution of the sentiment polarity scores of all proposed aspects.</p>
Full article ">
17 pages, 809 KiB  
Article
Intelligent Risk Evaluation for Investment Banking IPO Business Based on Text Analysis
by Lei Zhang, Chao Wang and Xiaoxing Liu
Information 2024, 15(8), 498; https://doi.org/10.3390/info15080498 - 20 Aug 2024
Viewed by 766
Abstract
By constructing a text quality analysis system and company quality analysis system based on a prospectus, the intelligent analysis method of investment banking IPO business risk is proposed based on the machine learning method and text analysis technology. Taking the Sci-Tech Innovation Board [...] Read more.
By constructing a text quality analysis system and company quality analysis system based on a prospectus, the intelligent analysis method of investment banking IPO business risk is proposed based on the machine learning method and text analysis technology. Taking the Sci-Tech Innovation Board in China as a sample, the empirical analysis results show that the text quality and the company quality disclosed in the prospectus can affect the withdrawal rate of investment banking IPO business. By carrying out a text analysis and machine learning on the text quality and company quality, the risk of investment banking IPO business can be predicted intelligently and effectively. The research results can not only improve the business efficiency of investment banking IPO, and save resource cost, but also improve the standardization and authenticity of investment banking IPO business. Full article
(This article belongs to the Section Information Applications)
Show Figures

Figure 1

Figure 1
<p>Mean square error of random forest.</p>
Full article ">Figure 2
<p>Importance ranking of feature selection.</p>
Full article ">
14 pages, 291 KiB  
Article
Minimum Mean Squared Error Estimation and Mutual Information Gain
by Jerry Gibson
Information 2024, 15(8), 497; https://doi.org/10.3390/info15080497 - 20 Aug 2024
Viewed by 1094
Abstract
Information theoretic quantities such as entropy, entropy rate, information gain, and relative entropy are often used to understand the performance of intelligent agents in learning applications. Mean squared error has not played a role in these analyses, primarily because it is not felt [...] Read more.
Information theoretic quantities such as entropy, entropy rate, information gain, and relative entropy are often used to understand the performance of intelligent agents in learning applications. Mean squared error has not played a role in these analyses, primarily because it is not felt to be a viable performance indicator in these scenarios. We build on a new quantity, the log ratio of entropy powers, to establish that minimum mean squared error (MMSE) estimation, prediction, and smoothing are directly connected to mutual information gain or loss in an agent learning system modeled by a Markov chain for many probability distributions of interest. Expressions for mutual information gain or loss are developed for MMSE estimation, prediction, and smoothing, and an example for fixed lag smoothing is presented. Full article
(This article belongs to the Special Issue Fundamental Problems of Information Studies)
26 pages, 3537 KiB  
Article
From Data to Insight: Transforming Online Job Postings into Labor-Market Intelligence
by Giannis Tzimas, Nikos Zotos, Evangelos Mourelatos, Konstantinos C. Giotopoulos and Panagiotis Zervas
Information 2024, 15(8), 496; https://doi.org/10.3390/info15080496 - 20 Aug 2024
Viewed by 1915
Abstract
In the continuously changing labor market, understanding the dynamics of online job postings is crucial for economic and workforce development. With the increasing reliance on Online Job Portals, analyzing online job postings has become an essential tool for capturing real-time labor-market trends. This [...] Read more.
In the continuously changing labor market, understanding the dynamics of online job postings is crucial for economic and workforce development. With the increasing reliance on Online Job Portals, analyzing online job postings has become an essential tool for capturing real-time labor-market trends. This paper presents a comprehensive methodology for processing online job postings to generate labor-market intelligence. The proposed methodology encompasses data source selection, data extraction, cleansing, normalization, and deduplication procedures. The final step involves information extraction based on employer industry, occupation, workplace, skills, and required experience. We address the key challenges that emerge at each step and discuss how they can be resolved. Our methodology is applied to two use cases: the first focuses on the analysis of the Greek labor market in the tourism industry during the COVID-19 pandemic, revealing shifts in job demands, skill requirements, and employment types. In the second use case, a data-driven ontology is employed to extract skills from job postings using machine learning. The findings highlight that the proposed methodology, utilizing NLP and machine-learning techniques instead of LLMs, can be applied to different labor market-analysis use cases and offer valuable insights for businesses, job seekers, and policymakers. Full article
(This article belongs to the Special Issue Second Edition of Predictive Analytics and Data Science)
Show Figures

Figure 1

Figure 1
<p>Methodology.</p>
Full article ">Figure 2
<p>Raw data DB structure.</p>
Full article ">Figure 3
<p>Information-extraction DB structure.</p>
Full article ">Figure 4
<p>Data-cleaning and -preparation procedure.</p>
Full article ">Figure 5
<p>Industry extraction.</p>
Full article ">Figure 6
<p>Surface-form quality-score examples.</p>
Full article ">
27 pages, 19187 KiB  
Article
Analyzing Tor Browser Artifacts for Enhanced Web Forensics, Anonymity, Cybersecurity, and Privacy in Windows-Based Systems
by Muhammad Shanawar Javed, Syed Muhammad Sajjad, Danish Mehmood, Khawaja Mansoor, Zafar Iqbal, Muhammad Kazim and Zia Muhammad
Information 2024, 15(8), 495; https://doi.org/10.3390/info15080495 - 19 Aug 2024
Cited by 2 | Viewed by 2459
Abstract
The Tor browser is widely used for anonymity, providing layered encryption for enhanced privacy. Besides its positive uses, it is also popular among cybercriminals for illegal activities such as trafficking, smuggling, betting, and illicit trade. There is a need for Tor Browser forensics [...] Read more.
The Tor browser is widely used for anonymity, providing layered encryption for enhanced privacy. Besides its positive uses, it is also popular among cybercriminals for illegal activities such as trafficking, smuggling, betting, and illicit trade. There is a need for Tor Browser forensics to identify its use in unlawful activities and explore its consequences. This research analyzes artifacts generated by Tor on Windows-based systems. The methodology integrates forensic techniques into incident responses per NIST SP (800-86), exploring areas such as registry, storage, network, and memory using tools like bulk-extractor, autopsy, and regshot. We propose an automated PowerShell script that detects Tor usage and retrieves artifacts with minimal user interaction. Finally, this research performs timeline analysis and artifact correlation for a contextual understanding of event sequences in memory and network domains, ultimately contributing to improved incident response and accountability. Full article
(This article belongs to the Special Issue Cybersecurity, Cybercrimes, and Smart Emerging Technologies)
Show Figures

Figure 1

Figure 1
<p>Tor components.</p>
Full article ">Figure 2
<p>Working principles of Tor Browser.</p>
Full article ">Figure 3
<p>Flow diagram for the proposed methodology used to analyze Tor Browser artifacts.</p>
Full article ">Figure 4
<p>Results and onion URLs found from memory dumps. Dumps of onion links are highlighted in blue color.</p>
Full article ">Figure 5
<p>Tor Browser installer and artifacts left behind in Windows registry.</p>
Full article ">Figure 6
<p>Tor Browser installed location found in the data recovery process. The installer location is highlighted in blue color.</p>
Full article ">Figure 7
<p>Tor artifacts in prefetch.</p>
Full article ">Figure 8
<p>Hosts from PCAP in Network Miner.</p>
Full article ">Figure 9
<p>Credentials found using Network Miner.</p>
Full article ">Figure 10
<p>An overview of different running processes.</p>
Full article ">Figure 11
<p>Tor usage timeline detected during analysis.</p>
Full article ">Figure 12
<p>Timestamps of visited onion sites retrieved during the analysis.</p>
Full article ">Figure 13
<p>Flow diagram for proposed automated script, which can be used to retrieve artifacts from Tor Browser in Windows environment.</p>
Full article ">Figure 14
<p>Results obtained after executing the developed script.</p>
Full article ">
19 pages, 1050 KiB  
Article
Enhancing Biomedical Question Answering with Large Language Models
by Hua Yang, Shilong Li and Teresa Gonçalves
Information 2024, 15(8), 494; https://doi.org/10.3390/info15080494 - 19 Aug 2024
Viewed by 1537
Abstract
In the field of Information Retrieval, biomedical question answering is a specialized task that focuses on answering questions related to medical and healthcare domains. The goal is to provide accurate and relevant answers to the posed queries related to medical conditions, treatments, procedures, [...] Read more.
In the field of Information Retrieval, biomedical question answering is a specialized task that focuses on answering questions related to medical and healthcare domains. The goal is to provide accurate and relevant answers to the posed queries related to medical conditions, treatments, procedures, medications, and other healthcare-related topics. Well-designed models should efficiently retrieve relevant passages. Early retrieval models can quickly retrieve passages but often with low precision. In contrast, recently developed Large Language Models can retrieve documents with high precision but at a slower pace. To tackle this issue, we propose a two-stage retrieval approach that initially utilizes BM25 for a preliminary search to identify potential candidate documents; subsequently, a Large Language Model is fine-tuned to evaluate the relevance of query–document pairs. Experimental results indicate that our approach achieves comparative performances on the BioASQ and the TREC-COVID datasets. Full article
(This article belongs to the Special Issue Editorial Board Members’ Collection Series: "Information Processes")
Show Figures

Figure 1

Figure 1
<p>An overview of the proposed two-stage BM25-LLMs retrieval model. The first stage employs BM25 for sparse search, and the second stage utilizes a Large Language Model for accurate re-ranking.</p>
Full article ">Figure 2
<p>Evaluation metrics at various top-k values, illustrating the Mean Average Precision (MAP), Recall, and Accuracy for varying sizes of the document retrieval set in this experiment. It demonstrates that a top-100 retrieval set is the k-value beyond which increases in the retrieval set size do not result in proportional improvements in accuracy.</p>
Full article ">Figure 3
<p>The prompt instructs the model to decide if the passage answers the query by responding with ‘Yes’ or ‘No’. The input consists of a query (A) and a passage (B), which are processed by a Large Language Model. The value of the token ‘Yes’ is found in the logits generated by the model for use.</p>
Full article ">Figure 4
<p>In-batch hard negative mining. The figure illustrates the process of in-batch hard negative mining, where the negative samples are selected from the same batch as the positive samples. Solid arrows represent positive sample pairs with high similarity (1.0), while dashed arrows indicate negative sample pairs with lower similarity.</p>
Full article ">Figure 5
<p>Distribution of query lengths with fitted log-normal curve. Observed data (blue) shows raw frequencies; fitted log-normal distribution (red) indicates approximate log-normal behavior with modal length around 5 words.</p>
Full article ">Figure 6
<p>Q–Q plot of log-transformed frequencies. Points roughly follow a straight line, suggesting reasonable agreement with a normal distribution.</p>
Full article ">Figure 7
<p>This histogram illustrates the distribution of query lengths within the TREC-COVID dataset.</p>
Full article ">Figure 8
<p>BioASQ BM25 hyperparameter tuning.</p>
Full article ">Figure 9
<p>TREC-COVID BM25 hyperparameter tuning.</p>
Full article ">Figure 10
<p>Performance comparison of different models across various k-values using MAP and NDCG metrics.</p>
Full article ">
14 pages, 1516 KiB  
Article
Early Recurrence Prediction of Hepatocellular Carcinoma Using Deep Learning Frameworks with Multi-Task Pre-Training
by Jian Song, Haohua Dong, Youwen Chen, Xianru Zhang, Gan Zhan, Rahul Kumar Jain and Yen-Wei Chen
Information 2024, 15(8), 493; https://doi.org/10.3390/info15080493 - 17 Aug 2024
Cited by 1 | Viewed by 997
Abstract
Post-operative early recurrence (ER) of hepatocellular carcinoma (HCC) is a major cause of mortality. Predicting ER before treatment can guide treatment and follow-up protocols. Deep learning frameworks, known for their superior performance, are widely used in medical imaging. However, they face challenges due [...] Read more.
Post-operative early recurrence (ER) of hepatocellular carcinoma (HCC) is a major cause of mortality. Predicting ER before treatment can guide treatment and follow-up protocols. Deep learning frameworks, known for their superior performance, are widely used in medical imaging. However, they face challenges due to limited annotated data. We propose a multi-task pre-training method using self-supervised learning with medical images for predicting the ER of HCC. This method involves two pretext tasks: phase shuffle, focusing on intra-image feature representation, and case discrimination, focusing on inter-image feature representation. The effectiveness and generalization of the proposed method are validated through two different experiments. In addition to predicting early recurrence, we also apply the proposed method to the classification of focal liver lesions. Both experiments show that the multi-task pre-training model outperforms existing pre-training (transfer learning) methods with natural images, single-task self-supervised pre-training, and DINOv2. Full article
(This article belongs to the Special Issue Intelligent Image Processing by Deep Learning)
Show Figures

Figure 1

Figure 1
<p>Overview of proposed method: (<b>a</b>) pretext task; (<b>b</b>) target task.</p>
Full article ">Figure 2
<p>The architecture of the ResNet18 network. Different colors are used to mark the five stages of convolutional layers in ResNet18.</p>
Full article ">Figure 3
<p>(<b>A</b>–<b>C</b>) NC, ART, and PV phases, respectively. The region of interest (ROI) is the bounding box of the tumor.</p>
Full article ">Figure 4
<p>Evolution patterns of four FLLs in multi-phase CT.</p>
Full article ">
14 pages, 4112 KiB  
Article
A Feasibility Study of a Respiratory Rate Measurement System Using Wearable MOx Sensors
by Mitsuhiro Fukuda, Jaakko Hyry, Ryosuke Omoto, Takunori Shimazaki, Takumi Kobayashi and Daisuke Anzai
Information 2024, 15(8), 492; https://doi.org/10.3390/info15080492 - 16 Aug 2024
Viewed by 785
Abstract
Accurately obtaining a patient’s respiratory rate is crucial for promptly identifying any sudden changes in their condition during emergencies. Typically, the respiratory rate is assessed through a combination of impedance change measurements and electrocardiography (ECG). However, impedance measurements are prone to interference from [...] Read more.
Accurately obtaining a patient’s respiratory rate is crucial for promptly identifying any sudden changes in their condition during emergencies. Typically, the respiratory rate is assessed through a combination of impedance change measurements and electrocardiography (ECG). However, impedance measurements are prone to interference from body movements. Conversely, a capnometer coupled with a ventilator offers a method of measuring the respiratory rate that is unaffected by body movements. However, capnometers are mainly used to evaluate respiration when using a ventilator or an Ambu bag by measuring the CO2 concentration at the breathing circuit, and they are not used only to measure the respiratory rate. Furthermore, capnometers are not suitable as wearable devices because they require intubation or a mask that covers the nose and mouth to prevent air leaks during the measurement. In this study, we developed a reliable system for measuring the respiratory rate utilizing a small wearable MOx sensor that is unaffected by body movements and not connected to the breathing circuit. Subsequently, we conducted experimental assessments to gauge the accuracy of the rate estimation achieved by the system. In order to avoid the effects of abnormal states on the estimation accuracy, we also evaluated the classification performance for distinguishing between normal and abnormal respiration using a one-class SVM-based approach. The developed system achieved 80% for both true positive and true negative rates. Our experimental findings reveal that the respiratory rate can be precisely determined without being influenced by body movements. Full article
Show Figures

Figure 1

Figure 1
<p>MOx sensor working principle, where oxygen is adsorbed onto the surface of the metal oxide.</p>
Full article ">Figure 2
<p>Circuit configuration of the wearable MOx sensor for detecting (eCO<sub>2</sub>) from the VOC and TVOC levels.</p>
Full article ">Figure 3
<p>Measured waveforms of the wearable MOx sensor, showing the eCO<sub>2</sub> concentration peaks (indicated by the arrows).</p>
Full article ">Figure 4
<p>Experimental environment: (<b>a</b>) physical setup overview, (<b>b</b>) attachment of the wearable MOx sensor in the mouth shield, and (<b>c</b>) equipment configuration diagram.</p>
Full article ">Figure 5
<p>Experimental results at rest: (<b>a</b>) eCO<sub>2</sub> concentrations and (<b>b</b>) normalized filter output, where the idling time of the eCO<sub>2</sub> sensor is indicated with the blue bar.</p>
Full article ">Figure 6
<p>Experimental results of time waveforms for eCO<sub>2</sub> concentration during body movement: (<b>a</b>) eCO<sub>2</sub> concentrations and (<b>b</b>) normalized filter output.</p>
Full article ">Figure 7
<p>Comparison of the estimated respiratory rates at rest and during body movement.</p>
Full article ">Figure 8
<p>(<b>a</b>) eCO<sub>2</sub> concentration obtained in this experiment (cough, yawn), and (<b>b</b>) normalized filter output (cough, yawn), where the blue, orange, and yellow bars indicate the idling of the sensor, coughing, and yawning, respectively.</p>
Full article ">Figure 9
<p>Comparison of the estimated respiratory rates for coughing and yawning.</p>
Full article ">Figure 10
<p>eCO<sub>2</sub> concentration in a rest state with an observed outlier.</p>
Full article ">Figure 11
<p>Examples of short-time Fourier transform (−5–5 Hz) measurements, showing the differences in the nasal, mouth, cough, and yawn frequency characteristics.</p>
Full article ">Figure 12
<p>Changes in the estimation accuracy for true negative rate (TNR), false positive rate (FPR), true positive rate (TPR), and false negative rate (FNR).</p>
Full article ">
18 pages, 1873 KiB  
Article
Beyond Supervised: The Rise of Self-Supervised Learning in Autonomous Systems
by Hamed Taherdoost
Information 2024, 15(8), 491; https://doi.org/10.3390/info15080491 - 16 Aug 2024
Viewed by 1977
Abstract
Supervised learning has been the cornerstone of many successful medical imaging applications. However, its reliance on large labeled datasets poses significant challenges, especially in the medical domain, where data annotation is time-consuming and expensive. In response, self-supervised learning (SSL) has emerged as a [...] Read more.
Supervised learning has been the cornerstone of many successful medical imaging applications. However, its reliance on large labeled datasets poses significant challenges, especially in the medical domain, where data annotation is time-consuming and expensive. In response, self-supervised learning (SSL) has emerged as a promising alternative, leveraging unlabeled data to learn meaningful representations without explicit supervision. This paper provides a detailed overview of supervised learning and its limitations in medical imaging, underscoring the need for more efficient and scalable approaches. The study emphasizes the importance of the area under the curve (AUC) as a key evaluation metric in assessing SSL performance. The AUC offers a comprehensive measure of model performance across different operating points, which is crucial in medical applications, where false positives and negatives have significant consequences. Evaluating SSL methods based on the AUC allows for robust comparisons and ensures that models generalize well to real-world scenarios. This paper reviews recent advances in SSL for medical imaging, demonstrating their potential to revolutionize the field by mitigating challenges associated with supervised learning. Key results show that SSL techniques, by leveraging unlabeled data and optimizing performance metrics like the AUC, can significantly improve the diagnostic accuracy, scalability, and efficiency in medical image analysis. The findings highlight SSL’s capability to reduce the dependency on labeled datasets and present a path forward for more scalable and effective medical imaging solutions. Full article
(This article belongs to the Special Issue Emerging Research on Neural Networks and Anomaly Detection)
Show Figures

Figure 1

Figure 1
<p>SSL concept.</p>
Full article ">Figure 2
<p>AUC through TPR and FPR.</p>
Full article ">Figure 3
<p>Process of selecting papers for review.</p>
Full article ">Figure 4
<p>Overview of methodologies and applications in included studies.</p>
Full article ">
23 pages, 1044 KiB  
Article
Optimized Early Prediction of Business Processes with Hyperdimensional Computing
by Fatemeh Asgarinejad, Anthony Thomas, Ryan Hildebrant, Zhenyu Zhang, Shangping Ren, Tajana Rosing and Baris Aksanli
Information 2024, 15(8), 490; https://doi.org/10.3390/info15080490 - 16 Aug 2024
Viewed by 747
Abstract
There is a growing interest in the early prediction of outcomes in ongoing business processes. Predictive process monitoring distills knowledge from the sequence of event data generated and stored during the execution of processes and trains models on this knowledge to predict outcomes [...] Read more.
There is a growing interest in the early prediction of outcomes in ongoing business processes. Predictive process monitoring distills knowledge from the sequence of event data generated and stored during the execution of processes and trains models on this knowledge to predict outcomes of ongoing processes. However, most state-of-the-art methods require the training of complex and inefficient machine learning models and hyper-parameter optimization as well as numerous input data to achieve high performance. In this paper, we present a novel approach based on Hyperdimensional Computing (HDC) for predicting the outcome of ongoing processes before their completion. We highlight its simplicity, efficiency, and high performance while utilizing only a subset of the input data, which helps in achieving a lower memory demand and faster and more effective corrective measures. We evaluate our proposed method on four publicly available datasets with a total of 12 binary prediction tasks. Our proposed method achieves an average 6% higher area under the ROC curve (AUC) and up to a 14% higher F1-score, while yielding a 20× earlier prediction than state-of-the-art conventional machine learning- and neural network-based models. Full article
(This article belongs to the Special Issue Second Edition of Predictive Analytics and Data Science)
Show Figures

Figure 1

Figure 1
<p>(<b>a</b>) HDC encoding and training. Each training data point (business process sequence) is encoded to a hypervector, and then in the training phase, it is added up to the proper class hypervector based on its label. This operation is a simple element-wise addition. In (<b>b</b>), HDC Inference, a business process query sequence is similarly encoded and compared with all the class hypervectors. The index of the class with the highest similarity is returned as a prediction result.</p>
Full article ">Figure 2
<p>AUC score over different prefixes (process lengths) for BPIC2012—Approved dataset with different D values.</p>
Full article ">Figure 3
<p>Proposed HDC framework steps: encoding, which includes (1) projecting unique event IDs and unique attributes into hypervectors and (2) binding attributes’ hypervectors to the events, encoding tri-grams, and superimposing for generating the hypervector corresponding to each input trace; training (3), wherein class hypervectors are generated with the element-wise summation of elements belonging to the class; querying (4), where the label of a query data is predicted using a similarity metric; and (5) retraining, which includes updating class vectors based on the correct/incorrect label prediction of the query data.</p>
Full article ">Figure 4
<p>AUC score over different prefixes (process lengths) for BPIC2012—Approved dataset with different <span class="html-italic">n</span> values in <span class="html-italic">n</span>—gram encoding.</p>
Full article ">Figure 5
<p>AUC score over different prefixes (process lengths) for BPIC2012—Approved dataset for different methods. Red and orange curves, respectively, depict Att-HDC and HDC, which yield higher AUCs in early prefixes of traces.</p>
Full article ">Figure 6
<p>AUC and F1-score for various training data ratios on the BPIC2012-Approved dataset.</p>
Full article ">
15 pages, 956 KiB  
Article
Enhancing Brain Tumour Multi-Classification Using Efficient-Net B0-Based Intelligent Diagnosis for Internet of Medical Things (IoMT) Applications
by Amna Iqbal, Muhammad Arfan Jaffar and Rashid Jahangir
Information 2024, 15(8), 489; https://doi.org/10.3390/info15080489 - 16 Aug 2024
Viewed by 1173
Abstract
Brain tumour disease develops due to abnormal cell proliferation. The early identification of brain tumours is vital for their effective treatment. Most currently available examination methods are laborious, require extensive manual instructions, and produce subpar findings. The EfficientNet-B0 architecture was used to diagnose [...] Read more.
Brain tumour disease develops due to abnormal cell proliferation. The early identification of brain tumours is vital for their effective treatment. Most currently available examination methods are laborious, require extensive manual instructions, and produce subpar findings. The EfficientNet-B0 architecture was used to diagnose brain tumours using magnetic resonance imaging (MRI). The fine-tuned EffeceintNet B0 model was proposed for the Internet of Medical Things (IoMT) environment. The fine-tuned EfficientNet-B0 architecture was employed to classify four different stages of brain tumours from the MRI images. The fine-tuned model showed 99% accuracy in the detection of four different classes of brain tumour detection (glioma, no tumour, meningioma, and pituitary). The proposed model performed very well in the detection of the pituitary class with a precision of 0.95, recall of 0.98, and F1 score of 0.96. The proposed model also performed very well in the detection of the no-tumour class with values of 0.99, 0.90, and 0.94 for precision, recall, and the F1 score, respectively. The precision, recall, and F1 scores for Glioma and Meningioma classes were also high. The proposed solution has several implications for enhancing clinical investigations of brain tumours. Full article
Show Figures

Figure 1

Figure 1
<p>Normal brain and brain with tumour.</p>
Full article ">Figure 2
<p>Proposed IoMT architecture for the MRI-based classification of brain tumours.</p>
Full article ">Figure 3
<p>Flowchart of the proposed methodology for model training and evaluation.</p>
Full article ">Figure 4
<p>MRI scans of brain tumour.</p>
Full article ">Figure 5
<p>EfficientNet-B0 architecture.</p>
Full article ">Figure 6
<p>Training and validation: (<b>a</b>) accuracy (<b>b</b>) loss of simple EfficientNet-B0.</p>
Full article ">Figure 7
<p>Testing confusion matrix for EfficientNet-B0.</p>
Full article ">Figure 8
<p>Testing ROC for EfficientNet-B0.</p>
Full article ">Figure 9
<p>Training and validation: (<b>a</b>) accuracy; (<b>b</b>) loss of proposed EfficientNet-B0.</p>
Full article ">Figure 10
<p>Testing confusion matrix for proposed EfficientNet-B0.</p>
Full article ">Figure 11
<p>Testing ROC for proposed EfficientNet-B0.</p>
Full article ">
12 pages, 1035 KiB  
Article
A Methodology to Distribute On-Chip Voltage Regulators to Improve the Security of Hardware Masking
by Soner Seçkiner and Selçuk Köse
Information 2024, 15(8), 488; https://doi.org/10.3390/info15080488 - 16 Aug 2024
Viewed by 788
Abstract
Hardware masking is used to protect against side-channel attacks by splitting sensitive information into different parts, called hardware masking shares. Ideally, a side-channel attack would only work if all these parts were completely independent. But in real-world VLSI implementations, things are not perfect. [...] Read more.
Hardware masking is used to protect against side-channel attacks by splitting sensitive information into different parts, called hardware masking shares. Ideally, a side-channel attack would only work if all these parts were completely independent. But in real-world VLSI implementations, things are not perfect. Information from a hardware masking share can leak to another, making it possible for side-channel attacks to succeed without needing data from every hardware masking share. The theoretically supposed independence of these shares often does not hold up in practice. The effectiveness of hardware masking is reduced because of the parasitic impedance that stems from power delivery networks or the internal structure of the integrated circuit. When the coupling effect and noise spread among the hardware masking shares powered by the same power delivery network, side-channel attacks can be carried out with fewer measurements. To address this, we propose a new method of distributing on-chip voltage regulators to improve hardware masking security. The benefits of distributed on-chip voltage regulators are evident. Placing the regulators close to the load minimizes power loss due to resistive losses in the power delivery network. Localized regulation allows for more efficient adjustments to the varying power demands of different chip sections, improving overall power efficiency. Additionally, distributed regulators can quickly respond to power demand changes, maintaining stable voltage levels for high-performance circuits, leading to improved control over noise. We introduce a new DLDO voltage regulator that uses random clocking and randomizing limit cycle oscillations to enhance security. Our simulations show that with these distributed DLDO regulators, the t-test value can be as low as 2.019, and typically, a circuit with a t-test value below 4.5 is considered secure. Full article
(This article belongs to the Special Issue Hardware Security and Trust)
Show Figures

Figure 1

Figure 1
<p>Simplified PDN model with masking shares and other circuitry where <span class="html-italic">R</span> is parasitic resistance, <math display="inline"><semantics> <msub> <mi>C</mi> <mrow> <mi>d</mi> <mi>e</mi> <mi>c</mi> <mi>a</mi> <mi>p</mi> </mrow> </msub> </semantics></math> is decoupling capacitor, and <span class="html-italic">C</span> is the parasitic capacitance.</p>
Full article ">Figure 2
<p>Proposed partition framework for the NxM grid power delivery network, where <span class="html-italic">t</span>-score is the <span class="html-italic">t</span>-test value.</p>
Full article ">Figure 3
<p>Power grid divided into four quadrants.</p>
Full article ">Figure 4
<p>Proposed DLDO.</p>
Full article ">Figure 5
<p>Schematic of bi-directional shift register [<a href="#B24-information-15-00488" class="html-bibr">24</a>,<a href="#B26-information-15-00488" class="html-bibr">26</a>].</p>
Full article ">Figure 6
<p>Activity of a bi-directional shift register [<a href="#B24-information-15-00488" class="html-bibr">24</a>].</p>
Full article ">Figure 7
<p>Description of the experiments and security analysis framework.</p>
Full article ">Figure 8
<p>The t-test scores at the location of voltage regulators and first-order hardware masking with two shares. The voltage regulators are colored orange with a circle, and hardware masking shares are colored blue with a star. (<b>a</b>) One ideal voltage regulator at the center and two hardware masking shares at the edges. (<b>b</b>) Four ideal voltage regulators at the center of the quadrants and two hardware masking shares at the edges. (<b>c</b>) Seven ideal voltage regulators at the center of the quadrants and two hardware masking shares at the edges.</p>
Full article ">Figure 9
<p>The <span class="html-italic">t</span>-test scores at the location of voltage regulators and first-order hardware masking with two shares. The voltage regulators are colored green with a circle and hardware masking shares are colored blue with a star. (<b>a</b>) One conventional DLDO at the center and two hardware masking shares at the edges of the power grid. (<b>b</b>) Four conventional DLDOs at the center of the quadrants and two hardware masking shares at the edges. (<b>c</b>) Seven conventional DLDOs at the center of the quadrants and two hardware masking shares at the edges.</p>
Full article ">Figure 10
<p>The <span class="html-italic">t</span>-test scores at the location of the proposed DLDO voltage regulators and first-order hardware masking with two shares. The voltage regulators are colored red with a circle, and hardware masking shares are colored blue with a star. (<b>a</b>) One proposed DLDO voltage regulator at the center and two hardware masking shares at the edges. (<b>b</b>) Four proposed DLDOs at the center of quadrants and two hardware masking shares at the edges.</p>
Full article ">Figure 11
<p>Seven conventional DLDOs are located according to different topologies. The conventional DLDO voltage regulators are colored green with a circle and hardware masking shares are colored blue with a star. (<b>a</b>) Seven conventional DLDOs at the center of the quadrants and two hardware masking shares at the edges of the power grid. (<b>b</b>) Seven conventional DLDOs according to the top–bottom topology. Two hardware masking shares at the edges of the power grid [<a href="#B27-information-15-00488" class="html-bibr">27</a>,<a href="#B28-information-15-00488" class="html-bibr">28</a>]. (<b>c</b>) Seven conventional DLDOs according to the daisy chain topology. Two hardware masking shares at the edges of the power grid [<a href="#B29-information-15-00488" class="html-bibr">29</a>].</p>
Full article ">
35 pages, 8757 KiB  
Review
From Information to Knowledge: A Role for Knowledge Networks in Decision Making and Action Selection
by Jagmeet S. Kanwal
Information 2024, 15(8), 487; https://doi.org/10.3390/info15080487 - 15 Aug 2024
Viewed by 1093
Abstract
The brain receives information via sensory inputs through the peripheral nervous system and stores a small subset as memories within the central nervous system. Short-term, working memory is present in the hippocampus whereas long-term memories are distributed within neural networks throughout the brain. [...] Read more.
The brain receives information via sensory inputs through the peripheral nervous system and stores a small subset as memories within the central nervous system. Short-term, working memory is present in the hippocampus whereas long-term memories are distributed within neural networks throughout the brain. Elegant studies on the mechanisms for memory storage and the neuroeconomic formulation of human decision making have been recognized with Nobel Prizes in Physiology or Medicine and in Economics, respectively. There is a wide gap, however, in our understanding of how memories of disparate bits of information translate into “knowledge”, and the neural mechanisms by which knowledge is used to make decisions. I propose that the conceptualization of a “knowledge network” for the creation, storage and recall of knowledge is critical to start bridging this gap. Knowledge creation involves value-driven contextualization of memories through cross-validation via certainty-seeking behaviors, including rumination or reflection. Knowledge recall, like memory, may occur via oscillatory activity that dynamically links multiple networks. These networks may show correlated activity and interactivity despite their presence within widely separated regions of the nervous system, including the brainstem, spinal cord and gut. The hippocampal–amygdala complex together with the entorhinal and prefrontal cortices are likely components of multiple knowledge networks since they participate in the contextual recall of memories and action selection. Sleep and reflection processes and attentional mechanisms mediated by the habenula are expected to play a key role in knowledge creation and consolidation. Unlike a straightforward test of memory, determining the loci and mechanisms for the storage and recall of knowledge requires the implementation of a naturalistic decision-making paradigm. By formalizing a neuroscientific concept of knowledge networks, we can experimentally test their functionality by recording large-scale neural activity during decision making in awake, naturally behaving animals. These types of studies are difficult but important also for advancing knowledge-driven as opposed to big data-driven models of artificial intelligence. A knowledge network-driven understanding of brain function may have practical implications in other spheres, such as education and the treatment of mental disorders. Full article
Show Figures

Figure 1

Figure 1
<p>The parallel–hierarchical processing of sensory inputs leads to the extraction of information from features and objects defined by the information-bearing elements or IBEs. The physical proximity and/or temporal coherence of extracted features create objects, and that of objects within a scene creates perceptual associations or memories. Most of this processing is accomplished via ascending lemniscal pathways in the brain. Thalamocortical loops facilitate egocentric selection by neurons tuned to the parameters of an incoming stimulus, and signal amplification occurs via descending projections (see arrows) [<a href="#B67-information-15-00487" class="html-bibr">67</a>].</p>
Full article ">Figure 2
<p>A schematic showing behavior-driven flow of information to extract object- and scene-specific cues for creating navigational knowledge. Associative memories are typically created via valence-driven idiothetic cues, or show statistically significant coincidence of occurrence in the form of allothetic cues. KNs are expected to play an important role in top-down modulation for sensory selection by sustaining attention at various levels of sensory processing and may be modified by reward- and aversion-driven associative memory mechanisms.</p>
Full article ">Figure 3
<p>(<b>A</b>). Networks and cell types discovered for navigation in rodents, bats and primates, including humans. Lines connect to the brain structures where they are located and arrows show the directionality in which networks and one cell type carves the receptive field of another. (<b>B</b>). Examples of research questions that can be addressed using a KN-driven approach to provide a foundational understanding of the neuroscience of knowledge. These approaches can help us to define the physiological properties, configuration, extent and plasticity of KNs. MEC: medial entorhinal cortex.</p>
Full article ">Figure 4
<p>Schematic showing the localization of oscillatory activity (top-left) in the ventral hippocampus within a navigational context (see <a href="#information-15-00487-f002" class="html-fig">Figure 2</a>) and for memory functions, and its complex interconnectivity with the amygdala for context-dependent associative learning. The lateral amygdala (LH) receives processed sensory inputs from the cortex and thalamus. These inputs contain information about features and objects in the sensory landscape. The subiculum, presubiculum and parasubiculum are extensions of the CA1, CA2 and CA3 regions of the hippocampus and they all receive outputs from the basal amygdala (BA). The basolateral amygdala (BLA) has reciprocal connections with the prefrontal cortex. Projections to the hypothalamus trigger hormonal changes, and those to the periaqueductal gray in the brainstem control respiration, heart rate and vocalization.</p>
Full article ">Figure 5
<p>Knowledge network construction and distinction. (<b>A</b>) A generic multilayered network for signal extraction or associative learning, including recurrence (looping back arrows on green circles) that can be either excitatory or inhibitory, and feedforward (arrows) and feedback (filled circles) connections. The green, blue and orange circles belong to hidden layers of the network. (<b>B</b>) A series of modular networks showing the bottom-up flow of information to extract IBEs, features and objects from the sensory landscape (see <a href="#information-15-00487-f001" class="html-fig">Figure 1</a>). (<b>C</b>) A putative knowledge network consisting of multiple distributed memory modules, predictive and attentional, and multimodal cross-validating networks as the basic components. This type of network contains multimodal representations, and has long-distance connections with inputs that can modify as well as query the network via sensory and cephalic triggers. Information is expected to flow bidirectionally (gray arrows) between these networks for decision making and action selection. The entire network can be triggered by either query or contextual signals. Transient coupling between networks (dashed arrows) can occur via oscillatory activity (see text for details). Physical dissociation via synaptic degradation or temporal unbinding over time is also possible via the de-correlation of neural activity. Prediction mismatch triggers large mismatch negativity-evoked potentials, leading to error signals and potential re-learning that may result in a reconfiguration of the knowledge network.</p>
Full article ">Figure 6
<p>A line diagram depicting the relationship between sensory inputs, motor control and behavior within a sense of self (idiothethic) in both physical (time and space) and physiological terms. The reciprocal interconnectivity of the various functions emphasizes the importance of and role of various neural states and mechanisms that constitute components of KNs. The behavior or motor output is typically triggered in response to a query generated by signals from the internal or the external environment. Physiological states, such as hunger, drive an organism to attend to contextual or allothetic cues in its environment and activate central motor programs in the brain to trigger behavioral actions. These in turn become the source of sensory inputs that allow an organism to monitor its movement and location within its relevant region of space. Some of this feedback leads to contextual learning through motivational cues related to reward and fear. Direct idiothetic feedback (diagonal arrow) can modulate coordinated and directed motor activity.</p>
Full article ">Figure 7
<p>Flow-chart showing the evolution of organisms with or without “brains” (shaded boxes) within the animal kingdom. Early forms of life (over 540 million years ago) did not require knowledge to adapt to their environment. Their action selection was directly determined by sensors and effectors operating at either molecular, cellular or simple reflex levels. Later, as organisms and their brains became more complex, just-in-time information availability was supplanted by KNs that allowed a brain to be proactive in terms of evaluating its environment, goal-setting and action selection. Some molluscan species with large brains, such as the octopus, and protochordates are expected to have rudimentary KNs, whereas the neurally advanced species, such as cetaceans and primates, have the capacity of knowledge abstraction and manipulation. Humans are assumed to have the most advanced KNs capable of symbolic representation through cultural evolution and education.</p>
Full article ">Figure 8
<p>Information flow to extract object- and scene-specific information and create memories as well as knowledge. KNs are expected to play an important role in top-down modulation for sensory selection by sustaining attention at various levels of sensory processing.</p>
Full article ">Figure 9
<p>A diagrammatic representation of a “pyramid-inversion model” for the transfer of information over time from the sensory to the brain environment via a series of filters (<b>left</b>). The conversion to knowledge (<b>right</b>) occurs via oscillatory mechanisms (<b>middle</b>) involving the transfer of short-term to long-term memories. The theta band (0.2 to 0.8 Hz oscillations), during the up-state in nonREM sleep, consolidates memories and integrates them with contextual and cross-validated information as knowledge. Delta waves attenuate sleep spindle activity occurring during the down-state of sleep leading to the forgetting and incorporation of new memories [<a href="#B240-information-15-00487" class="html-bibr">240</a>]. Together with sharp wave ripples (140 to 200 Hz) during sleep and sustained reflection, this activity can modify memories stored within KNs [<a href="#B247-information-15-00487" class="html-bibr">247</a>,<a href="#B250-information-15-00487" class="html-bibr">250</a>].</p>
Full article ">
16 pages, 1563 KiB  
Article
Assessment in the Age of Education 4.0: Unveiling Primitive and Hidden Parameters for Evaluation
by Anil Verma, Parampreet Kaur and Aman Singh
Information 2024, 15(8), 486; https://doi.org/10.3390/info15080486 - 15 Aug 2024
Viewed by 913
Abstract
This study delves into the nuanced aspects that influence the quality of education within the Education 4.0 framework. Education 4.0 epitomizes a contemporary educational paradigm leveraging IoT devices, sensors, and actuators to facilitate real-time and continuous assessment, thereby enhancing student evaluation methodologies. Within [...] Read more.
This study delves into the nuanced aspects that influence the quality of education within the Education 4.0 framework. Education 4.0 epitomizes a contemporary educational paradigm leveraging IoT devices, sensors, and actuators to facilitate real-time and continuous assessment, thereby enhancing student evaluation methodologies. Within this context, the study scrutinizes the pivotal role of infrastructure, learning environment, and faculty, acknowledged as fundamental determinants of educational excellence. Identifying five discrete yet crucial hidden parameters, awareness, accessibility, participation, satisfaction, and academic loafing, this paper meticulously examines their ramifications within the Education 4.0 landscape. Employing a comparative analysis encompassing pre- and post-implementation scenarios, the research assesses the transformative impact of Education 4.0 on the educational sector while dissecting the influence of these hidden parameters across these temporal contexts. The findings underscore the substantial enhancements introduced by Education 4.0, including the provision of real-time and continuous assessment mechanisms, heightened accessibility to educational resources, and amplified student engagement levels. Notably, the study advocates for bolstering stakeholders’ accountability as a strategic measure to mitigate academic loafing within an ambient educational milieu. In essence, this paper offers invaluable insights into the intricate interplay between hidden parameters and educational quality, elucidating the pivotal role of Education 4.0 in catalyzing advancements within the education industry. Full article
Show Figures

Figure 1

Figure 1
<p>Direct/Primitive Parameters contributing for quality assessment.</p>
Full article ">Figure 2
<p>Direct/primitive parameters for assessment.</p>
Full article ">Figure 3
<p>Flowchart of investigation methodology.</p>
Full article ">Figure 4
<p>Pre-Education 4.0 mathematical model.</p>
Full article ">Figure 5
<p>Post-Education 4.0 mathematical model.</p>
Full article ">
25 pages, 4636 KiB  
Article
Application of Multi-Source Remote Sensing Data and Machine Learning for Surface Soil Moisture Mapping in Temperate Forests of Central Japan
by Kyaw Win, Tamotsu Sato and Satoshi Tsuyuki
Information 2024, 15(8), 485; https://doi.org/10.3390/info15080485 - 15 Aug 2024
Cited by 1 | Viewed by 2130
Abstract
Surface soil moisture (SSM) is a key parameter for land surface hydrological processes. In recent years, satellite remote sensing images have been widely used for SSM estimation, and many methods based on satellite-derived spectral indices have also been used to estimate the SSM [...] Read more.
Surface soil moisture (SSM) is a key parameter for land surface hydrological processes. In recent years, satellite remote sensing images have been widely used for SSM estimation, and many methods based on satellite-derived spectral indices have also been used to estimate the SSM content in various climatic conditions and geographic locations. However, achieving an accurate estimation of SSM content at a high spatial resolution remains a challenge. Therefore, improving the precision of SSM estimation through the synergies of multi-source remote sensing data has become imperative, particularly for informing forest management practices. In this study, the integration of multi-source remote sensing data with random forest and support vector machine models was conducted using Google Earth Engine in order to estimate the SSM content and develop SSM maps for temperate forests in central Japan. The synergy of Sentinel-2 and terrain factors, such as elevation, slope, aspect, slope steepness, and valley depth, with the random forest model provided the most suitable approach for SSM estimation, yielding the highest accuracy values (overall accuracy for testing = 91.80%, Kappa = 87.18%, r = 0.98) for the temperate forests of central Japan. This finding provides more valuable information for SSM mapping, which shows promise for precision forestry applications. Full article
(This article belongs to the Special Issue Machine Learning and Artificial Intelligence with Applications)
Show Figures

Figure 1

Figure 1
<p>Overview of the study site: (<b>a</b>) Map of Japan. (<b>b</b>) Map of the University of Tokyo Chiba Forest. (<b>c</b>) Sample locations (the symbol of each point represents each sample).</p>
Full article ">Figure 1 Cont.
<p>Overview of the study site: (<b>a</b>) Map of Japan. (<b>b</b>) Map of the University of Tokyo Chiba Forest. (<b>c</b>) Sample locations (the symbol of each point represents each sample).</p>
Full article ">Figure 2
<p>SSM ground-truth data values for the different forest types of the study site.</p>
Full article ">Figure 3
<p>Result of variable importance among spectral indices.</p>
Full article ">Figure 4
<p>Flowchart of the data processing and analysis steps.</p>
Full article ">Figure 5
<p>Accuracy results of five synergies used by (<b>a</b>) RF model and (<b>b</b>) SVM model.</p>
Full article ">Figure 6
<p>Spatial distribution of SSM in temperate forests of UTCBF: (<b>a</b>) synergy 1 by RF, (<b>b</b>) synergy 1 by SVM, (<b>c</b>) synergy 2 by RF, (<b>d</b>) synergy 2 by SVM, (<b>e</b>) synergy 3 by RF, (<b>f</b>) synergy 3 by SVM, (<b>g</b>) synergy 4 by RF, (<b>h</b>) synergy 4 by SVM, (<b>i</b>) synergy 5 by RF, and (<b>j</b>) synergy 5 by SVM.</p>
Full article ">
48 pages, 894 KiB  
Review
Earlier Decision on Detection of Ransomware Identification: A Comprehensive Systematic Literature Review
by Latifa Albshaier, Seetah Almarri and M. M. Hafizur Rahman
Information 2024, 15(8), 484; https://doi.org/10.3390/info15080484 - 14 Aug 2024
Viewed by 3833
Abstract
Cybersecurity is normally defined as protecting systems against all kinds of cyberattacks; however, due to the rapid and permanent expansion of technology and digital transformation, the threats are also increasing. One of those new threats is ransomware, which is a form of malware [...] Read more.
Cybersecurity is normally defined as protecting systems against all kinds of cyberattacks; however, due to the rapid and permanent expansion of technology and digital transformation, the threats are also increasing. One of those new threats is ransomware, which is a form of malware that aims to steal user’s money. Ransomware is a form of malware that encrypts a victim’s files. The attacker then demands a ransom from the victim to restore access to the data upon a large payment. Ransomware is a way of stealing money in which a user’s files are encrypted and the decrypted key is held by the attacker until a ransom amount is paid by the victim. This systematic literature review (SLR) highlights recent papers published between 2020 and 2024. This paper examines existing research on early ransomware detection methods, focusing on the signs, frameworks, and techniques used to identify and detect ransomware before it causes harm. By analyzing a wide range of academic papers, industry reports, and case studies, this review categorizes and assesses the effectiveness of different detection methods, including those based on signatures, behavior patterns, and machine learning (ML). It also looks at new trends and innovative strategies in ransomware detection, offering a classification of detection techniques and pointing out the gaps in current research. The findings provide useful insights for cybersecurity professionals and researchers, helping guide future efforts to develop strong and proactive ransomware detection systems. This review emphasizes the need for ongoing improvements in detection technologies to keep up with the constantly changing ransomware threat landscape. Full article
(This article belongs to the Special Issue Cybersecurity, Cybercrimes, and Smart Emerging Technologies)
Show Figures

Figure 1

Figure 1
<p>Paper selection for literature review using PRISMA [<a href="#B7-information-15-00484" class="html-bibr">7</a>,<a href="#B8-information-15-00484" class="html-bibr">8</a>].</p>
Full article ">Figure 2
<p>Total values received by ransomware attackers in the last 5 years [<a href="#B10-information-15-00484" class="html-bibr">10</a>].</p>
Full article ">Figure 3
<p>How the ransomware attacks work [<a href="#B11-information-15-00484" class="html-bibr">11</a>].</p>
Full article ">Figure 4
<p>Types of ransomware attacks.</p>
Full article ">Figure 5
<p>Artificial intelligence techniques [<a href="#B35-information-15-00484" class="html-bibr">35</a>].</p>
Full article ">Figure 6
<p>Machine learning: detection algorithm.</p>
Full article ">
15 pages, 5521 KiB  
Article
A Historical Handwritten French Manuscripts Text Detection Method in Full Pages
by Rui Sang, Shili Zhao, Yan Meng, Mingxian Zhang, Xuefei Li, Huijie Xia and Ran Zhao
Information 2024, 15(8), 483; https://doi.org/10.3390/info15080483 - 14 Aug 2024
Viewed by 798
Abstract
Historical handwritten manuscripts pose challenges to automated recognition techniques due to their unique handwriting styles and cultural backgrounds. In order to solve the problems of complex text word misdetection, omission, and insufficient detection of wide-pitch curved text, this study proposes a high-precision text [...] Read more.
Historical handwritten manuscripts pose challenges to automated recognition techniques due to their unique handwriting styles and cultural backgrounds. In order to solve the problems of complex text word misdetection, omission, and insufficient detection of wide-pitch curved text, this study proposes a high-precision text detection method based on improved YOLOv8s. Firstly, the Swin Transformer is used to replace C2f at the end of the backbone network to solve the shortcomings of fine-grained information loss and insufficient learning features in text word detection. Secondly, the Dysample (Dynamic Upsampling Operator) method is used to retain more detailed features of the target and overcome the shortcomings of information loss in traditional upsampling to realize the text detection task for dense targets. Then, the LSK (Large Selective Kernel) module is added to the detection head to dynamically adjust the feature extraction receptive field, which solves the cases of extreme aspect ratio words, unfocused small text, and complex shape text in text detection. Finally, in order to overcome the CIOU (Complete Intersection Over Union) loss in target box regression with unclear aspect ratio, insensitive to size change, and insufficient correlation between target coordinates, Gaussian Wasserstein Distance (GWD) is introduced to modify the regression loss to measure the similarity between the two bounding boxes in order to obtain high-quality bounding boxes. Compared with the State-of-the-Art methods, the proposed method achieves optimal performance in text detection, with the precision and [email protected] reaching 86.3% and 82.4%, which are 8.1% and 6.7% higher than the original method, respectively. The advancement of each module is verified by ablation experiments. The experimental results show that the method proposed in this study can effectively realize complex text detection and provide a powerful technical means for historical manuscript reproduction. Full article
Show Figures

Figure 1

Figure 1
<p>An example of a historical French handwritten manuscript (Source: gallica.bnf.fr/Bibliothèque nationale de France. Département des manuscrits. Français 17238, <a href="https://gallica.bnf.fr/ark:/12148/btv1b90616522/f5.item.r=Traductions%20et%20extraits%20de%20livres%20chinois" target="_blank">https://gallica.bnf.fr/ark:/12148/btv1b90616522/f5.item.r=Traductions%20et%20extraits%20de%20livres%20chinois</a>) (accessed on 18 July 2024).</p>
Full article ">Figure 2
<p>Comparison of super-resolution effects of different models on blurred French text (Source: gallica.bnf.fr/Bibliothèque nationale de France. Département des manuscrits. Français 17239, <a href="https://gallica.bnf.fr/ark:/12148/btv1b90615371/f49.item.r=francais%2017239" target="_blank">https://gallica.bnf.fr/ark:/12148/btv1b90615371/f49.item.r=francais%2017239</a>) (accessed on 18 July 2024).</p>
Full article ">Figure 3
<p>Improved YOLOv8s network architecture (the red box shows the improved module).</p>
Full article ">Figure 4
<p>Swin Transformer network module structure schematic diagram.</p>
Full article ">Figure 5
<p>Dysample module structure schematic diagram.</p>
Full article ">Figure 6
<p>LSK module structure schematic diagram.</p>
Full article ">Figure 7
<p>Visualized detection results of different text detection models (Source: gallica.bnf.fr/Bibliothèque nationale de France. Département des manuscrits. Français 17239, <a href="https://gallica.bnf.fr/ark:/12148/btv1b90615371/f16.item.r=francais%2017239" target="_blank">https://gallica.bnf.fr/ark:/12148/btv1b90615371/f16.item.r=francais%2017239</a>) (accessed on 18 July 2024).</p>
Full article ">Figure 8
<p>Detection results of original and enhanced image visualization (Source: gallica.bnf.fr/Bibliothèque nationale de France. Département des manuscrits. Français 17239, <a href="https://gallica.bnf.fr/ark:/12148/btv1b90615371/f16.item.r=francais%2017239" target="_blank">https://gallica.bnf.fr/ark:/12148/btv1b90615371/f16.item.r=francais%2017239</a>) (accessed on 18 July 2024).</p>
Full article ">
14 pages, 1357 KiB  
Article
Combined-Step-Size Affine Projection Andrew’s Sine Estimate for Robust Adaptive Filtering
by Yuhao Wan and Wenyuan Wang
Information 2024, 15(8), 482; https://doi.org/10.3390/info15080482 - 14 Aug 2024
Viewed by 715
Abstract
Recently, an affine-projection-like M-estimate (APLM) algorithm has gained popularity for its ability to effectively handle impulsive background disturbances. Nevertheless, the APLM algorithm’s performance is negatively affected by steady-state misalignment. To address this issue while maintaining equivalent computational complexity, a robust cost function based [...] Read more.
Recently, an affine-projection-like M-estimate (APLM) algorithm has gained popularity for its ability to effectively handle impulsive background disturbances. Nevertheless, the APLM algorithm’s performance is negatively affected by steady-state misalignment. To address this issue while maintaining equivalent computational complexity, a robust cost function based on the Andrew’s sine estimator (ASE) is introduced and a corresponding affine-projection Andrew’s sine estimator (APASE) algorithm is proposed in this paper. To further enhance the tracking capability and accelerate the convergence rate, we develop the combined-step-size APASE (CSS-APASE) algorithm using a combination of two different step sizes. A series of simulation studies are conducted in system identification and echo cancellation scenarios, which confirms that the proposed algorithms can attain reduced misalignment compared to other currently available algorithms in cases of impulsive noise. Meanwhile, we also establish a bound on the learning rate to ensure the stability of the proposed algorithms. Full article
(This article belongs to the Special Issue Signal Processing and Machine Learning, 2nd Edition)
Show Figures

Figure 1

Figure 1
<p>Comparison of different c values in system identification: (<b>a</b>) NMSD; (<b>b</b>) EMSE.</p>
Full article ">Figure 2
<p>Comparison of APA, APSA, APLM, APASE, and CSS-APASE algorithms in system identification with white Gaussian noise: (<b>a</b>) NMSD; (<b>b</b>) EMSE.</p>
Full article ">Figure 3
<p>Comparison of APA, APSA, APLM, APASE, and CSS-APASE algorithms with both impulsive noise and white Gaussian noise: (<b>a</b>) NMSD; (<b>b</b>) EMSE.</p>
Full article ">Figure 4
<p>Comparison of APA, APSA, APLM, APASE, and CSS-APASE algorithms with both impulsive noise and uniformly distributed noise: (<b>a</b>) NMSD; (<b>b</b>) EMSE.</p>
Full article ">Figure 5
<p>(<b>a</b>) Impulse response of echo path. (<b>b</b>) Speech input signal.</p>
Full article ">Figure 6
<p>Comparison of the NMSD (dB) of APA, APSA, APLM, APASE, and CSS-APASE algorithms in echo cancellation for speech input signal.</p>
Full article ">Figure 7
<p>Comparison of the NMSD (dB) of APA, APSA, APLM, APASE, and CSS-APASE algorithms in echo cancellation for AR(1)-related signal.</p>
Full article ">
10 pages, 594 KiB  
Article
Performance Management Decision-Making Model: Case Study on Foreign Language Learning Curriculums
by Kuen-Suan Chen, Chun-Min Yu, Chun-Hung Yu and Yen-Po Chen
Information 2024, 15(8), 481; https://doi.org/10.3390/info15080481 - 14 Aug 2024
Viewed by 773
Abstract
Foreign language learning courses can be regarded as a service operation system, and a complete foreign language learning course performance evaluation model can help improve the effectiveness of student learning. The performance evaluation matrix (PEM) is an excellent tool for evaluation and resource [...] Read more.
Foreign language learning courses can be regarded as a service operation system, and a complete foreign language learning course performance evaluation model can help improve the effectiveness of student learning. The performance evaluation matrix (PEM) is an excellent tool for evaluation and resource management decision making, and the administrator uses the satisfaction and the importance indices to establish evaluation coordinate points based on the rules of statistical testing. The coordinate points of all service items are plotted into the PEM to grasp the full picture and to make decisions on what to improve or to consider resource transfers so as to elevate the overall satisfaction of the entire service. However, plotting all the coordinate points on the PEM can only be performed by programming, which will lead to limitations in practice. Therefore, instead of the above evaluation rules, this article uses the confidence intervals of decision-making indicators to form a validity evaluation table, to decide which teaching service items should be improved, maintained, or transferred to improve the satisfaction of the entire service system. This form of performance evaluation can be completed using any commonly used word-processing software, so it is easy to apply and promote. Finally, this article provides an applied example to illustrate the proposed method. Full article
(This article belongs to the Special Issue New Applications in Multiple Criteria Decision Analysis II)
Show Figures

Figure 1

Figure 1
<p>Performance evaluation matrix for foreign language teaching.</p>
Full article ">
19 pages, 1196 KiB  
Article
AI-Driven QoS-Aware Scheduling for Serverless Video Analytics at the Edge
by Dimitrios Giagkos, Achilleas Tzenetopoulos, Dimosthenis Masouros, Sotirios Xydis, Francky Catthoor and Dimitrios Soudris
Information 2024, 15(8), 480; https://doi.org/10.3390/info15080480 - 13 Aug 2024
Viewed by 1324
Abstract
Today, video analytics are becoming extremely popular due to the increasing need for extracting valuable information from videos available in public sharing services through camera-driven streams in IoT environments. To avoid data communication overheads, a common practice is to have computation close to [...] Read more.
Today, video analytics are becoming extremely popular due to the increasing need for extracting valuable information from videos available in public sharing services through camera-driven streams in IoT environments. To avoid data communication overheads, a common practice is to have computation close to the data source rather than Cloud offloading. Typically, video analytics are organized as separate tasks, each with different resource requirements (e.g., computational- vs. memory-intensive tasks). The serverless computing paradigm forms a promising approach for mapping such types of applications, enabling fine-grained deployment and management in a per-function, and per-device manner. However, there is a tradeoff between QoS adherence and resource efficiency. Performance variability due to function co-location and prevalent resource heterogeneity make maintaining QoS challenging. At the same time, resource efficiency is essential to avoid waste, such as unnecessary power consumption and CPU reservation. In this paper, we present Darly, a QoS-, interference- and heterogeneity-aware Deep Reinforcement Learning-based Scheduler for serverless video analytics deployments on top of distributed Edge nodes. The proposed framework incorporates a DRL agent that exploits performance counters to identify the levels of interference and the degree of heterogeneity in the underlying Edge infrastructure. It combines this information along with user-defined QoS requirements to improve resource allocations by deciding the placement, migration, or horizontal scaling of serverless functions. We evaluate Darly on a typical Edge cluster with a real-world workflow composed of commonly used serverless video analytics functions and show that our approach achieves efficient scheduling of the deployed functions by satisfying multiple QoS requirements for up to 91.6% (Profile-based) of the total requests under dynamic conditions. Full article
Show Figures

Figure 1

Figure 1
<p>Video analytics workflow.</p>
Full article ">Figure 2
<p>Impact of interference on serverless functions.</p>
Full article ">Figure 3
<p>Impact of heterogeneity on serverless functions.</p>
Full article ">Figure 4
<p>Framework overview.</p>
Full article ">Figure 5
<p>Comparative evaluation of different schedulers during training of the DRL.</p>
Full article ">Figure 6
<p>Interference level, QoS Quotient, and decision-making of the DRL agent under different scheduling policies.</p>
Full article ">
Previous Issue
Back to TopTop