[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
Next Article in Journal
Task Offloading with LLM-Enhanced Multi-Agent Reinforcement Learning in UAV-Assisted Edge Computing
Previous Article in Journal
Fibre Optic Method for Detecting Oil Fluorescence in Marine Sediments
Previous Article in Special Issue
Semantic Scene Completion in Autonomous Driving: A Two-Stream Multi-Vehicle Collaboration Approach
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Driving-Related Cognitive Abilities Prediction Based on Transformer’s Multimodal Fusion Framework

Faculty of Information Science and Technology, Beijing University of Technology, Beijing 100124, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(1), 174; https://doi.org/10.3390/s25010174
Submission received: 18 November 2024 / Revised: 25 December 2024 / Accepted: 27 December 2024 / Published: 31 December 2024
(This article belongs to the Special Issue Intelligent Sensing and Computing for Smart and Autonomous Vehicles)

Abstract

:
With the increasing complexity of urban roads and rising traffic flow, traffic safety has become a critical societal concern. Current research primarily addresses drivers’ attention, reaction speed, and perceptual abilities, but comprehensive assessments of cognitive abilities in complex traffic environments are lacking. This study, grounded in cognitive science and neuropsychology, identifies and quantitatively evaluates ten cognitive components related to driving decision-making, execution, and psychological states by analyzing video footage of drivers’ actions. Physiological data (e.g., Electrocardiogram (ECG), Electrodermal Activity (EDA)) and non-physiological data (e.g., Eye Tracking (ET)) are collected from simulated driving scenarios. A dual-branch Transformer network model is developed to extract temporal features from multimodal data, integrating these features through a weight adjustment strategy to predict driving-related cognitive abilities. Experiments on a multimodal driving dataset from the Computational Physiology Laboratory at the University of Houston, USA, yield an Accuracy (ACC) of 0.9908 and an F1-score of 0.9832, confirming the model’s effectiveness. This method effectively combines scale measurements and driving behavior under secondary tasks to assess cognitive abilities, providing a novel approach for driving risk assessment and traffic safety strategy development.

1. Introduction

In modern society, driving has become an integral part of daily life. According to a survey, 94% of traffic accidents are caused by driver-related factors [1]. Beyond mastering vehicle operation skills, drivers must also possess a wide range of complex cognitive abilities to handle various traffic situations and emergencies [2,3,4,5]. Therefore, gaining a comprehensive understanding of the cognitive abilities involved in driving has significant theoretical and practical value. This knowledge is critical for assessing whether drivers are capable of safely operating vehicles, helping them understand their driving performance, and enabling governments to conduct more scientific medical examinations.
In recent years, cognitive science and neuropsychology have been introduced into driving research [6,7,8]. By studying cognitive processes such as perception, response, and memory in driving, researchers like Moran et al. [9] have proposed a theoretical framework that divides driving behavior into several interrelated cognitive sub-capacities, including attention allocation, rapid response, processing speed, working memory (WM) capacity, reaction ability, judgment, anticipatory ability, and perceptual skills. A deficiency in any of these abilities can lead to serious traffic accidents and heightened driving risks. Young drivers (aged 17–25) and older drivers (aged 60 and above) are particularly considered high-risk groups [10]. Younger drivers, especially those newly licensed, may encounter unfamiliar driving challenges and exhibit impulsivity, leading to relatively weaker cognitive driving abilities [11,12]. In older adults, cognitive abilities tend to decline with age [13,14].
Most current studies on driving-related cognitive abilities assessments focus on specific cognitive components such as attention [15,16,17,18], reaction speed [19,20,21], working memory (WM) capacity [22,23,24], or perceptual ability [9,25,26]. For example, Zatmeh-Kanj and Toledo [16] proposed the Gipps’ Model (GM) for car-following based on vehicle data, a classical model used to assess driving attention. In this study, participants were tasked with following a lead vehicle that randomly changed speeds while maintaining a consistent distance. Data including average speed, speed variability, acceleration, deceleration times, and distance to the lead vehicle were used as inputs for the GM model to evaluate driver attention levels. Similarly, Chen et al. [17] analyzed behaviors that distract attention while driving, examining four secondary tasks: answering math questions, solving problems, texting, and normal driving. By observing changes in drivers’ behavior and physiological data during these tasks, they developed a neural architecture search network to perform detailed analyses of attention-diverting activities. Mohammed et al. [18] proposed a semi-supervised lightweight vision transformer method based on pseudo-labeling, which effectively detects driver distraction behaviors in natural driving environments by utilizing both unlabeled and a small amount of labeled data for model training. This approach achieved an average accuracy of 95.43% on the StateFarm dataset. Stanisław et al. [20] examined reaction time by measuring dynamic vehicle parameters such as response time for the accelerator, brake, and steering inputs, presenting the data as a Time-To-Collision (TTC) function. Their study demonstrated that these parameters effectively evaluated driver reaction ability under various conditions, providing a scientific basis for enhancing driving safety. Additionally, Hajinoroozi et al. [21] integrated Electroencephalogram (EEG) data with reaction times to assess drivers’ states. Their findings suggested that a reaction time (RT) of ≤0.7 s indicates good driving performance, whereas an RT of ≥2.1 s reflects poor performance. EEG signals helped distinguish between drivers with strong and weak driving capabilities. Broadbent et al. [24] found that by setting tasks with varying levels of working memory (WM) demands and employing biometric methods such as Functional Near-Infrared Spectroscopy (FNIRS) and eye-tracking, it is possible to effectively evaluate drivers’ WM capacity and its impact on driving performance. Their findings indicate that WM capacity significantly influences drivers’ reaction times and overall performance, particularly under high-load tasks. In a study analyzing the effect of perceptual abilities on driving, Moran et al. [9] employed newly developed testing methods that included the accuracy of recognizing potential hazards, the reaction time required to anticipate these hazards, and the time needed to take action to avoid them. They recommended incorporating cognitive function assessments into graduated licensing systems to help identify young drivers whose cognitive skills may be insufficient for safe driving. Moghaddam and Ayati [27] found that Type A drivers, characterized by competitiveness and a sense of urgency, are more prone to high-risk driving behaviors and have higher accident rates. In contrast, Type B drivers, who tend to exhibit more relaxed and patient driving styles, are generally safer. Therefore, when developing traffic safety strategies and interventions, personality traits should be considered.
These studies primarily analyze driver distraction, reaction capabilities, and other factors by combining secondary tasks with subjective and objective data measurement methods. However, they focus on individual components of cognitive driving abilities without fully considering the complexity of the driving process and the multifaceted nature of cognitive abilities.
Some subsequent studies have attempted to analyze the effects of driving-related cognitive abilities holistically and comprehensively, for example, Depestele et al. [28] set up various driving scenarios to assess cognitive abilities and conduct driving simulator tests. They found that cognitive capabilities are significantly positively correlated with driving performance, both in elderly and young individuals. Ledger et al. [29] targeted two groups with a higher risk of collisions, namely young and elderly drivers, using various cognitive measurement tools such as the Rey Complex Figure Test (CFT), Trail Making Test (TMT), and Mini-Mental State Examination (MMSE). Through statistical analysis, they observed the relationship between drivers’ cognitive abilities and driving performance. They found that, regardless of age, drivers’ cognitive abilities are a crucial factor in assessing driving performance. Mullen et al. [30] recorded the performance of elderly drivers in driving behaviors such as turning, braking, and accelerating using a driving simulator and statistical analysis methods. Additionally, they employed a series of cognitive measurement scales to assess the cognitive abilities of elderly drivers, including short-term memory, spatial working capacity, processing speed, and attention. The study results indicate that driving-related cognitive abilities are an important basis for assessing the driving performance of elderly drivers. However, Stolwyk et al. [31] found that the correlation between on-road driving performance and standard neuropsychological tests was not significant, especially for healthy individuals without significant cognitive impairment. The studies mentioned above demonstrate that evaluating driving-related cognitive abilities through various scales, task scenarios, and driving simulators to observe driver performance in behaviors such as steering, braking, and acceleration is effective. Regardless of age, there is a significant positive correlation between cognitive abilities and driving performance. However, these studies mainly use statistical methods to analyze data relationships, which perform inadequately when handling high-dimensional data and fail to fully extract common features.
Based on the research surveyed above, it is found that current studies mostly investigate the correlation between a single element of driving-related cognitive abilities and driving performance. Even when some studies attempt to integrate these driving-related cognitive abilities, they have only used scales or have not comprehensively synthesized all aspects of driving-related cognitive abilities. In addition, most of the methods analyzed in these studies use machine learning or statistical models, although machine learning has advantages in speed and efficiency, its performance is heavily dependent on the researcher’s expertise and the selection of data features. In the processing of high-dimensional and complex data, feature selection and data preprocessing require extensive specialized knowledge. Therefore, it is necessary to further utilize advanced modeling algorithms to comprehensively explore the integrated impact of different driving-related cognitive abilities on driving behavior.
In recent years, with the continuous emergence of big data and the improvement of computational capabilities, sensor technology has also been evolving rapidly. Sensors used for driving behavior analysis include physiological and non-physiological measurement sensors. Physiological or neurophysiological measurements include ECG, EEG, EDA, galvanic skin response (GSR), and respiratory rate [32]. Non-physiological measurement sensors include vehicle signals such as speed, acceleration, lane deviation, and braking force, eye trackers measuring eye displacement in the X/Y directions, and depth cameras for capturing facial emotional signals [33,34,35,36]. The continuous advancement of these sensor technologies has led to a more comprehensive acquisition of driving behavior data, which provides a richer source of information for assessing a driver’s cognitive ability to drive. At the same time, this poses a challenge to the processing of such multi-source data. The large amount of multi-source data makes it challenging to obtain high-dimensional features. Therefore, the distributed and hierarchical feature characteristics of deep learning have emerged as effective methods for processing these data and have found wide application in driving behavior assessment. For this purpose, Mou et al. [32] proposed a novel dual-channel feature extraction model CNN (Convolutional Neural Network) -Trans-SA based on CNN and Transformer. This model takes the eye, vehicle, and physiological data as input features to perform fine-grained classification of driver distraction levels, including normal driving, cognitive distraction, emotional distraction, and sensory-motor distraction. The classification ACC of the model reached 99.802%. Arvin et al. [33] employed the concept of volatility by integrating a Data Acquisition System (DAS) with various sensors such as front and in-car cameras, front sensors, and accelerometers to collect vehicle motion, cabin video, and surrounding environment data streams. They utilized a combination of a one-dimensional 1D-Convolutional Neural Network (1D-CNN) and Long Short-Term Memory (LSTM) to build a 1DCNN-LSTM model for processing these data. This model extracts local dependencies and volatility from time-series data to achieve distraction classification of drivers, achieving the model classification ACC of 95.45% in their study. Mou et al. [34] proposed an attention-based framework for multimodal fusion of driver stress detection. The framework utilizes an attention-based CNN and LSTM model to fuse non-invasive data, including eye data, vehicle data, and environmental data. With this model, features can be automatically extracted from each modality and features from different modalities can be weighted through a self-attention mechanism to increase attention to important features. The method achieves the model average classification ACC of 95.5%. Arefnezhad et al. [35] proposed a combined CNN and Recurrent Neural Network (RNN) based method for driver sleepiness detection. The method takes five vehicle-based measurements as inputs to the network, including lateral deviation from the road centerline, lateral acceleration, traverse angular velocity, steering wheel angle, and steering wheel speed. In the RNN layer, LSTM and Gated Recurrent Unit (GRU) are used. By combining with CNN (CNN-LSTM), the method achieved the highest model classification ACC of 96.0%. Vyas et al. [36] proposed a transformer-based end-to-end driver behavior classification framework called Trans-DBC that aims to classify unsafe driving behaviors. The framework enables the analysis of driver behavior by effectively using accelerometer, gyroscope, and GPS multivariate time-series data combined with learning short-term and long-term time dependencies. The results of the study show that the framework can achieve 95.38% classification ACC.
The above research indicates that multimodal data provides rich feature information. To effectively predict driving-related cognitive abilities, it is essential to maximally extract the common features from these data. This requires designing an appropriate network model to reduce the interference caused by different distribution characteristics at the data source level, mapping the features into a high-dimensional space, and maximally integrating the common features across all data to effectively predict driving-related cognitive abilities. Table 1. presents a summary table form of the current state of research related to the assessment of driving behavior mentioned above, which includes the research subjects, the signals used, the assessment methods, and the experimental results.
In summary, although existing research has explored various cognitive factors related to driving performance in different driving scenarios, most studies are limited to the analysis of a single cognitive element, which leads to an incomplete understanding of the factors influencing driving behavior. This study systematically analyzes the components of driving-related cognitive abilities and their interactions with driving behavior from three dimensions: cognitive decision-making, vehicle control, and psychological state. By improving the model, we are able to extract key common features from multimodal sensor data, allowing for a more accurate and comprehensive evaluation of driving-related cognitive abilities. Therefore, this study proposes a method for assessing driving-related cognitive abilities based on multimodal data. This method initially introduces a comprehensive way to quantify driving-related cognitive abilities. By utilizing driving behavior data across different scenarios, it extracts ten fundamental cognitive elements comprising driving-related cognitive abilities: attention allocation, processing speed, working memory capacity, reaction ability, judgment, anticipation, perception, stress resistance, anxiety level, and Type A/B personality traits, further refining the characterization of driving-related cognitive abilities. Concurrently, a predictive method for driving-related cognitive abilities based on multimodal driving data is proposed. This approach aims to identify relationships between sensor data and driver cognitive abilities, utilizing the Transformer self-attention mechanism and multimodal fusion techniques to predict drivers’ cognitive capabilities. The proposed method in this paper contains the following main contributions:
  • Introduction of a Multidimensional Cognitive Ability Evaluation System: This study defines and quantifies cognitive factors affecting driving safety from three dimensions—cognitive decision-making, vehicle control, and psychological state. It extracts and quantifies multiple cognitive ability indicators related to driving, thereby providing a comprehensive assessment of drivers’ cognitive capabilities and addressing the limitations of previous studies that focused solely on single cognitive dimensions.
  • Development of a Multimodal Data Feature Extraction Model: To address the challenge of effectively extracting common features between multimodal data in traditional models, this study proposes a dual-branch Transformer network. The dual-branch structure eliminates interference between input data and effectively learns high-dimensional features of the data. Additionally, for different data sources, a dynamic weight allocation-based feature fusion mechanism is introduced, enhancing the feature representation ability and predictive performance of the final model.

2. Materials and Methods

2.1. The Datasets

This study utilizes the multimodal driving dataset published by Salah Taamneh et al. [37], which comprises 68 participants, including 34 young participants (aged 18–27 years, mean = 22.41 ± 1.7 years) and 34 elderly participants (aged 61–86 years, mean = 69.97 ± 6.6 years). The dataset was collected using a driving simulator manufactured by Realtime Technologies, Inc. (located in Royal Oak, MI, USA). During the simulated driving task, participants were required to drive along a 10.9 km straight section of a four-lane highway with a speed limit of 70 km/h. There were two dedicated lanes in each direction, with participants driving in the right lane (R), while the opposite lane had heavy traffic (more than 12 vehicles per kilometer). The left lane (L) was under construction, and traffic signposts were located on both sides of the right lane (R). During this task, participants had to complete various driving tasks, including secondary and failure-driving tasks, which involved distracted driving scenarios such as answering questions, texting, etc., while also facing challenges like construction zone crossings and speed limit signs. In the failure-driving task phase, a vehicle malfunction caused an unintended acceleration event before the traffic light at an intersection turned green, propelling the car forward and causing a collision with another vehicle entering the intersection. Throughout these simulated driving tasks, a range of sensors recorded both physiological and non-physiological data of the drivers for analysis.
1.
Physiological sensors include the following:
  • Skin galvanic response signals were collected using the Shimmer3 GSR sensor (Shimmer, Dublin, Ireland) with a sampling frequency of 25 Hz to record the driver’s physiological arousal state.
  • Heart rate and respiratory rate were collected using the Zephyr BioHarness 3.0 (Zephyr Technology, Annapolis, Maryland) with a sampling frequency of 1 Hz.
  • The Tau 640 long-wave infrared (LWIR) camera (FLIR Commercial Systems, Goleta, California) with a frame rate of 7.5 fps was used to capture perspiration signals around the nose area.
2.
Non-physiological sensors include the following:
  • Eye movement parameters, such as gaze position and pupil diameter, were recorded using the faceLAB system (Seeing Machines, Canberra, Australia) with a sampling frequency of 25 Hz to analyze the driver’s attention state.
  • Data on vehicle speed, acceleration, braking, steering, lane deviation, and lane position were collected using the driving simulator manufactured by Realtime Technologies, Inc. (Royal Oak, Michigan) with a sampling frequency of 58.8 Hz to analyze vehicle control state.
  • Emotional state data were recorded using the HD Pro Webcam C920 (Logitech, Newark, CA, USA) monochrome zoom camera with a frame rate of 15 fps. The data were pre-processed and analyzed to assess changes in the driver’s emotional state.

2.2. General Structure of the Framework

In this study, a Transformer-based Dual-branch multimodal feature fusion network model is proposed for realizing the prediction of a driver’s driving-related cognitive abilities. Figure 1. shows the general technical framework of this paper.
In Figure 1, the shaded blocks are the main innovations of this study. In this study, the components of driving-related cognitive ability are extracted and quantified by the Driving-Related Cognitive Abilities Extraction Module, and the driving-related cognitive ability labels are generated by the Driving-Related Cognitive Abilities Label Generation Module by assigning weights to each component to generate weighted scores. Meanwhile, this study proposes a Dual-branch transformer network in the Driving-Related Cognitive Abilities Prediction Network Model, based on the VIT network, the input features are sent to different branches for feature extraction, and the extracted features are fused with the AFF network, which enhances the feature extraction capability of the model. The overall technical framework of this paper consists of the following four main modules.
  • Data Preprocessing Module: Due to the complexity and diversity of the data contained in multimodal driving data, the raw data need to be preprocessed before feature extraction. This module initially removes unstable data recorded before the start of the experiment and addresses any missing data. It also eliminates motion artifacts caused by actions such as blinking and selects an appropriate sliding window to provide sufficient samples for training.
  • Driving-Related Cognitive Abilities Extraction Module: This module is designed to assess the performance of drivers during various secondary tasks and malfunction handling tasks. It extracts and quantitatively represents the components of driving cognitive-related abilities based on this performance.
  • Driving-Related Cognitive Abilities Label Generation Module: This module assigns weights to the quantified components of driving-related cognitive abilities based on their importance and performs a weighted sum. Labels for driving-related cognitive abilities are then generated based on the weighted scores.
  • Driving-Related Cognitive Abilities Prediction Network Model: After the above modules, clean driving data and labels for driving-related cognitive abilities were obtained. At this point, an appropriate network model is required to extract and fuse features from the processed data, enabling precise prediction of driving-related cognitive abilities.
In this paper, the above four modules are described in detail in Section 2.3, Section 2.4 and Section 2.5.

2.3. Data Preprocessing Module

Before performing the prediction task, the 15-dimensional multimodal driving data needs to be preprocessed to eliminate noise, missing values, and outliers from the data to ensure ACC and completeness. Consider the input X is an n-dimensional time-series signal, and using Min–Max normalization, X is normalized between 0 and 1 to obtain X . The normalization formula is as follows:
X = X X m i n X m a x X m i n
In this study, a sliding window approach with overlap was used to partition the normalized multimodal driving data into appropriate sizes. These sliding window sizes and step sizes were chosen based on the results of previous studies on multimodal driving behavior analysis [17,21,32,33]. Time series of fixed length and width were generated through the sliding window approach. These sequences were designed to overlap between neighboring windows to maintain continuity in the window sequences and provide enough samples for training, enabling the network to better acquire features. The sequences generated through the windowing method are upscaled into uniformly sized 2D image data, which are aligned with the original labels and used as data for model training.
To address the issue of missing data in multimodal driving datasets, particularly the asynchrony observed among multiple sensors during the initial recording phase (i.e., the first 1–10 s), this study initially excluded these asynchronous data records. Subsequently, linear interpolation was employed to impute the missing values for the synchronized data, thereby ensuring the dataset’s high quality and completeness. This ensures that the data can be effectively utilized for network training.

2.4. Driving-Related Cognitive Abilities Label Setting

Driving is a process that involves the brain’s rapid recognition of external information, decision-making, and vehicle control through actions such as steering or braking. This process is also influenced by the driver’s psychological state. To deeply investigate and quantify the key factors affecting the driving process, this study designed a set of driving-related cognitive abilities assessment methods, aiming to quantitatively assess the performance of drivers in various aspects of thinking decision-making, vehicle maneuvering, and changes in psychological state during driving activities.
Driving-related cognitive abilities refer to the driver’s ability to process information and make decisions, precisely control the vehicle, and effectively regulate psychological states during the driving process. Driving-related cognitive abilities can reflect the driver’s reception and processing of information, rapid execution of decision-making, and changes in psychological state. This study comprehensively examined the changes in driving behavior and classified driving-related cognitive abilities into three dimensions based on the different needs of driving tasks: thinking decision-making, vehicle manipulation, and psychological regulation, which contain ten basic elements that make up driving-related cognitive abilities. The specific details are as follows:
  • Cognitive decision-making dimension: This explores how drivers process received traffic information and quickly make decisions based on this information. It includes the assessment of drivers’ attention allocation, working memory, and anticipation abilities.
  • Vehicle control dimension: This directly reflects how drivers translate decisions into vehicle operations, covering assessments of drivers’ processing speed, reaction capability, judgment, and perceptual skills.
  • Psychological regulation dimension: This examines how drivers influence decision-making and action execution through the regulation of their psychological state, assessing drivers’ stress resilience, anxiety levels, and Type A/B personality traits.
To quantitatively assess these components of driving-related cognitive abilities, this study integrates subjective measures (e.g., questionnaires) with objective measures (e.g., behavioral analysis). Subjective measures reflect the self-assessed abilities of the driver, while objective measures provide data on the actual driving performance of the driver. This study describes in detail how to extract and quantify the components of driving-related cognitive abilities based on the video data in Section 2.4.1. In Section 2.4.2, a systematic assessment method is designed to assign weights to the quantified indicators and weight the scores, and the driving-related cognitive ability is finally graded as “Insufficient”, “Fair”, “Good “ and “Excellent”.

2.4.1. Driving-Related Cognitive Abilities Extraction Module

This study analyzes the performance of drivers in various tasks recorded in driving videos, observing their performance in tasks that require focused attention, rapid response, accurate decision-making, and handling complex situations to extract the components of driving-related cognitive abilities. In driving cognitive ability assessment, elements such as attention allocation and perception ability do not exist in isolation, and there may be differences in cognitive abilities involved in normal driving and dangerous situations [38]. Additionally, these driving-related cognitive abilities have varying priorities in different task scenarios, emphasizing different aspects. The key cognitive abilities involved in driving can be extracted through different task scenarios. For example, multiple studies [6,39,40] have pointed out that while both reaction ability and processing speed involve reaction time, reaction ability mainly refers to the driver’s braking time in emergencies, while processing speed refers to the driver’s information processing efficiency when faced with complex road conditions. Perception ability involves the driver’s awareness of traffic signs, vehicles, and the relative position of the surrounding environment. These reflect different cognitive processes in emergencies and when handling secondary tasks. The specific methods for extracting the ten elements of driving-related cognitive abilities within the three dimensions are presented in Table 2.
Table 2 lists the foundational data sources for each driving-related cognitive ability. The dimension features corresponding to each data source are extracted through specific driving task scenarios. For example, reaction ability is evaluated through a simulated emergency braking task, while processing speed is assessed by the completion time of secondary tasks in complex road conditions. By analyzing task performance across different scenarios, we effectively extract the driver’s performance across various cognitive ability dimensions. Due to the difference in the experimental setup, members of group A in the faulty driving task needed to complete additional tasks, thus creating a bias in the calculation of reaction time. To eliminate this effect, the study used the method of time compensation [20]. It standardized the reaction time data for the two subject groups. This was achieved by using the difference between the average reaction times of groups A and B as compensation. The method of reaction time compensation is calculated by Equation (2).
C R T = ( R T A R T B ) N
where C R T denotes compensated reaction time, R T A denotes the group a reaction time, R T B denotes the group’s reaction time, and N denotes the total number of people.
To accurately quantify driving-related cognitive abilities, this study follows these steps:
  • First, directly quantifiable indicators such as reaction times and scale scores are kept in their original values for standardization.
  • Second, qualitative behavioral performances, such as the degree of collision, are numerically processed. For example, ’no collision’ is marked as 0, and ’collision’ is marked as 1. This step completes the numerical transformation of all data.
  • Finally, the formula Z = ( X μ ) σ is applied to standardize the numerically transformed data.
where Z denotes normalization, X denotes indicator value, μ denotes sample mean, and σ denotes sample standard deviation.
These steps produce quantified driving-related cognitive abilities, which are then used for generating labels in Section 2.4.2.

2.4.2. Driving-Related Cognitive Abilities Label Generation Module

To map driving-related cognitive abilities to different levels, the data were first processed using the G-Rules-IQR (Gaussian-Rules Interquartile Range, abbreviated as G-Rules-IQR) method [41] to adjust the distribution closer to a normal distribution. Then, based on the first quartile (Q1), the second quartile (Q2, median), and the third quartile (Q3) of the data, all scores were mapped into four ranges. These four ranges represent different levels of driving abilities: Excellent (above Q3), Good (between Q2 and Q3), Fair (between Q1 and Q2), and Insufficient (below Q1). Converting the quantified driving-related cognitive abilities into a driving-related cognitive abilities label may be accomplished by steps 1–3.
  • Analysis of the Importance of Components of Driving-Related Cognitive Abilities: Anstey et al. [42] found that thought decision-making directly reflects the driver’s response to environmental changes and decision-making efficiency, hence it is of the highest importance. Vehicle control reflects the driver’s actual execution ability in controlling the vehicle and responding to road conditions, making it the second most important. Psychological adjustment, while having a non-negligible impact on driving safety, has the least influence compared to thought decision-making and vehicle control abilities that directly determine driving response and vehicle control, thus it is of the lowest importance.
  • Generation of Weighted Scores for Components of Driving-Related Cognitive Abilities: Based on step 1, the importance of three dimensions of driving-related cognitive abilities from high to low are: Cognitive decision-making ( d i m . 1 ), vehicle control ( d i m . 2 ), and psychological regulation ( d i m . 3 ). The weights for each cognitive ability dimension were determined using the Principal Component Analysis (PCA) method [43]. Specifically, PCA was applied to the indicators extracted in Table 2 to identify their linear combination coefficients in each principal component. Based on the variance contribution rate of each principal component, the weights W 1 , W 2 , W 3 were calculated and assigned to d i m . 1 , d i m . 2 , and d i m . 3 , respectively. The driving-related cognitive abilities score ( S c o r e for short) is obtained through weighting. W i and S c o r e can be calculated using Equations (3) and (4). The distribution of driving-related cognitive abilities weights is shown in Table 3.
W i = k = 1 K   L i k V k
S c o r e = i = 1 3   W i d i m . i
where L i k denotes the loading of the i th variable on the k th principal component, V k denotes the ratio of variance explained by the k th principal component, and K denotes the number of principal components selected.
Table 3. Driving-related cognitive abilities weight distribution table.
Table 3. Driving-related cognitive abilities weight distribution table.
Driving-Related Cognitive Abilities Dimensions Importance   Weight   W i Weighting
Thinking Decision Making W 1 0.3898
Vehicle Handling W 2 0.3182
Mental State W 3 0.2920
  • Generation of Driving Cognitive-Related Ability Labels: Driving-related cognitive abilities labels are divided into four levels, 0–3, representing four levels of driving-related cognitive abilities: Insufficient, Fair, Good, and Excellent. These four levels are determined based on the S c o r e derived from Equation (4). The calculation of the S c o r e incorporates the weighted values of ten components of driving-related cognitive abilities, reflecting the comprehensive level of driving-related cognitive abilities. Since these components of driving-related cognitive abilities have different distribution characteristics in terms of centrality and dispersion, the weighted S c o r e results in a multivariate composite distribution. To effectively categorize Scores into levels, this study employs the G-Rules-IQR method. This method does not rely on the normal distribution properties of the data and can determine the score thresholds Q i for “Score” by calculating the interquartile range (IQR). This method maps the continuous “Score” to four discrete intervals {[ S c o r e m i n , Q 1 ), [ Q 1 , Q 2 ), [ Q 2 , Q 3 ), [ Q 3 , S c o r e m a x ]}, achieving the classification of driving-related cognitive abilities levels and thus generating driving-related cognitive abilities labels. The Q i can be calculated using Equation (5).
Q i = μ + σ × z p i   ( i = 1,2 , 3 )
where Q i corresponds to the index of the quartile ( Q 1 denotes the first quartile, Q 2 denotes the second quartile, Q 3 denotes the third quartile), μ denotes the mean of the sample S c o r e , σ denotes the standard deviation of the sample S c o r e , and z p i denotes the different percentiles of S c o r e .
The mapping relationship between the driving-related cognitive abilities labels and S c o r e is calculated by Equation (6).
M s c o r e = 0 ,                   Score m i n S c o r e < Q 1 1 ,                   Q 1 S c o r e < Q 2 2 ,                   Q 2 S c o r e < Q 3 3 ,                   Q 3 S c o r e Score m a x
where 0–3 denotes the driving-related cognitive abilities labels corresponding to the four levels of driving-related cognitive abilities: Insufficient, Fair, Good, and Excellent, respectively. Score m i n and Score m a x denotes the minimum and maximum values, respectively, and Q 1 Q 3 denotes the S c o r e thresholds.
Driving-related cognitive abilities labels were obtained through the processing of the above two modules. The subjects were regrouped according to the labels and the labeled data were fed into the Section 2.5 Driving-related cognitive abilities Prediction Network Module for final driving-related cognitive abilities prediction.

2.5. Driving-Related Cognitive Abilities Prediction Network Model

In the field of driving behavior analysis, effectively extracting and fusing multimodal data from different sensors presents a challenge. The common practice in current research is to directly concatenate the time-series data of different modalities, such as respiratory rate, eye movement coordinates, and vehicle dynamics, without considering the inherent differences in feature expression and diversity among the input data. This approach can easily lead to confusion of information and loss of key features. Furthermore, when dealing with complex driving behavior data, traditional data processing and feature extraction methods based on rules or manually designed features struggle to effectively capture the high-dimensional nonlinear characteristics of the data, thereby affecting the ACC of predictions.
To address the above problems, this study proposes an improved Dual-branch Transformer network structure. It aims to effectively extract and fuse temporal features from heterogeneous data from multiple sources, such as physiological and non-physiological, to adapt to the needs of driving-related cognitive abilities prediction. The structure addresses the differences in the feature expression and feature diversity of multi-source heterogeneous data by employing two parallel encoders, each dedicated to the extraction of physiological and non-physiological features, respectively. This approach resolves the issue of consistency in the feature expression of input data. Moreover, to further weigh the relative importance of the features extracted by physiological and non-physiological encoders, the Attentional Feature Fusion (AFF) algorithm [44] is introduced in the feature fusion stage of this study. This algorithm optimizes the fusion of physiological and non-physiological features through a dynamic weight allocation mechanism and precisely adjusts the relative contribution of these two features in the information representation process, thereby effectively enhancing the ACC of the model in predicting driving-related cognitive abilities. The network model structure proposed in this study is illustrated in Figure 2 and is based on the ViT [45] design. It comprises three main components: a Multi-source Heterogeneous Feature Conversion Module, a Dual-branch Transformer Encoder Module, and an AFF module based on dynamic weight assignment.
  • Multi-source Heterogeneous Feature Conversion Module: This module is responsible for separating physiological data from non-physiological data from multi-modal driving time-series data, and converting them from the original 2D time-series data to 2D image data, which meets the needs of the dual-branch Transformer encoder module for feature extraction from the input data.
  • Dual-branch Transformer Encoder Module: This module adopts an improved dual-branch structure, which allows physiological and non-physiological data to be feature extracted by independent transformer encoders. This dual-branch structure allows the physiological branch to focus on capturing changes in the detailed features of the physiological signals in the short term, and the non-physiological branch to focus on capturing changes in the long-term features such as vehicle operating modes. A self-attention mechanism is used within each branch to optimize the capture of dynamic time-series data features and enhance the feature extraction capability of the model for multimodal driving time-series data.
  • AFF module based on dynamic weight assignment: after physiological and non-physiological features are extracted by the Dual-branch Transformer encoder module, the attention weights of global and local feature channels are extracted by the AFF module using global pooling and point-by-point convolution, where global pooling is in charge of capturing the global feature information and point-by-point convolution is used for extracting and strengthening local feature details. Finally, the performance of the prediction model is further optimized by feature fusion.

2.5.1. Multi-Source Heterogeneous Feature Conversion Module

This module processes the input multimodal driving time-series data, which are categorized into physiological and non-physiological data according to their physiological properties. The input data are in the form of X R H × W , where H and W represent the height and width of the data, respectively. By splitting and reshaping operations, X is split into physiological data X p R H × W × C and non-physiological data X n p R H × W × C , where C is the number of channels. The processed data are converted into 2D image format for feature extraction by the Dual-branch Transformer encoder module.
In order to adequately represent the temporal characteristics of the time-series data, the sliding window method is used to generate the set of temporal windows D. Each temporal window X i R H × W ( 1 i m ) denotes the i th window. For physiological data X p there are timing windows D p denoted as:
D p = X 1 P , y 1 p , X 2 P , y 2 p , , X i P , y i p , 1 < i < m
where X i P denotes the i th window, and y i p is the corresponding category label for the temporal window. The set of temporal windows D p contains m windows and all temporal windows will be used as model inputs for further feature extraction.

2.5.2. Dual-Branch Transformer Encoder Module

For nonlinear time-series feature extraction of multimodal driving time-series data, this study proposes a Dual-branch Transformer encoder module, which adopts two parallel Transformer encoder structures for feature extraction of physiological data X p and non-physiological data X n p processed by a multi-source heterogeneous feature transformation module, respectively, so that the module not only focuses effectively on the changes of the detailed features of the physiological parameters, such as heart rate, EDA, and other physiological parameters but also captures the changes of the long-term dynamic features of the non-physiological data regarding the operating modes of the vehicle.
First, the physiological data D p undergoes a linear projection layer, which maps the raw data containing four dimensions, including skin EDA, heart rate, respiratory rate, and nasal sweating signals, into a high-dimensional feature space, effectively capturing the complex high-dimensional features of the physiological data. Then, to preserve the temporal and label information of each window, we apply position encoding by adding sinusoidal and cosine functions with a fixed frequency. The position encoding introduces an additional positional marker vector E p o s for each temporal window, helping the model recognize the positional information of the sequential data. The label information is incorporated by adding a classification marker vector Z c l s to each window. Finally, the temporal branch feature vector set X p 0 is generated. The specific implementation of the linear projection and position encoding can be referred to in Equation (8).
X P 0 = [ Z c l s p ; d 1 p E p ; d 2 p E p ; ; d m p E p ] + E p o s p
where X p 0 denotes the initial features after linear projection and positional coding process, where E p is the linear projection matrix, d 1 p , d 2 p …, d m p are the temporal slice vectors.
After processing, the initial features X P 0 are fed into a Transformer encoder composed of a multi-head self-attention (MSA) mechanism and a feed-forward network (FFN). Each feed-forward network consists of two fully connected layers, with a dropout layer applied after each layer. The dropout rate is set to 0.2.
The encoded features X p 0 were obtained and next X p 0 is fed into a Transformer encoder consisting of a Multi-head Self-Attention ( M S A ) Mechanism and a Feedforward network ( FFN ). The MSA allows the model to encode a feature at one position taking into account the features at all the other positions in the sequence, enhancing the feature representation by learning the dependencies between the different positions. Each head processes the input independently in MSA, focusing on different parts of the sequence. While the FFN performs further nonlinear transformations on the output of the MSA.
For the encoding process of each Transformer encoder layer L, it can be expressed as:
X p l = M S A ( L N ( X p l 1 ) ) + X p l 1 , l = 1,2 , , L
X p l = F F N ( L N ( X p l ) ) + X p l
where LN denotes the layer normalization operation, which is used to normalize the distribution of input features before self-attention and FFN at each layer, stabilizing the training process and improving the generalization ability of the model. X p l and X p l represent the output features after MSA and FFN, respectively.
Moreover, in the process of physiological feature extraction by the encoder, to solve the problem of gradient vanishing in the deep network training, residual connections are added to ensure that the gradient can be better propagated to the shallower layers during backpropagation, thus helping the network to learn the feature representation more effectively.
For non-physiological data X n p , the process is the same as the physiological branch feature extraction process, whereas the inputs to the non-physiological branch are non-physiological data containing 11 dimensions of eye tracking (gaze point X and Y coordinates, pupil diameters of the left and right eyes), vehicle status (speed, acceleration, braking, lane offset, lane position, distance), and facial mood changes. These data are analyzed by Transformer’s MSA and FFN, and the learning process is enhanced by residual linkage to capture the long-term feature changes in the non-physiological data. After the above process, the final output features of the non-physiological branch are X n p l .
After processing in this module, physiological and non-physiological features can be extracted, providing key feature representations for subsequent feature fusion and prediction tasks.

2.5.3. AFF Module Based on Dynamic Weight Assignment

The input data X p and X n p after going through the Dual-branch Transformer encoder module, the obtained X p l and X n p l features containing complex timing dependencies and rich feature information. To effectively integrate these features for driving-related cognitive abilities prediction, an efficient fusion strategy needs to be designed to effectively integrate the key features contained in X p l and X n p l .
Traditional feature fusion methods, such as simple splicing or weighted averaging, which are simple to design, usually ignore the interrelationships between different features and their relative contribution to the final prediction task. This approach cannot fully utilize all the available information features, especially in scenarios with complex interactions between features, which can easily lead to the omission of critical information or overemphasis on redundant features, which in turn affects the ACC of prediction. Therefore, this study introduces the AFF feature fusion module based on dynamic weight assignment, which optimizes the fusion process of different modal features and enhances the prediction capability of the model. In contrast, AFF adopts an approach based on the attention mechanism to dynamically adjust the contribution of different feature sources, to capture the complementary information between modalities more effectively and retain the key features. The AFF network model is shown in Figure 3.
In Figure 3, where X p l and X n p l denotes the physiological and non-physiological features of the input, these two features go through the global extraction channel and the local extraction channel to obtain the attention weights, and then the obtained attention weights are multiplied with the input features, and finally aggregated to obtain the output Z, ⊗ denotes multiplication and ⊕ denotes addition.
In the AFF module based on dynamic weight assignment, the physiological features X p l and non-physiological features X n p l output from the last layer of the Dual-branch transformer encoder is first split into two streams, the first stream initially integrates the physiological and non-physiological features through the summation operation and passes the summed features through the point-wise convolution layer. In this layer, the feature channels are optimized by linear transformations to adjust the relative importance between features and enhance their nonlinear representations to extract key global information features. The second stream multiplies X p l and X n p l to extract key local information features. It uses a channel attention mechanism to dynamically weigh these features, adjusting their importance based on their relevance to the prediction task. Finally, the global and local features extracted by the two streams of addition and multiplication are fused to generate a comprehensive feature representation. The AFF process can be represented by Equation (10).
F f u s i o n = Attention ( X p l , X n p l ) ( X p l X n p l )
where “ Attention ” denotes the weight assignment function computed by the attention mechanism, ⊕ denotes the fusion operation of the features, and F f u s i o n is the final fused feature representation.
After processing in the above modules, a feature vector incorporating key information F f u s i o n is finally generated, which is used for driving-related cognitive abilities prediction after the FC layer.

3. Experimental Results and Analysis

In this section, the effectiveness of the proposed driving-related cognitive abilities prediction model is evaluated through a series of experiments. Additionally, the methodology of this study is compared with existing machine learning algorithms and deep learning algorithms. Ablation experiments were also conducted to validate the effectiveness of the proposed method in this paper. The computer equipment used for the experiments in this study has the following configurations: an Intel(R) Core (TM) i7-8750H @ 2.20GHz, an NVIDIA GeForce GTX 3090 graphics card, Python 3.6.0 (64-bit), and PyTorch 1.5.0. The version of Tensorflow used for the environment is 2.6.0.
Under the computation power of a GTX 3090 GPU, with a batch size of 128, embedding dimension of 128, 12 layers, and 8 attention heads, the dual-branch Transformer feature fusion network used in this study (Baseline + Dual-Branch + AFF) has patch sizes of 4 × 4 and 11 × 4, with a parameter count of 7.90 MB, a computational cost of 25.72 GFLOPS, and an inference speed of 312 FPS. The baseline ViT model, with a patch size of 15 × 15, has a parameter count of 3.97 MB, a computational cost of 5.23 GFLOPS, and an inference speed of 344 FPS.

3.1. Data Set Segmentation

This study conducted a series of validation experiments using large-scale multimodal driving datasets. This dataset is described in detail in Section 2.1. In this section, the construction of the training–testing dataset will be described.
To further validate the stability and generalization ability of the model, this study employed subject-based cross-validation. In this approach, each subject was treated as an independent test set, with the remaining subjects’ data used for training. This process was repeated, and model evaluations were obtained for each sample. The overall performance of the model was then assessed by averaging the evaluation results across all samples. The specific implementation of this method follows the description in [46].
During the experiments, the parameters were set to 0.9 and 0.999 using the Adam optimizer, and the learning rate was corrected using StepLR, with the step size set to 1, gamma set to 0.5, initial learning rate set to 6e-4, and the batch size to 128. Based on our experience with different window lengths, we considered the length of 60 s, and a continuous window with a step size of 5 s was chosen for sample acquisition eventually generating a total of 31,187 samples, of which the sample sizes for each category were 6649, 7747, 12,134, and 4657, respectively.

3.2. Evaluation Metrics

The performance of the proposed network is evaluated using ACC, Confusion Matrix, and F1-score as evaluation metrics.
The ACC is calculated as follows: where False Positives (FP) and False Negatives (FN) are the negative and positive samples that are incorrectly predicted, while True Positives (TP) and True Negatives (TN) are the positive and negative samples that are correctly predicted, respectively.
A C C = T P + T N T P + T N + F P + F N
The confusion matrix, also known as the error matrix, is a standard format for representing ACC evaluations in the form of an n-row, n-column matrix. Each column of the confusion matrix represents the predicted category, and the total number of each column indicates the number of data predicted to be in that category; each row represents the true category to which the data belongs, and the total number of data in each row indicates the number of data instances in that category. The value in each column represents the number of real data predicted to be in that category. As can be seen, the confusion matrix portrays the combined effect of the model and gives a clear picture of the number of correct or incorrect predictions for each category.
The F1-score is calculated by balancing precision (P) and recall (R) using the following formula.
F 1 = 2 × P × R P + R
Through the evaluation of the model using the above metrics, the F1-score, which takes into account both precision and recall, effectively assesses the model’s performance in handling class imbalance issues. Compared to ACC alone, it provides more detailed performance information. The confusion matrix presents a detailed classification of the model’s predictions, revealing the model’s performance and biases across different categories, thus helping us identify and improve deficiencies in certain categories.

3.3. Comparison Experiment

The comparison models chosen for this study are Support Vector Machine (SVM) [47], Random Forest (RF) [48], LSTM [49], VGG-16 [50], ResNet [51], CNN-LSTM [35]. These models achieve excellent prediction results in the literature related to driving behavior assessment mentioned above. Therefore, these six algorithms are chosen in this study as the comparison algorithms for this experiment.
In this experiment, all comparison algorithms used the same dataset processing methods as in this paper, i.e., the same missing data padding, data normalization, etc.
Table 4 shows the experimental results of all the algorithms using the ACC and F1-score metrics, and Figure 4 shows the experimental results of all the algorithms using the confusion matrix.
From the results of the comparison tests in the table, it can be observed that the model proposed in this study achieves high performance in terms of accuracy (Accuracy) and F1-score. Specifically, the model proposed in this study achieved an ACC rate of 0.9908 and an F1-score of 0.9832. Compared to SVM, RF, LSTM, VGG-16, ResNet, and CNN-LSTM, our model showed improvements in ACC of 36.13%, 13.25%, 33.47%, 32.09%, 4.37%, 3.11%, respectively, and improvements in F1-score of 38.31%, 13.28%, 34.64%, 49.00%, 5.68%, 5.30%, respectively. This demonstrates the superiority of the proposed model in predicting driving-related cognitive abilities. Both the CNN-LSTM and ResNet models achieved ACC rates exceeding 0.90. These models are frequently used for processing time-series data and effectively capture long-term and short-term features in the data. Specifically, CNN excels at extracting local features, while LSTM focuses on the temporal dependencies of sequential data, making CNN-LSTM highly effective in capturing complex temporal features. ResNet, based on deep residual networks, effectively addresses the vanishing gradient problem in deep networks through skip connections, thus demonstrating high predictive performance when processing multimodal data of driving behavior. In contrast, the RF model achieved an ACC and F1 score of 0.8583 and 0.8504, respectively. Although RF performs well with time-series data, it is less effective than our proposed model in predicting the cognitive abilities of drivers in tasks involving complex temporal features. The SVM, LSTM, and VGG-16 models demonstrated relatively lower performance, with ACC rates of 0.6295, 0.6561, and 0.6699, respectively. While SVM excels in handling high-dimensional data, it is limited in fusing multimodal and nonlinear data features. Although LSTM can handle temporal dependency data, it struggles to comprehensively capture spatial features in multimodal data when used alone. VGG-16, being primarily an image recognition model, lacks adaptability in handling multimodal time-series features, resulting in lower predictive ACC. In comparison, the dual-branch transformer network proposed in this study effectively divides multimodal driving data into physiological and non-physiological data and uses a dynamic weight-based feature fusion module to extract common features, thereby achieving effective prediction of driving-related cognitive abilities.
Figure 4 shows the confusion matrix experimental results of each algorithm in the driving-related cognitive abilities prediction task. From the results of the confusion matrix, it can be seen that the Dual-branch multimodal feature fusion network model proposed in this study has the best prediction results, with prediction ACC of 0.98, 0.98, 0.97, and 1.00 for each of the four categories, and more than 0.97 for each of the categories. Categorical performance on a few categories in particular (inadequate driving-related cognitive abilities and excellent driving-related cognitive abilities) also maintained a high level of stability. In contrast, the VGG-16 performed the worst in terms of driving-related cognitive abilities prediction. In particular, the prediction ACC for the categories of poor and excellent driving-related cognitive abilities were only 0.48 and 0.01, respectively. The VGG-16 predicted almost all the samples with excellent driving-related cognitive abilities into the category of average driving-related cognitive abilities, while the prediction ACC for the samples with poor driving-related cognitive abilities was only 0.48. This is mainly due to the imbalance of the distribution of driving-related cognitive abilities in reality, where the majority of the drivers have a moderate level of cognitive ability, with relatively few drivers at the extremes (highest and lowest). However, these minority groups, especially drivers with poor driving-related cognitive abilities, are at higher risk and require special attention. Therefore, this requires the model to have a high predictive power even when the sample is unbalanced. Results indicate that the proposed model, by effectively leveraging the Transformer architecture, can better capture and analyze the nuanced features of such data, thereby significantly improving prediction ACC.

3.4. Ablation Experiments

To verify the effect of the algorithm proposed in this study on the prediction ACC of driving-related cognitive abilities, ablation experiments will be conducted in this study and the results of the ablation experiments are shown in Table 5.
Table 5 shows the results of the ablation experiments performed in this study based on the Baseline. From the results of the ablation experiments, it can be seen that compared to the Baseline, the Dual-branch multimodal feature fusion network model (Baseline + Dual-Branch + AFF) proposed in this study achieved 0.9908 and 0.9832 in the metrics ACC and F1-Score, respectively, marking an improvement of 5.98% and 6.07% over the Baseline model, and an increase of 2.99% and 3.12% relative to using only the Baseline + Dual-Branch structure. This indicates that the dual-branch multimodal feature fusion model proposed in this study effectively enhanced the model’s ability to extract features from different modalities. Additionally, introducing the AFF module further optimized the integration process of features from different modalities, thereby improving the model’s prediction ACC. This improvement is attributed to the proposed dual-branch multimodal feature fusion network’s ability to adapt to the temporal distribution characteristics of different modal input data, reduce interference between different data sources, and fuse the output features of the dual-branch encoders at both local and global levels, effectively extracting the nonlinear time-series features of physiological and non-physiological data, and enhancing the model’s predictive precision.
Figure 5 shows a comparison of the experimental results in terms of Accuracy and F1-score, highlighting the improvements of our model over the baseline and comparison models. The results demonstrate that our model outperforms the baseline model as well as other comparison models, exhibiting superior performance.

4. Conclusions

Driving relies on multiple cognitive abilities. To address the relationships between these abilities and driving performance, as well as to reduce potential information loss or redundancy in multimodal data processing, this study proposes a high-precision framework for predicting driver cognitive abilities. The framework consists of four modules: data preprocessing, cognitive ability extraction, label generation, and a cognitive ability network model. The extraction and label generation modules analyze driving videos to identify ten cognitive components critical for decision-making, vehicle control, and psychological regulation. These components are quantified and weighted to generate cognitive ability labels.
In the cognitive ability network model, this study innovatively applies a Transformer structure to extract both physiological and non-physiological multimodal features, enhancing the model’s capacity for feature extraction and improving prediction accuracy (ACC). The experimental results demonstrate that the proposed framework achieves ACC and F1-scores of 0.9908 and 0.9832, respectively, outperforming RF, SVM, LSTM, ResNet, CNN-LSTM, and VGG-16 models. This prediction model could be applied in driving assessments for licensing, training programs to address individual weaknesses, and personalized driving assistance systems to offer real-time support and warnings. It provides a novel approach to driving risk assessment.
While this study successfully constructs a multimodal-based framework for predicting driving-related cognitive abilities, several limitations remain. The use of video-based analysis for driver behavior introduces the potential for human error, which could affect the objectivity and accuracy of the data. Future work could focus on automating key indicator extraction to enhance reliability. Additionally, factors such as driver experience (e.g., years licensed) may impact ability assessments and should be considered in future studies. The study’s limited focus on younger and older drivers restricts its applicability across all age groups, with insufficient data on middle-aged drivers (aged 27 to 60). Furthermore, the complexity of multimodal data collection may make device usage cumbersome, affecting the framework’s practical feasibility. Future research could focus on optimizing data acquisition devices by selecting a small number of high-precision sensors to reduce unnecessary hardware requirements, thus simplifying the data collection process and enhancing the system’s operability and practicality. Additionally, the introduction of lightweight deep learning models could reduce the number of model parameters and computational complexity, decreasing reliance on computational resources, and improving the system’s real-time responsiveness, computational efficiency, and deployability.

Author Contributions

Conceptualization, Y.L. and B.L.; methodology, Y.L.; software, W.Z.; validation, W.Z.; formal analysis, Y.L.; investigation, Y.L.; resources, W.Z.; data curation, B.L.; writing—original draft preparation, Y.L.; writing—review and editing, Y.L. and W.Z.; visualization, Y.L.; supervision, W.Z.; project administration, B.L.; funding acquisition, W.Z. All authors have read and agreed to the published version of the manuscript.

Funding

Special Project on Scientific Research and Development of Beijing Rehabilitation Hospital Affiliated to Capital Medical University (2022-001). Beijing Municipal Science and Technology Program: Research and Translational Application of Clinical Characteristic Diagnostic and Treatment Techniques in the Capital City (Project No. Z221100007422113).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained within the article.

Acknowledgments

I would like to thank Xu Fei of Huishou (Beijing) Technology Co. Ltd. for the helpful discussions, suggestions, and guidance in writing this paper.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Precht, L.; Keinath, A.; Krems, J.F. Identifying the main factors contributing to driving errors and traffic violations—Results from naturalistic driving data. Transp. Res. Part F Traffic Psychol. Behav. 2017, 49, 49–92. [Google Scholar] [CrossRef]
  2. Frazier, S.; McComb, S.A.; Hass, Z.; Pitts, B.J. The Moderating Effects of Task Complexity and Age on the Relationship between Automation Use and Cognitive Workload. Int. J. Hum.–Comput. Interact. 2024, 40, 1746–1764. [Google Scholar] [CrossRef]
  3. Ranney, T.A. Models of driving behavior: A review of their evolution. Accid. Anal. Prev. 1994, 26, 733–750. [Google Scholar] [CrossRef]
  4. Zicat, E.; Bennett, J.M.; Chekaluk, E.; Batchelor, J. Cognitive function and young drivers: The relationship between driving, attitudes, personality and cognition. Transp. Res. Part F Traffic Psychol. Behav. 2018, 55, 341–352. [Google Scholar] [CrossRef]
  5. Guo, M.; Hu, L.; Ye, L. Cognition and driving safety: How does the high-speed railway drivers’ cognitive ability affect safety performance? Transp. Res. Part F Traffic Psychol. Behav. 2019, 65, 10–22. [Google Scholar] [CrossRef]
  6. Vilchez, J.L.; Guaranda, M.E.R.; Polo, M.F.M.; Martínez, M.C.Á.; Castro, C.I.C.; Andrade, M.S.M.; Atiencia, W.X.T.; Alberca, D.O.; Valarezo, W.L.M. Cognitive psychology in traffic safety. Theor. Issues Ergon. Sci. 2024, 25, 474–494. [Google Scholar] [CrossRef]
  7. Groeger, J.A. Understanding Driving: Applying Cognitive Psychology to a Complex Everyday Task; Routledge: London, UK, 2013. [Google Scholar]
  8. Lohani, M.; Payne, B.R.; Strayer, D.L. A Review of Psychophysiological Measures to Assess Cognitive States in Real-World Driving. Front. Hum. Neurosci. 2019, 13, 57. [Google Scholar] [CrossRef] [PubMed]
  9. Moran, C.; Bennett, J.M.; Prabhakharan, P. The relationship between cognitive function and hazard perception in younger drivers. Transp. Res. Part F Traffic Psychol. Behav. 2020, 74, 104–119. [Google Scholar] [CrossRef]
  10. Ni, D.; Guo, F.; Zhou, Y.; Shi, C. Determination of Risk Perception of Drivers Using Fuzzy-Clustering Analysis for Road Safety. IEEE Access 2020, 8, 125501–125512. [Google Scholar] [CrossRef]
  11. Walshe, E.A.; Ward McIntosh, C.; Romer, D.; Winston, F.K. Executive Function Capacities, Negative Driving Behavior and Crashes in Young Drivers. Int. J. Environ. Res. Public Health 2017, 14, 1314. [Google Scholar] [CrossRef] [PubMed]
  12. Hatfield, J.; Williamson, A.; Kehoe, E.J.; Prabhakharan, P. An examination of the relationship between measures of impulsivity and risky simulated driving amongst young drivers. Accid. Anal. Prev. 2017, 103, 37–43. [Google Scholar] [CrossRef] [PubMed]
  13. Tefft, B. Rates of Motor Vehicle Crashes, Injuries and Deaths in Relation to Driver Age, United States, 2014–2015; AAA Foundation for Traffic Safety: Washington, DC, USA, 2017. [Google Scholar]
  14. Rupp, G.; Berka, C.; Meghdadi, A.H.; Karić, M.S.; Casillas, M.; Smith, S.; Rosenthal, T.; McShea, K.; Sones, E.; Marcotte, T.D. EEG-Based Neurocognitive Metrics May Predict Simulated and On-Road Driving Performance in Older Drivers. Front. Hum. Neurosci. 2019, 12, 532. [Google Scholar] [CrossRef] [PubMed]
  15. Zahabi, M.; Razak, A.M.A.; Shortz, A.E.; Mehta, R.K.; Manser, M. Evaluating advanced driver-assistance system trainings using driver performance, attention allocation, and neural efficiency measures. Appl. Ergon. 2020, 84, 103036. [Google Scholar] [CrossRef]
  16. Zatmeh-Kanj, S.; Toledo, T. Car following and microscopic traffic simulation under distracted driving. Transp. Res. Rec. 2021, 2675, 643–656. [Google Scholar] [CrossRef]
  17. Chen, J.; Jiang, Y.; Huang, Z.; Guo, X.; Wu, B.; Sun, L.; Wu, T. Fine-Grained Detection of Driver Distraction Based on Neural Architecture Search. IEEE Trans. Intell. Transp. Syst. 2021, 22, 5783–5801. [Google Scholar] [CrossRef]
  18. Mohammed, A.A.; Geng, X.; Wang, J.; Ali, Z. Driver distraction detection using semi-supervised lightweight vision transformer. Eng. Appl. Artif. Intell. 2024, 129, 107618. [Google Scholar] [CrossRef]
  19. Gonzalez, C. Task workload and cognitive abilities in dynamic decision making. Hum. Factors 2005, 47, 92–101. [Google Scholar] [CrossRef] [PubMed]
  20. Stanisław Jurecki, R.; Lech Stańczyk, T.; Jacek Jaśkiewicz, M. Driver’s reaction time in a simulated, complex road incident. Transport 2017, 32, 44–54. [Google Scholar] [CrossRef]
  21. Hajinoroozi, M.; Mao, Z.; Jung, T.-P.; Lin, C.-T.; Huang, Y. EEG-based prediction of driver’s cognitive performance by deep convolutional neural network. Signal Process. Image Commun. 2016, 47, 549–555. [Google Scholar] [CrossRef]
  22. Foy, H.J.; Chapman, P. Mental workload is reflected in driver behaviour, physiology, eye movements and prefrontal cortex activation. Appl. Ergon. 2018, 73, 90–99. [Google Scholar] [CrossRef] [PubMed]
  23. Bloomfield, P.M.; Green, H.; Gant, N. Cerebral haemodynamics during simulated driving: Changes in workload are detectable with functional near infrared spectroscopy. PLoS ONE 2021, 16, e0248533. [Google Scholar] [CrossRef]
  24. Broadbent, D.P.; D’Innocenzo, G.; Ellmers, T.J.; Parsler, J.; Szameitat, A.J.; Bishop, D.T. Cognitive load, working memory capacity and driving performance: A preliminary fNIRS and eye tracking study. Transp. Res. Part F Traffic Psychol. Behav. 2023, 92, 121–132. [Google Scholar] [CrossRef]
  25. Ma, Y.; Chen, S.; Xing, G.; Xie, Z.; Wang, F. A dynamic method to predict driving risk on sharp curves using multi-source data. Accid. Anal. Prev. 2023, 191, 107228. [Google Scholar] [CrossRef]
  26. Yuksel, A.S.; Atmaca, S. Driver’s black box: A system for driver risk assessment using machine learning and fuzzy logic. J. Intell. Transp. Syst. 2021, 25, 482–500. [Google Scholar] [CrossRef]
  27. Moghaddam, A.M.; Ayati, E. Introducing a risk estimation index for drivers: A case of Iran. Saf. Sci. 2014, 62, 90–97. [Google Scholar] [CrossRef]
  28. Depestele, S.; Ross, V.; Verstraelen, S.; Brijs, K.; Brijs, T.; van Dun, K.; Meesen, R. The impact of cognitive functioning on driving performance of older persons in comparison to younger age groups: A systematic review. Transp. Res. Part F Traffic Psychol. Behav. 2020, 73, 433–452. [Google Scholar] [CrossRef]
  29. Ledger, S.; Bennett, J.M.; Chekaluk, E.; Batchelor, J. Cognitive function and driving: Important for young and old alike. Transp. Res. Part F Traffic Psychol. Behav. 2019, 60, 262–273. [Google Scholar] [CrossRef]
  30. Mullen, N.; Chattha, H.; Weaver, B.; Bedard, M. Older driver performance on a simulator: Associations between simulated tasks and cognition. Adv. Transp. Stud. Spec. 2008, 31–42. [Google Scholar]
  31. Stolwyk, R.J.; Charlton, J.L.; Triggs, T.J.; Iansek, R.; Bradshaw, J.L. Neuropsychological function and driving ability in people with Parkinson’s disease. J. Clin. Exp. Neuropsychol. 2006, 28, 898–913. [Google Scholar] [CrossRef] [PubMed]
  32. Mou, L.; Chang, J.; Zhou, C.; Zhao, Y.; Ma, N.; Yin, B.; Jain, R.; Gao, W. Multimodal driver distraction detection using dual-channel network of CNN and Transformer. Expert Syst. Appl. 2023, 234, 121066. [Google Scholar] [CrossRef]
  33. Arvin, R.; Khattak, A.J.; Qi, H. Safety critical event prediction through unified analysis of driver and vehicle volatilities: Application of deep learning methods. Accid. Anal. Prev. 2021, 151, 105949. [Google Scholar] [CrossRef] [PubMed]
  34. Mou, L.; Zhou, C.; Zhao, P.; Nakisa, B.; Rastgoo, M.N.; Jain, R.; Gao, W. Driver stress detection via multimodal fusion using attention-based CNN-LSTM. Expert Syst. Appl. 2021, 173, 114693. [Google Scholar] [CrossRef]
  35. Arefnezhad, S.; Samiee, S.; Eichberger, A.; Frühwirth, M.; Kaufmann, C.; Klotz, E. Applying deep neural networks for multi-level classification of driver drowsiness using Vehicle-based measures. Expert Syst. Appl. 2020, 162, 113778. [Google Scholar] [CrossRef]
  36. Vyas, J.; Bhardwaj, N.; Bhumika; Das, D. TransDBC: Transformer for Multivariate Time-Series based Driver Behavior Classification. In Proceedings of the 2022 International Joint Conference on Neural Networks (IJCNN), Padua, Italy, 18–23 July 2022; pp. 1–8. [Google Scholar]
  37. Taamneh, S.; Tsiamyrtzis, P.; Dcosta, M.; Buddharaju, P.; Khatri, A.; Manser, M.; Ferris, T.; Wunderlich, R.; Pavlidis, I. A multimodal dataset for various forms of distracted driving. Sci. Data 2017, 4, 170110. [Google Scholar] [CrossRef] [PubMed]
  38. Kim, B.; Bishu, R.R. Cognitive abilities in driving: Differences between normal and hazardous situations. Ergonomics 2004, 47, 1037–1052. [Google Scholar] [CrossRef] [PubMed]
  39. Shinar, D. Driver information processing: Attention, perception, reaction time, and comprehension. In Traffic Safety and Human Behavior; Emerald Publishing Limited: Bingley, UK, 2017; pp. 189–256. [Google Scholar]
  40. Tapia, J.L.; Duñabeitia, J.A. Driving safety: Investigating the cognitive foundations of accident prevention. Heliyon 2023, 9, e21355. [Google Scholar] [CrossRef]
  41. Almutairi, M.; Stahl, F.; Bramer, M. Reg-rules: An explainable rule-based ensemble learner for classification. IEEE Access 2021, 9, 52015–52035. [Google Scholar] [CrossRef]
  42. Anstey, K.J.; Wood, J.; Lord, S.; Walker, J.G. Cognitive, sensory and physical factors enabling driving safety in older adults. Clin. Psychol. Rev. 2005, 25, 45–65. [Google Scholar] [CrossRef] [PubMed]
  43. Abdi, H.; Williams, L.J. Principal component analysis. Wiley Interdiscip. Rev. Comput. Stat. 2010, 2, 433–459. [Google Scholar] [CrossRef]
  44. Dai, Y.; Gieseke, F.; Oehmcke, S.; Wu, Y.; Barnard, K. Attentional feature fusion. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 3–8 January 2021; pp. 3560–3569. [Google Scholar]
  45. Dosovitskiy, A. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  46. Brookshire, G.; Kasper, J.; Blauch, N.M.; Wu, Y.C.; Glatt, R.; Merrill, D.A.; Gerrol, S.; Yoder, K.J.; Quirk, C.; Lucero, C.; et al. Data leakage in deep learning studies of translational EEG. Front. Neurosci. 2024, 18, 1373515. [Google Scholar] [CrossRef] [PubMed]
  47. Suthaharan, S. Support vector machine. In Machine Learning Models and Algorithms for Big Data Classification; Springer: Boston, MA, USA, 2016; pp. 207–235. [Google Scholar]
  48. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  49. Graves, A. Long short-term memory. In Supervised Sequence Labelling with Recurrent Neural Networks; Springer: Berlin/Heidelberg, Germany, 2012; pp. 37–45. [Google Scholar]
  50. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  51. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
Figure 1. Overall technical framework.
Figure 1. Overall technical framework.
Sensors 25 00174 g001
Figure 2. Multimodal feature fusion network model.
Figure 2. Multimodal feature fusion network model.
Sensors 25 00174 g002
Figure 3. AFF network model.
Figure 3. AFF network model.
Sensors 25 00174 g003
Figure 4. Confusion matrix results for SVM (a), RF (b), LSTM (c), VGG-16 (d), ResNet (e), CNN-LSTM (f), and our (g).
Figure 4. Confusion matrix results for SVM (a), RF (b), LSTM (c), VGG-16 (d), ResNet (e), CNN-LSTM (f), and our (g).
Sensors 25 00174 g004
Figure 5. Comparison of experimental visualization results.
Figure 5. Comparison of experimental visualization results.
Sensors 25 00174 g005
Table 1. Relevant studies on driving behavior assessment.
Table 1. Relevant studies on driving behavior assessment.
ReferenceSubject of StudyDataAssessment MethodPerformance (ACC)
Stanisław, et al. [20]Reaction SpeedAccelerator Pedal, Brakes, Steering WheelStatistical Analysis-
Broadbent, et al. [24]Working Memory CapacityFNIRS,
Eye Tracker
Statistical Analysis-
Moran, et al. [9]Hazard Perception AbilityCognitive Tests, Hazard Perception Test (HPT)Statistical Analysis-
Chen, et al. [17]DistractionEye, Vehicle, PhysiologicalDeep-CNN99.78%
Mohammed, et al. [18]DistractionDriving Image DataSemi-supervised Lightweight Hybrid VIT95.43%
Mou, et al. [32]DistractionEye, Vehicle, PhysiologicalCNN-Trans-SA99.80%
Arvin, et al. [33]DistractionVehicle Motion, Cabin Video, Environment1DCNN-LSTM95.45%
Mou, et al. [34]StressEye Data, Vehicle Data, Environmental DataCNN-LSTM95.5%
Arefnezhad, et al. [35]DrowsinessVehicleCNN-LSTM96.0%
Vyas, et al. [36]Driving SafetyAccelerometer, Gyroscope, and GPSTrans-DBC95.38%
Table 2. Driving-related cognitive abilities extraction method.
Table 2. Driving-related cognitive abilities extraction method.
DimensionDriving-Related Cognitive AbilitiesTask ScenarioData Sources and Extraction Method
Cognitive decision-makingAttention Allocation AbilityCompleting secondary tasks
Completing secondary tasks
Passing through an intersection
Record the number of times secondary tasks were avoided
Working Memory CapacityACC of completing secondary tasks
Anticipatory AbilityAccurate anticipation of traffic signals
Vehicle ControlReaction AbilityVehicle malfunction acceleration, a car suddenly entering the road from the leftTime from recognizing an emergency to execution (e.g., braking or evading)
Judgment AbilityRecord evasion strategies and extent of vehicle collision
Processing SpeedCompleting secondary tasksReaction time to complete secondary tasks
Perception AbilityPassing through an intersectionWhether the line is crossed during red light waiting
Psychological regulationStress ResistanceBefore and after simulated driving tasksNASA scale score
Anxiety LevelAnxiety scale score
Type A/B PersonalityType A/B personality scale score
Table 4. ACC and F1-score experimental results for all compared algorithms.
Table 4. ACC and F1-score experimental results for all compared algorithms.
Comparison AlgorithmEpochACCF1-Score
SVM-0.62950.6001
RF-0.85830.8504
LSTM1000.65610.6368
VGG-161000.66990.4932
ResNet1000.94710.9264
CNN-LSTM1000.95970.9302
Our1000.99080.9832
Table 5. Ablation experiments.
Table 5. Ablation experiments.
ExperimentsEpochACCF1-Score
Baseline1000.93100.9225
Baseline + Dual-Branch1000.96090.9520
Baseline + Dual-Branch + AFF (our)1000.99080.9832
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, Y.; Liu, B.; Zhang, W. Driving-Related Cognitive Abilities Prediction Based on Transformer’s Multimodal Fusion Framework. Sensors 2025, 25, 174. https://doi.org/10.3390/s25010174

AMA Style

Li Y, Liu B, Zhang W. Driving-Related Cognitive Abilities Prediction Based on Transformer’s Multimodal Fusion Framework. Sensors. 2025; 25(1):174. https://doi.org/10.3390/s25010174

Chicago/Turabian Style

Li, Yifan, Bo Liu, and Wenli Zhang. 2025. "Driving-Related Cognitive Abilities Prediction Based on Transformer’s Multimodal Fusion Framework" Sensors 25, no. 1: 174. https://doi.org/10.3390/s25010174

APA Style

Li, Y., Liu, B., & Zhang, W. (2025). Driving-Related Cognitive Abilities Prediction Based on Transformer’s Multimodal Fusion Framework. Sensors, 25(1), 174. https://doi.org/10.3390/s25010174

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop