Driving-Related Cognitive Abilities Prediction Based on Transformer’s Multimodal Fusion Framework
<p>Overall technical framework.</p> "> Figure 2
<p>Multimodal feature fusion network model.</p> "> Figure 3
<p>AFF network model.</p> "> Figure 4
<p>Confusion matrix results for SVM (<b>a</b>), RF (<b>b</b>), LSTM (<b>c</b>), VGG-16 (<b>d</b>), ResNet (<b>e</b>), CNN-LSTM (<b>f</b>), and our (<b>g</b>).</p> "> Figure 5
<p>Comparison of experimental visualization results.</p> ">
Abstract
:1. Introduction
- Introduction of a Multidimensional Cognitive Ability Evaluation System: This study defines and quantifies cognitive factors affecting driving safety from three dimensions—cognitive decision-making, vehicle control, and psychological state. It extracts and quantifies multiple cognitive ability indicators related to driving, thereby providing a comprehensive assessment of drivers’ cognitive capabilities and addressing the limitations of previous studies that focused solely on single cognitive dimensions.
- Development of a Multimodal Data Feature Extraction Model: To address the challenge of effectively extracting common features between multimodal data in traditional models, this study proposes a dual-branch Transformer network. The dual-branch structure eliminates interference between input data and effectively learns high-dimensional features of the data. Additionally, for different data sources, a dynamic weight allocation-based feature fusion mechanism is introduced, enhancing the feature representation ability and predictive performance of the final model.
2. Materials and Methods
2.1. The Datasets
- 1.
- Physiological sensors include the following:
- Skin galvanic response signals were collected using the Shimmer3 GSR sensor (Shimmer, Dublin, Ireland) with a sampling frequency of 25 Hz to record the driver’s physiological arousal state.
- Heart rate and respiratory rate were collected using the Zephyr BioHarness 3.0 (Zephyr Technology, Annapolis, Maryland) with a sampling frequency of 1 Hz.
- The Tau 640 long-wave infrared (LWIR) camera (FLIR Commercial Systems, Goleta, California) with a frame rate of 7.5 fps was used to capture perspiration signals around the nose area.
- 2.
- Non-physiological sensors include the following:
- Eye movement parameters, such as gaze position and pupil diameter, were recorded using the faceLAB system (Seeing Machines, Canberra, Australia) with a sampling frequency of 25 Hz to analyze the driver’s attention state.
- Data on vehicle speed, acceleration, braking, steering, lane deviation, and lane position were collected using the driving simulator manufactured by Realtime Technologies, Inc. (Royal Oak, Michigan) with a sampling frequency of 58.8 Hz to analyze vehicle control state.
- Emotional state data were recorded using the HD Pro Webcam C920 (Logitech, Newark, CA, USA) monochrome zoom camera with a frame rate of 15 fps. The data were pre-processed and analyzed to assess changes in the driver’s emotional state.
2.2. General Structure of the Framework
- Data Preprocessing Module: Due to the complexity and diversity of the data contained in multimodal driving data, the raw data need to be preprocessed before feature extraction. This module initially removes unstable data recorded before the start of the experiment and addresses any missing data. It also eliminates motion artifacts caused by actions such as blinking and selects an appropriate sliding window to provide sufficient samples for training.
- Driving-Related Cognitive Abilities Extraction Module: This module is designed to assess the performance of drivers during various secondary tasks and malfunction handling tasks. It extracts and quantitatively represents the components of driving cognitive-related abilities based on this performance.
- Driving-Related Cognitive Abilities Label Generation Module: This module assigns weights to the quantified components of driving-related cognitive abilities based on their importance and performs a weighted sum. Labels for driving-related cognitive abilities are then generated based on the weighted scores.
- Driving-Related Cognitive Abilities Prediction Network Model: After the above modules, clean driving data and labels for driving-related cognitive abilities were obtained. At this point, an appropriate network model is required to extract and fuse features from the processed data, enabling precise prediction of driving-related cognitive abilities.
2.3. Data Preprocessing Module
2.4. Driving-Related Cognitive Abilities Label Setting
- Cognitive decision-making dimension: This explores how drivers process received traffic information and quickly make decisions based on this information. It includes the assessment of drivers’ attention allocation, working memory, and anticipation abilities.
- Vehicle control dimension: This directly reflects how drivers translate decisions into vehicle operations, covering assessments of drivers’ processing speed, reaction capability, judgment, and perceptual skills.
- Psychological regulation dimension: This examines how drivers influence decision-making and action execution through the regulation of their psychological state, assessing drivers’ stress resilience, anxiety levels, and Type A/B personality traits.
2.4.1. Driving-Related Cognitive Abilities Extraction Module
- First, directly quantifiable indicators such as reaction times and scale scores are kept in their original values for standardization.
- Second, qualitative behavioral performances, such as the degree of collision, are numerically processed. For example, ’no collision’ is marked as 0, and ’collision’ is marked as 1. This step completes the numerical transformation of all data.
- Finally, the formula is applied to standardize the numerically transformed data.
2.4.2. Driving-Related Cognitive Abilities Label Generation Module
- Analysis of the Importance of Components of Driving-Related Cognitive Abilities: Anstey et al. [42] found that thought decision-making directly reflects the driver’s response to environmental changes and decision-making efficiency, hence it is of the highest importance. Vehicle control reflects the driver’s actual execution ability in controlling the vehicle and responding to road conditions, making it the second most important. Psychological adjustment, while having a non-negligible impact on driving safety, has the least influence compared to thought decision-making and vehicle control abilities that directly determine driving response and vehicle control, thus it is of the lowest importance.
- Generation of Weighted Scores for Components of Driving-Related Cognitive Abilities: Based on step 1, the importance of three dimensions of driving-related cognitive abilities from high to low are: Cognitive decision-making (), vehicle control (), and psychological regulation (). The weights for each cognitive ability dimension were determined using the Principal Component Analysis (PCA) method [43]. Specifically, PCA was applied to the indicators extracted in Table 2 to identify their linear combination coefficients in each principal component. Based on the variance contribution rate of each principal component, the weights , , were calculated and assigned to , , and , respectively. The driving-related cognitive abilities score ( for short) is obtained through weighting. and can be calculated using Equations (3) and (4). The distribution of driving-related cognitive abilities weights is shown in Table 3.
Driving-Related Cognitive Abilities Dimensions | Weighting | |
---|---|---|
Thinking Decision Making | 0.3898 | |
Vehicle Handling | 0.3182 | |
Mental State | 0.2920 |
- Generation of Driving Cognitive-Related Ability Labels: Driving-related cognitive abilities labels are divided into four levels, 0–3, representing four levels of driving-related cognitive abilities: Insufficient, Fair, Good, and Excellent. These four levels are determined based on the derived from Equation (4). The calculation of the incorporates the weighted values of ten components of driving-related cognitive abilities, reflecting the comprehensive level of driving-related cognitive abilities. Since these components of driving-related cognitive abilities have different distribution characteristics in terms of centrality and dispersion, the weighted results in a multivariate composite distribution. To effectively categorize Scores into levels, this study employs the G-Rules-IQR method. This method does not rely on the normal distribution properties of the data and can determine the score thresholds for “Score” by calculating the interquartile range (IQR). This method maps the continuous “Score” to four discrete intervals {[, ), [, ), [, ), [, ]}, achieving the classification of driving-related cognitive abilities levels and thus generating driving-related cognitive abilities labels. The can be calculated using Equation (5).
2.5. Driving-Related Cognitive Abilities Prediction Network Model
- Multi-source Heterogeneous Feature Conversion Module: This module is responsible for separating physiological data from non-physiological data from multi-modal driving time-series data, and converting them from the original 2D time-series data to 2D image data, which meets the needs of the dual-branch Transformer encoder module for feature extraction from the input data.
- Dual-branch Transformer Encoder Module: This module adopts an improved dual-branch structure, which allows physiological and non-physiological data to be feature extracted by independent transformer encoders. This dual-branch structure allows the physiological branch to focus on capturing changes in the detailed features of the physiological signals in the short term, and the non-physiological branch to focus on capturing changes in the long-term features such as vehicle operating modes. A self-attention mechanism is used within each branch to optimize the capture of dynamic time-series data features and enhance the feature extraction capability of the model for multimodal driving time-series data.
- AFF module based on dynamic weight assignment: after physiological and non-physiological features are extracted by the Dual-branch Transformer encoder module, the attention weights of global and local feature channels are extracted by the AFF module using global pooling and point-by-point convolution, where global pooling is in charge of capturing the global feature information and point-by-point convolution is used for extracting and strengthening local feature details. Finally, the performance of the prediction model is further optimized by feature fusion.
2.5.1. Multi-Source Heterogeneous Feature Conversion Module
2.5.2. Dual-Branch Transformer Encoder Module
2.5.3. AFF Module Based on Dynamic Weight Assignment
3. Experimental Results and Analysis
3.1. Data Set Segmentation
3.2. Evaluation Metrics
3.3. Comparison Experiment
3.4. Ablation Experiments
4. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- Precht, L.; Keinath, A.; Krems, J.F. Identifying the main factors contributing to driving errors and traffic violations—Results from naturalistic driving data. Transp. Res. Part F Traffic Psychol. Behav. 2017, 49, 49–92. [Google Scholar] [CrossRef]
- Frazier, S.; McComb, S.A.; Hass, Z.; Pitts, B.J. The Moderating Effects of Task Complexity and Age on the Relationship between Automation Use and Cognitive Workload. Int. J. Hum.–Comput. Interact. 2024, 40, 1746–1764. [Google Scholar] [CrossRef]
- Ranney, T.A. Models of driving behavior: A review of their evolution. Accid. Anal. Prev. 1994, 26, 733–750. [Google Scholar] [CrossRef]
- Zicat, E.; Bennett, J.M.; Chekaluk, E.; Batchelor, J. Cognitive function and young drivers: The relationship between driving, attitudes, personality and cognition. Transp. Res. Part F Traffic Psychol. Behav. 2018, 55, 341–352. [Google Scholar] [CrossRef]
- Guo, M.; Hu, L.; Ye, L. Cognition and driving safety: How does the high-speed railway drivers’ cognitive ability affect safety performance? Transp. Res. Part F Traffic Psychol. Behav. 2019, 65, 10–22. [Google Scholar] [CrossRef]
- Vilchez, J.L.; Guaranda, M.E.R.; Polo, M.F.M.; Martínez, M.C.Á.; Castro, C.I.C.; Andrade, M.S.M.; Atiencia, W.X.T.; Alberca, D.O.; Valarezo, W.L.M. Cognitive psychology in traffic safety. Theor. Issues Ergon. Sci. 2024, 25, 474–494. [Google Scholar] [CrossRef]
- Groeger, J.A. Understanding Driving: Applying Cognitive Psychology to a Complex Everyday Task; Routledge: London, UK, 2013. [Google Scholar]
- Lohani, M.; Payne, B.R.; Strayer, D.L. A Review of Psychophysiological Measures to Assess Cognitive States in Real-World Driving. Front. Hum. Neurosci. 2019, 13, 57. [Google Scholar] [CrossRef] [PubMed]
- Moran, C.; Bennett, J.M.; Prabhakharan, P. The relationship between cognitive function and hazard perception in younger drivers. Transp. Res. Part F Traffic Psychol. Behav. 2020, 74, 104–119. [Google Scholar] [CrossRef]
- Ni, D.; Guo, F.; Zhou, Y.; Shi, C. Determination of Risk Perception of Drivers Using Fuzzy-Clustering Analysis for Road Safety. IEEE Access 2020, 8, 125501–125512. [Google Scholar] [CrossRef]
- Walshe, E.A.; Ward McIntosh, C.; Romer, D.; Winston, F.K. Executive Function Capacities, Negative Driving Behavior and Crashes in Young Drivers. Int. J. Environ. Res. Public Health 2017, 14, 1314. [Google Scholar] [CrossRef] [PubMed]
- Hatfield, J.; Williamson, A.; Kehoe, E.J.; Prabhakharan, P. An examination of the relationship between measures of impulsivity and risky simulated driving amongst young drivers. Accid. Anal. Prev. 2017, 103, 37–43. [Google Scholar] [CrossRef] [PubMed]
- Tefft, B. Rates of Motor Vehicle Crashes, Injuries and Deaths in Relation to Driver Age, United States, 2014–2015; AAA Foundation for Traffic Safety: Washington, DC, USA, 2017. [Google Scholar]
- Rupp, G.; Berka, C.; Meghdadi, A.H.; Karić, M.S.; Casillas, M.; Smith, S.; Rosenthal, T.; McShea, K.; Sones, E.; Marcotte, T.D. EEG-Based Neurocognitive Metrics May Predict Simulated and On-Road Driving Performance in Older Drivers. Front. Hum. Neurosci. 2019, 12, 532. [Google Scholar] [CrossRef] [PubMed]
- Zahabi, M.; Razak, A.M.A.; Shortz, A.E.; Mehta, R.K.; Manser, M. Evaluating advanced driver-assistance system trainings using driver performance, attention allocation, and neural efficiency measures. Appl. Ergon. 2020, 84, 103036. [Google Scholar] [CrossRef]
- Zatmeh-Kanj, S.; Toledo, T. Car following and microscopic traffic simulation under distracted driving. Transp. Res. Rec. 2021, 2675, 643–656. [Google Scholar] [CrossRef]
- Chen, J.; Jiang, Y.; Huang, Z.; Guo, X.; Wu, B.; Sun, L.; Wu, T. Fine-Grained Detection of Driver Distraction Based on Neural Architecture Search. IEEE Trans. Intell. Transp. Syst. 2021, 22, 5783–5801. [Google Scholar] [CrossRef]
- Mohammed, A.A.; Geng, X.; Wang, J.; Ali, Z. Driver distraction detection using semi-supervised lightweight vision transformer. Eng. Appl. Artif. Intell. 2024, 129, 107618. [Google Scholar] [CrossRef]
- Gonzalez, C. Task workload and cognitive abilities in dynamic decision making. Hum. Factors 2005, 47, 92–101. [Google Scholar] [CrossRef] [PubMed]
- Stanisław Jurecki, R.; Lech Stańczyk, T.; Jacek Jaśkiewicz, M. Driver’s reaction time in a simulated, complex road incident. Transport 2017, 32, 44–54. [Google Scholar] [CrossRef]
- Hajinoroozi, M.; Mao, Z.; Jung, T.-P.; Lin, C.-T.; Huang, Y. EEG-based prediction of driver’s cognitive performance by deep convolutional neural network. Signal Process. Image Commun. 2016, 47, 549–555. [Google Scholar] [CrossRef]
- Foy, H.J.; Chapman, P. Mental workload is reflected in driver behaviour, physiology, eye movements and prefrontal cortex activation. Appl. Ergon. 2018, 73, 90–99. [Google Scholar] [CrossRef] [PubMed]
- Bloomfield, P.M.; Green, H.; Gant, N. Cerebral haemodynamics during simulated driving: Changes in workload are detectable with functional near infrared spectroscopy. PLoS ONE 2021, 16, e0248533. [Google Scholar] [CrossRef]
- Broadbent, D.P.; D’Innocenzo, G.; Ellmers, T.J.; Parsler, J.; Szameitat, A.J.; Bishop, D.T. Cognitive load, working memory capacity and driving performance: A preliminary fNIRS and eye tracking study. Transp. Res. Part F Traffic Psychol. Behav. 2023, 92, 121–132. [Google Scholar] [CrossRef]
- Ma, Y.; Chen, S.; Xing, G.; Xie, Z.; Wang, F. A dynamic method to predict driving risk on sharp curves using multi-source data. Accid. Anal. Prev. 2023, 191, 107228. [Google Scholar] [CrossRef]
- Yuksel, A.S.; Atmaca, S. Driver’s black box: A system for driver risk assessment using machine learning and fuzzy logic. J. Intell. Transp. Syst. 2021, 25, 482–500. [Google Scholar] [CrossRef]
- Moghaddam, A.M.; Ayati, E. Introducing a risk estimation index for drivers: A case of Iran. Saf. Sci. 2014, 62, 90–97. [Google Scholar] [CrossRef]
- Depestele, S.; Ross, V.; Verstraelen, S.; Brijs, K.; Brijs, T.; van Dun, K.; Meesen, R. The impact of cognitive functioning on driving performance of older persons in comparison to younger age groups: A systematic review. Transp. Res. Part F Traffic Psychol. Behav. 2020, 73, 433–452. [Google Scholar] [CrossRef]
- Ledger, S.; Bennett, J.M.; Chekaluk, E.; Batchelor, J. Cognitive function and driving: Important for young and old alike. Transp. Res. Part F Traffic Psychol. Behav. 2019, 60, 262–273. [Google Scholar] [CrossRef]
- Mullen, N.; Chattha, H.; Weaver, B.; Bedard, M. Older driver performance on a simulator: Associations between simulated tasks and cognition. Adv. Transp. Stud. Spec. 2008, 31–42. [Google Scholar]
- Stolwyk, R.J.; Charlton, J.L.; Triggs, T.J.; Iansek, R.; Bradshaw, J.L. Neuropsychological function and driving ability in people with Parkinson’s disease. J. Clin. Exp. Neuropsychol. 2006, 28, 898–913. [Google Scholar] [CrossRef] [PubMed]
- Mou, L.; Chang, J.; Zhou, C.; Zhao, Y.; Ma, N.; Yin, B.; Jain, R.; Gao, W. Multimodal driver distraction detection using dual-channel network of CNN and Transformer. Expert Syst. Appl. 2023, 234, 121066. [Google Scholar] [CrossRef]
- Arvin, R.; Khattak, A.J.; Qi, H. Safety critical event prediction through unified analysis of driver and vehicle volatilities: Application of deep learning methods. Accid. Anal. Prev. 2021, 151, 105949. [Google Scholar] [CrossRef] [PubMed]
- Mou, L.; Zhou, C.; Zhao, P.; Nakisa, B.; Rastgoo, M.N.; Jain, R.; Gao, W. Driver stress detection via multimodal fusion using attention-based CNN-LSTM. Expert Syst. Appl. 2021, 173, 114693. [Google Scholar] [CrossRef]
- Arefnezhad, S.; Samiee, S.; Eichberger, A.; Frühwirth, M.; Kaufmann, C.; Klotz, E. Applying deep neural networks for multi-level classification of driver drowsiness using Vehicle-based measures. Expert Syst. Appl. 2020, 162, 113778. [Google Scholar] [CrossRef]
- Vyas, J.; Bhardwaj, N.; Bhumika; Das, D. TransDBC: Transformer for Multivariate Time-Series based Driver Behavior Classification. In Proceedings of the 2022 International Joint Conference on Neural Networks (IJCNN), Padua, Italy, 18–23 July 2022; pp. 1–8. [Google Scholar]
- Taamneh, S.; Tsiamyrtzis, P.; Dcosta, M.; Buddharaju, P.; Khatri, A.; Manser, M.; Ferris, T.; Wunderlich, R.; Pavlidis, I. A multimodal dataset for various forms of distracted driving. Sci. Data 2017, 4, 170110. [Google Scholar] [CrossRef] [PubMed]
- Kim, B.; Bishu, R.R. Cognitive abilities in driving: Differences between normal and hazardous situations. Ergonomics 2004, 47, 1037–1052. [Google Scholar] [CrossRef] [PubMed]
- Shinar, D. Driver information processing: Attention, perception, reaction time, and comprehension. In Traffic Safety and Human Behavior; Emerald Publishing Limited: Bingley, UK, 2017; pp. 189–256. [Google Scholar]
- Tapia, J.L.; Duñabeitia, J.A. Driving safety: Investigating the cognitive foundations of accident prevention. Heliyon 2023, 9, e21355. [Google Scholar] [CrossRef]
- Almutairi, M.; Stahl, F.; Bramer, M. Reg-rules: An explainable rule-based ensemble learner for classification. IEEE Access 2021, 9, 52015–52035. [Google Scholar] [CrossRef]
- Anstey, K.J.; Wood, J.; Lord, S.; Walker, J.G. Cognitive, sensory and physical factors enabling driving safety in older adults. Clin. Psychol. Rev. 2005, 25, 45–65. [Google Scholar] [CrossRef] [PubMed]
- Abdi, H.; Williams, L.J. Principal component analysis. Wiley Interdiscip. Rev. Comput. Stat. 2010, 2, 433–459. [Google Scholar] [CrossRef]
- Dai, Y.; Gieseke, F.; Oehmcke, S.; Wu, Y.; Barnard, K. Attentional feature fusion. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 3–8 January 2021; pp. 3560–3569. [Google Scholar]
- Dosovitskiy, A. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
- Brookshire, G.; Kasper, J.; Blauch, N.M.; Wu, Y.C.; Glatt, R.; Merrill, D.A.; Gerrol, S.; Yoder, K.J.; Quirk, C.; Lucero, C.; et al. Data leakage in deep learning studies of translational EEG. Front. Neurosci. 2024, 18, 1373515. [Google Scholar] [CrossRef] [PubMed]
- Suthaharan, S. Support vector machine. In Machine Learning Models and Algorithms for Big Data Classification; Springer: Boston, MA, USA, 2016; pp. 207–235. [Google Scholar]
- Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
- Graves, A. Long short-term memory. In Supervised Sequence Labelling with Recurrent Neural Networks; Springer: Berlin/Heidelberg, Germany, 2012; pp. 37–45. [Google Scholar]
- Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
- He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
Reference | Subject of Study | Data | Assessment Method | Performance (ACC) |
---|---|---|---|---|
Stanisław, et al. [20] | Reaction Speed | Accelerator Pedal, Brakes, Steering Wheel | Statistical Analysis | - |
Broadbent, et al. [24] | Working Memory Capacity | FNIRS, Eye Tracker | Statistical Analysis | - |
Moran, et al. [9] | Hazard Perception Ability | Cognitive Tests, Hazard Perception Test (HPT) | Statistical Analysis | - |
Chen, et al. [17] | Distraction | Eye, Vehicle, Physiological | Deep-CNN | 99.78% |
Mohammed, et al. [18] | Distraction | Driving Image Data | Semi-supervised Lightweight Hybrid VIT | 95.43% |
Mou, et al. [32] | Distraction | Eye, Vehicle, Physiological | CNN-Trans-SA | 99.80% |
Arvin, et al. [33] | Distraction | Vehicle Motion, Cabin Video, Environment | 1DCNN-LSTM | 95.45% |
Mou, et al. [34] | Stress | Eye Data, Vehicle Data, Environmental Data | CNN-LSTM | 95.5% |
Arefnezhad, et al. [35] | Drowsiness | Vehicle | CNN-LSTM | 96.0% |
Vyas, et al. [36] | Driving Safety | Accelerometer, Gyroscope, and GPS | Trans-DBC | 95.38% |
Dimension | Driving-Related Cognitive Abilities | Task Scenario | Data Sources and Extraction Method |
---|---|---|---|
Cognitive decision-making | Attention Allocation Ability | Completing secondary tasks Completing secondary tasks Passing through an intersection | Record the number of times secondary tasks were avoided |
Working Memory Capacity | ACC of completing secondary tasks | ||
Anticipatory Ability | Accurate anticipation of traffic signals | ||
Vehicle Control | Reaction Ability | Vehicle malfunction acceleration, a car suddenly entering the road from the left | Time from recognizing an emergency to execution (e.g., braking or evading) |
Judgment Ability | Record evasion strategies and extent of vehicle collision | ||
Processing Speed | Completing secondary tasks | Reaction time to complete secondary tasks | |
Perception Ability | Passing through an intersection | Whether the line is crossed during red light waiting | |
Psychological regulation | Stress Resistance | Before and after simulated driving tasks | NASA scale score |
Anxiety Level | Anxiety scale score | ||
Type A/B Personality | Type A/B personality scale score |
Comparison Algorithm | Epoch | ACC | F1-Score |
---|---|---|---|
SVM | - | 0.6295 | 0.6001 |
RF | - | 0.8583 | 0.8504 |
LSTM | 100 | 0.6561 | 0.6368 |
VGG-16 | 100 | 0.6699 | 0.4932 |
ResNet | 100 | 0.9471 | 0.9264 |
CNN-LSTM | 100 | 0.9597 | 0.9302 |
Our | 100 | 0.9908 | 0.9832 |
Experiments | Epoch | ACC | F1-Score |
---|---|---|---|
Baseline | 100 | 0.9310 | 0.9225 |
Baseline + Dual-Branch | 100 | 0.9609 | 0.9520 |
Baseline + Dual-Branch + AFF (our) | 100 | 0.9908 | 0.9832 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Li, Y.; Liu, B.; Zhang, W. Driving-Related Cognitive Abilities Prediction Based on Transformer’s Multimodal Fusion Framework. Sensors 2025, 25, 174. https://doi.org/10.3390/s25010174
Li Y, Liu B, Zhang W. Driving-Related Cognitive Abilities Prediction Based on Transformer’s Multimodal Fusion Framework. Sensors. 2025; 25(1):174. https://doi.org/10.3390/s25010174
Chicago/Turabian StyleLi, Yifan, Bo Liu, and Wenli Zhang. 2025. "Driving-Related Cognitive Abilities Prediction Based on Transformer’s Multimodal Fusion Framework" Sensors 25, no. 1: 174. https://doi.org/10.3390/s25010174
APA StyleLi, Y., Liu, B., & Zhang, W. (2025). Driving-Related Cognitive Abilities Prediction Based on Transformer’s Multimodal Fusion Framework. Sensors, 25(1), 174. https://doi.org/10.3390/s25010174