[go: up one dir, main page]
More Web Proxy on the site http://driver.im/ skip to main content
research-article

Recognition of emotions using multimodal physiological signals and an ensemble deep learning model

Published: 01 March 2017 Publication History

Abstract

An ensemble of deep classifiers is built for recognizing emotions using multimodal physiological signals.The higher-level abstractions of physiological features of each modality are separately extracted by deep hidden neurons in member stacked-autoencoders.The minimal structure of the deep model is identified according to a structural loss function for local geometrical information preservation.The physiological feature abstractions are merged via an adjacent-graph based fusion network with hierarchical layers. Background and ObjectiveUsing deep-learning methodologies to analyze multimodal physiological signals becomes increasingly attractive for recognizing human emotions. However, the conventional deep emotion classifiers may suffer from the drawback of the lack of the expertise for determining model structure and the oversimplification of combining multimodal feature abstractions. MethodsIn this study, a multiple-fusion-layer based ensemble classifier of stacked autoencoder (MESAE) is proposed for recognizing emotions, in which the deep structure is identified based on a physiological-data-driven approach. Each SAE consists of three hidden layers to filter the unwanted noise in the physiological features and derives the stable feature representations. An additional deep model is used to achieve the SAE ensembles. The physiological features are split into several subsets according to different feature extraction approaches with each subset separately encoded by a SAE. The derived SAE abstractions are combined according to the physiological modality to create six sets of encodings, which are then fed to a three-layer, adjacent-graph-based network for feature fusion. The fused features are used to recognize binary arousal or valence states. ResultsDEAP multimodal database was employed to validate the performance of the MESAE. By comparing with the best existing emotion classifier, the mean of classification rate and F-score improves by 5.26%. ConclusionsThe superiority of the MESAE against the state-of-the-art shallow and deep emotion classifiers has been demonstrated under different sizes of the available physiological instances.

References

[1]
M. Soleymani, J. Lichtenauer, T. Pun, M. Pantic, A multi-modal database for affect recognition and implicit tagging, IEEE Trans. Affect. Comput., 3 (2012) 42-55.
[2]
G. Fanelli, J. Gall, H. Romsdorfer, T. Weise, L. Van Gool, A 3-D audio-visual corpus of affective communication, IEEE Trans. on Multimedia, 12 (2010) 591-598.
[3]
Z. Yin, J. Zhang, Operator functional state classification using least-square support vector machine based recursive feature elimination technique, Comput. Methods Prog. Biomed., 113 (2014) 101-115.
[4]
Z. Zeng, M. Pantic, G.I. Roisman, T.S. Huang, A survey of affect recognition methods: audio, visual, and spontaneous expressions, IEEE Trans. Pattern Anal. Mach. Intell., 31 (2009) 39-58.
[5]
J. Kim, E. Andre, Emotion recognition based on physiological changes in music listening, IEEE Trans. Pattern Anal. Mach. Intell., 30 (2008) 2067-2083.
[6]
A. Hanjalic, L.-Q. Xu, Affective video content representation and modeling, IEEE Trans. Multimedia, 7 (2005) 143-154.
[7]
C. Brunner, C. Vidaurre, M. Billinger, C. Neuper, A comparison of univariate, vector, bilinear autoregressive, and band power features for brain-computer interfaces, Med. Biol. Eng. Comput., 49 (2011) 1337-1346.
[8]
N. Birbaumer, Breaking the silence: brain-computer interfaces (BCI) for communication and motor control, Psychophysiology, 43 (2006) 517-532.
[9]
J. Zhang, Z. Yin, R. Wang, Recognition of mental workload levels under complex human-machine collaboration by using physiological features and adaptive support vector machines, IEEE Trans. Hum. Mach. Syst., 45 (2015) 200-214.
[10]
F. Agrafioti, D. Hatzinakos, A.K. Anderson, ECG pattern analysis for emotion detection, IEEE Trans. Affect. Comput., 3 (2012) 102-115.
[11]
R.W. Picard, E. Vyzas, J. Healey, Toward machine emotional intelligence: analysis of affective physiological state, IEEE Trans. Pattern Anal. Mach. Intell., 23 (2001) 1175-1191.
[12]
R. Cowie, E. Douglas-Cowie, Emotion recognition in human-computer interaction, IEEE Signal Process. Mag., 18 (2001) 32-80.
[13]
J. Kim, E. Andre, Emotion-specific dichotomous classification and feature-level fusion of multichannel biosignals for automatic emotion recognition, in: Proceedings of IEEE Inter-national Conference on Multisensor Fusion and Integration for Intelligent Systems, August 20-22, 2008.
[14]
H. Lee, A. Shackman, D. Jackson, R. Davidson, Test-retest reliability of voluntary emotion regulation, Psychophysiology, 46 (2009) 874-879.
[15]
I.C. Christie, B.H. Friedman, Autonomic specificity of discrete emotion and dimensions of affective space: a multivariate approach, Int. J. Psychophysiol., 51 (2004) 143-153.
[16]
G. Chanel, C. Rebetez, M. Btrancourt, T. Pun, Emotion assessment from physiological signals for adaptation of game difficulty, IEEE Trans. Syst. Man Cybern. A Syst. Humans, 41 (2011) 1052-1063.
[17]
O. AlZoubi, S.K. D'Mello, R.A. Calvo, Detecting naturalistic expressions of nonbasic affect using physiological signals, IEEE Trans. Affect. Comput., 3 (2012) 298-310.
[18]
L. Brown, B. Grundlehner, J. Penders, Towards wireless emotional valence detection from EEG, in: Proceedings of Engineering in Medicine and Biology Society, EMBC, 2011 Annual International Conference of the IEEE 2011.
[19]
R.J. Davidson, Affective neuroscience and psychophysiology: toward a synthesis, Psychophysiology, 40 (2003) 655-665.
[20]
Z. Yin, J. Zhang, Identification of temporal variations in mental workload using locally-linear-embedding -based EEG feature reduction and support-vector-machine-based clustering and classification technique, Comput. Methods Prog. Biomed., 115 (2014) 119-134.
[21]
G. Lee, M. Kwon, S. Kavuri-Sri, M. Lee, Emotion recognition based on 3D fuzzy visual and EEG features in movie clips, Neurocomputing, 144 (2014) 560-568.
[22]
J. Cannon, P.A. Krokhmal, Y. Chen, R. Murphey, Detection of temporal changes in psychophysiological data using statistical process control methods, Comput. Methods Prog. Biomed., 107 (2012) 367-381.
[23]
N. Martini, D. Menicucci, L. Sebastiani, R. Bedini, A. Pingitore, N. Vanello, A. Gemignani, The dynamics of EEG gamma responses to unpleasant visual stimuli: from local activity to functional connectivity, NeuroImage, 60 (2012) 922-932.
[24]
D. Nie, X.-W. Wang, L.-C. Shi, B.-L. Lu, EEG-based emotion recognition during watching movies, In: Proceedings of the 5th International IEEE/EMBS Conference on Neural Engineering (NER) 2011.
[25]
G. Borghini, L. Astolfi, G. Vecchiato, D. Mattia, F. Babiloni, Measuring neurophysiological signals in aircraft pilots and car drivers for the assessment of mental workload, fatigue and drowsiness, Neurosci. Biobehav. Rev., 44 (2014) 44.
[26]
G.K. Verma, U.S. Tiwary, Multimodal fusion framework: a multiresolution approach for emotion classification and recognition from physiological signal, NeuroImage, 102 (2014) 162-172.
[27]
M. Balconi, C. Lucchiari, EEG correlates (event-related desynchronization) of emotional face elaboration: a temporal analysis, Neurosci. Lett., 392 (2006) 118-123.
[28]
C.A. Frantzidis, C. Bratsas, C.L. Papadelis, E. Konstantinidis, C. Pappas, P.D. Bamidis, Toward emotion aware computing: an integrated approach using multichannel neurophysiological recordings and affective visual stimuli, IEEE Trans. Inf. Technol. Biomed., 14 (2010) 589-597.
[29]
M. Balconi, G. Mazza, Brain oscillations and BIS/BAS (behavioral inhibition/activation system) effects on processing masked emotional cues: ERS/ERD and coherence measures of alpha band, Int. J. Psychophysiol., 74 (2009) 158-165.
[30]
E.I. Konstantinidis, C.A. Frantzidis, C. Pappas, P.D. Bamidis, Real time emotion aware applications: a case study employing emotion evocative pictures and neuro-physiological sensing enhanced by graphic processor units, Comput. Methods Prog. Biomed., 107 (2012) 16-27.
[31]
M. Khezri, M. Firoozabadi, A.R. Sharafat, Reliable emotion recognition system based on dynamic adaptive fusion of forehead biopotentials and physiological signals, Comput. Methods Prog. Biomed., 122 (2015) 149-164.
[32]
J. Atkinson, D. Campos, Improving BCI-based emotion recognition by combining EEG feature selection and kernel classifiers, Expert. Syst. Appl., 47 (2016) 35-41.
[33]
S. Liu, D. Zhang, M. Xu, H. Qi, F. He, X. Zhao, P. Zhou, L. Zhang, D. Ming, Randomly dividing homologous samples leads to overinflated accuracies for emotion recognition, Int. J. Psychophysiol., 96 (2015) 29-37.
[34]
C. Li, C. Xu, Z. Feng, Analysis of physiological for emotion recognition with the IRS model, Neurocomputing, 178 (2016) 103-111.
[35]
J. Christensen, J. Estepp, G. Wilson, C. Russell, The effects of day-to-day variability of physiological data on operator functional state classification, NeuroImage, 59 (2012) 57-63.
[36]
F. Laurent, M. Valderrama, M. Besserve, M. Guillard, J.-P. Lachaux, J. Martinerie, G. Florence, Multimodal information improves the rapid detection of mental fatigue, Biomed. Signal. Process. Control., 8 (2013) 400-408.
[37]
Y. Shin, S. Lee, M. Ahn, H. Cho, S.C. Jun, H.-N. Lee, Simple adaptive sparse representation based classification schemes for EEG based braincomputer interface applications, Comput. Biol. Med., 66 (2015) 29-38.
[38]
D. Iacoviello, A. Petracca, M. Spezialetti, G. Placidi, A real-time classification algorithm for EEG-based BCI driven by self-induced emotions, Comput. Methods Prog. Biomed., 122 (2015) 293-303.
[39]
R.M. Mehmood, H.J. Lee, A novel feature extraction method based on late positive potential for emotion recognition in human brain signal patterns, Comput. Electr. Eng., 53 (2016) 444-457.
[40]
D. Wang, Y. Shang, Modeling physiological data with deep belief networks, Int. J. Inf. Educ. Technol., 3 (2013) 505-511.
[41]
X. Li, P. Zhang, D. Song, G. Yu, Y. Hou, B. Hu, EEG based emotion identification using unsupervised deep feature learning, in: SIGIR2015 Workshop on Neuro-Physiological Methods in IR Research, 13 August 2015.
[42]
K. Li, X. Li, Y. Zhang, A. Zhang, Affective state recognition from EEG with deep belief networks, in: IEEE 13th International Conference on Bioinformatics and Biomedicine, 2013.
[43]
X. Jia, K. Li, X. Li, A. Zhang, A Novel semi-supervised deep learning framework for affective state recognition on EEG signals, in: IEEE 14th International Conference on Bioinformatics and Bioengineering, 2014.
[44]
S. Jirayucharoensak, S. Pan-Ngum, P. Israsena, EEG-based emotion recognition using deep learning network with principal component based covariate shift adaptation, Scientific World J., 1 (2014).
[45]
E. Acar, F. Hopfgartner, S. Albayrak, Understanding affective content of music videos through learned representations, in: 20th Anniversary International Conference, January 2014.
[46]
J. Zhang, Z. Yin, R. Wang, Pattern classification of instantaneous cognitive task-load through GMM clustering, laplacian eigenmap and ensemble SVMs, IEEE/ACM Trans. Comput. Biol. Bioinf. (May 2016).
[47]
Y. Bengio, P. Lamblin, D. Popovici, H. Larochelle, Greedy layer-wise training of deep networks, Adv. Neural. Inf. Process. Syst., 19 (2007) 153-160.
[48]
S. Koelstra, C. Muehl, M. Soleymani, J.-S. Lee, A. Yazdani, T.E. Ebrahimi, T. Pun, A. Nijholt, I.Y. Patras, DEAP: a database for emotion analysis using physiological signals, IEEE Trans. Affect. Comput., 3 (2012) 18-31.
[49]
Y. Liu, X. Feng, Z. Zhou, Multimodal video classification with stacked contractive autoencoders, Sig. Process., 120 (2016) 761-766.
[50]
J. Li, Z. Struzik, L. Zhang, A. Cichocki, Feature learning from incomplete EEG with denoising autoencoder, Neurocomputing, 165 (2015) 23-31.
[51]
M. Belkin, P. Niyogi, Laplacian Eigenmaps and spectral techniques for embedding and clustering, Adv. Neural. Inf. Process. Syst. (2002) 585-591.
[52]
L. van der Maaten, E. Postma, J. van den Herik, Dimensionality Reduction, Tilburg University, 2009.
[53]
Y. Liu, O. Sourina, EEG-based valence level recognition for real-time applications, in: IEEE International Conference on Cyberworlds (CW), 2012, pp. 53-60.
[54]
D.S. Naser, G. Saha, Recognition of emotions induced by music videos using DT-CWPT, in: Indian Conference on Medical Informatics and Telemedicine (ICMIT) IEEE, 2013, pp. 53-57.
[55]
J. Chen, B. Hua, P. Moore, X. Zhang, X. Ma, Electroencephalogram-based emotion assessment system using ontology and data mining techniques, Appl. Soft. Comput., 30 (2015) 663-674.
[56]
H.J. Yoon, S.Y. Chung, EEG-based emotion estimation using Bayesian weighted-log-posterior function and perceptron convergence algorithm, Comput. Biol. Med., 43 (2013) 2230-2237.

Cited By

View all
  • (2025)Multimodal emotion recognition by fusing complementary patterns from central to peripheral neurophysiological signals across feature domainsEngineering Applications of Artificial Intelligence10.1016/j.engappai.2025.110004143:COnline publication date: 1-Mar-2025
  • (2024)pSCOUTER: Time-Series Emotion Classification Using Contactless Measured Multimodal Biosignals in Medical DiagnosisCompanion of the 2024 on ACM International Joint Conference on Pervasive and Ubiquitous Computing10.1145/3675094.3677582(46-50)Online publication date: 5-Oct-2024
  • (2024)AttX: Attentive Cross-Connections for Fusion of Wearable Signals in Emotion RecognitionACM Transactions on Computing for Healthcare10.1145/36537225:3(1-24)Online publication date: 18-Sep-2024
  • Show More Cited By
  1. Recognition of emotions using multimodal physiological signals and an ensemble deep learning model

    Recommendations

    Comments

    Please enable JavaScript to view thecomments powered by Disqus.

    Information & Contributors

    Information

    Published In

    cover image Computer Methods and Programs in Biomedicine
    Computer Methods and Programs in Biomedicine  Volume 140, Issue C
    March 2017
    298 pages

    Publisher

    Elsevier North-Holland, Inc.

    United States

    Publication History

    Published: 01 March 2017

    Author Tags

    1. Affective computing
    2. Deep learning
    3. Emotion recognition
    4. Ensemble learning
    5. Physiological signals

    Qualifiers

    • Research-article

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)0
    • Downloads (Last 6 weeks)0
    Reflects downloads up to 02 Mar 2025

    Other Metrics

    Citations

    Cited By

    View all
    • (2025)Multimodal emotion recognition by fusing complementary patterns from central to peripheral neurophysiological signals across feature domainsEngineering Applications of Artificial Intelligence10.1016/j.engappai.2025.110004143:COnline publication date: 1-Mar-2025
    • (2024)pSCOUTER: Time-Series Emotion Classification Using Contactless Measured Multimodal Biosignals in Medical DiagnosisCompanion of the 2024 on ACM International Joint Conference on Pervasive and Ubiquitous Computing10.1145/3675094.3677582(46-50)Online publication date: 5-Oct-2024
    • (2024)AttX: Attentive Cross-Connections for Fusion of Wearable Signals in Emotion RecognitionACM Transactions on Computing for Healthcare10.1145/36537225:3(1-24)Online publication date: 18-Sep-2024
    • (2024)Bridge Graph Attention Based Graph Convolution Network With Multi-Scale Transformer for EEG Emotion RecognitionIEEE Transactions on Affective Computing10.1109/TAFFC.2024.339487315:4(2042-2054)Online publication date: 30-Apr-2024
    • (2024)Hierarchical multimodal-fusion of physiological signals for emotion recognition with scenario adaption and contrastive alignmentInformation Fusion10.1016/j.inffus.2023.102129103:COnline publication date: 1-Mar-2024
    • (2024)“The reasons you believe …”Computer Methods and Programs in Biomedicine10.1016/j.cmpb.2024.108257254:COnline publication date: 1-Sep-2024
    • (2024)A Resilient Overlay for Human Emotion Recognition Using Mixed Frameworks in Machine-Human InteractionsSN Computer Science10.1007/s42979-024-02762-z5:4Online publication date: 8-Apr-2024
    • (2024)A lightweight deep neural network for detection of mental states from physiological signalsInnovations in Systems and Software Engineering10.1007/s11334-022-00470-620:3(405-412)Online publication date: 1-Sep-2024
    • (2023)Semisupervised Deep Features of Time-Frequency Maps for Multimodal Emotion RecognitionInternational Journal of Intelligent Systems10.1155/2023/36081152023Online publication date: 1-Jan-2023
    • (2023)Empirical study of emotional state recognition using multimodal fusion frameworkProceedings of the 2023 5th International Conference on Image, Video and Signal Processing10.1145/3591156.3591160(27-35)Online publication date: 24-Mar-2023
    • Show More Cited By

    View Options

    View options

    Figures

    Tables

    Media

    Share

    Share

    Share this Publication link

    Share on social media