Spoken language, in addition to serving as a primary vehicle for externalizing linguistic structures and meaning, acts as a carrier of various sources of information, including background, age, gender, membership in social structures, as well as physiological, pathological and emotional states. These sources of information are more than just ancillary to the main purpose of linguistic communication: Humans react to the various non-linguistic factors encoded in the speech signal, shaping and adjusting their interactions to satisfy interpersonal and social protocols.
Computer science, artificial intelligence and computational linguistics have devoted much active research to systems that aim to model the production and recovery of linguistic lexico-semantic structures from speech. However, less attention has been devoted to systems that model and understand the paralinguistic and extralinguistic information in the signal. As the breadth and nature of human-computer interaction escalates to levels previously reserved for human-to-human communication, there is a growing need to endow computational systems with human-like abilities which facilitate the interaction and make it more natural. Of paramount importance amongst these is the human ability to make inferences regarding the affective content of our exchanges.
This thesis proposes a framework for the recognition of affective qualifiers from prosodic-acoustic parameters extracted from spoken language. It is argued that modeling the affective prosodic variation of speech can be approached by integrating acoustic parameters from various prosodic time scales, summarizing information from more localized (e.g., syllable level) to more global prosodic phenomena (e.g., utterance level). In this framework speech is structurally modeled as a dynamically evolving hierarchical model in which levels of the hierarchy are determined by prosodic constituency and contain parameters that evolve according to dynamical systems. The acoustic parameters have been chosen to reflect four main components of speech thought to reflect paralinguistic and affect-specific information: intonation, loudness, rhythm and voice quality. The thesis addresses the contribution of each of these components separately, and evaluates the full model by testing it on datasets of acted and of spontaneous speech perceptually annotated with affective labels, and by comparing it against human performance benchmarks. (Copies available exclusively from MIT Libraries, Rm. 14-0551, Cambridge, MA 02139-4307. Ph. 617-253-5668; Fax 617-253-1690.)
Cited By
- Li Y Semi-Supervised Learning for Multimodal Speech and Emotion Recognition Proceedings of the 2021 International Conference on Multimodal Interaction, (817-821)
- Mottelson A and Hornbæk K An affect detection technique using mobile commodity sensors in the wild Proceedings of the 2016 ACM International Joint Conference on Pervasive and Ubiquitous Computing, (781-792)
- Li Y, Contreras J and Salazar L Predicting Voice Elicited Emotions Proceedings of the 21th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, (1969-1978)
- Reyes-Vargas M, Sánchez-Gutiérrez M, Rufiner L, Albornoz M, Vignolo L, Martínez-Licona F and Goddard-Close J Hierarchical Clustering and Classification of Emotions in Human Speech Using Confusion Matrices Proceedings of the 15th International Conference on Speech and Computer - Volume 8113, (162-169)
- Hennig S Candidacy of physiological measurements for implicit control of emotional speech synthesis Proceedings of the 4th international conference on Affective computing and intelligent interaction - Volume Part II, (208-215)
- Chang K, Fisher D, Canny J and Hartmann B How's my mood and stress? Proceedings of the 6th International Conference on Body Area Networks, (71-77)
- Eyben F, Wöllmer M and Schuller B Opensmile Proceedings of the 18th ACM international conference on Multimedia, (1459-1462)
- Álvarez A, Cearreta I, López J, Arruti A, Lazkano E, Sierra B and Garay N Application of feature subset selection based on evolutionary algorithms for automatic emotion recognition in speech Proceedings of the 2007 international conference on Advances in nonlinear speech processing, (273-281)
- Inanoglu Z and Caneel R Emotive alert Proceedings of the 10th international conference on Intelligent user interfaces, (251-253)
- Vemuri S and Bender W (2004). Next-Generation Personal Memory Aids, BT Technology Journal, 22:4, (125-138), Online publication date: 1-Oct-2004.
Index Terms
- A computational model for the automatic recognition of affect in speech
Recommendations
Cued Speech automatic recognition in normal-hearing and deaf subjects
This article discusses the automatic recognition of Cued Speech in French based on hidden Markov models (HMMs). Cued Speech is a visual mode which, by using hand shapes in different positions and in combination with lip patterns of speech, makes all the ...