[go: up one dir, main page]
More Web Proxy on the site http://driver.im/ skip to main content
10.5555/1838206.1838370acmotherconferencesArticle/Chapter ViewAbstractPublication PagesaamasConference Proceedingsconference-collections
research-article

Evaluating models of speaker head nods for virtual agents

Published: 10 May 2010 Publication History

Abstract

Virtual human research has often modeled nonverbal behaviors based on the findings of psychological research. In recent years, however, there have been growing efforts to use automated, data-driven approaches to find patterns of nonverbal behaviors in video corpora and even thereby discover new factors that have not been previously documented. However, there have been few studies that compare how the behaviors generated by different approaches are interpreted by people. In this paper, we present an evaluation study to compare the perception of nonverbal behaviors, more specifically head nods, generated by different approaches. Studies have shown that head nods serve a variety of communicative functions and that the head is in constant motion during speaking turns. To evaluate the different approaches of head nod generation, we asked human subjects to evaluate videos of a virtual agent displaying nods generated by a human, by a machine learning data-driven approach, or by a hand-crafted rule-based approach. Results show that there is a significant effect on the perception of head nods in terms of appropriate nod occurrence, especially between the data-driven approach and the rule-based approach.

References

[1]
K. Bergmann and S. Kopp. GNetic-using bayesian decision networks for iconic gesture generation. In Proceedings of the 9th International Conference on Intelligent Virtual Agents, pages 76--89, 2009.
[2]
W. Breitfuss, H. Prendinger, and M. Ishizuka. Automated generation of non-verbal behavior for virtual embodied characters. In ICMI '07: Proceedings of the 9th international conference on Multimodal interfaces, pages 319--322, New York, NY, USA, 2007. ACM.
[3]
C. Busso, Z. Deng, M. Grimm, U. Neumann, and S. Narayanan. Rigid head motion in expressive speech animation: Analysis and synthesis. IEEE Transactions on Audio, Speech and Language Processing, 15(3):1075--1086, 2007.
[4]
J. Cassell. More than just another pretty face: Embodied conversational interface agents. Communications of the ACM, 43:70--78, 2000.
[5]
J. Cassell, H. H. Vilhjálmsson, and T. Bickmore. BEAT: the behavior expression animation toolkit. In SIGGRAPH '01: Proceedings of the 28th annual conference on Computer graphics and interactive techniques, pages 477--486, New York, NY, USA, 2001. ACM.
[6]
U. Hadar, T. J. Steiner, and F. C. Rose. Head movement during listening turns in conversation. Journal of Nonverbal Behavior, 9(4):214--228, 1985.
[7]
D. Heylen. Challenges ahead: Head movements and other social acts in conversations. In AISB 2005, Social Presence Cues Symposium, 2005.
[8]
L. Huang, L.-P. Morency, and J. Gratch. Parasocial consensus sampling: Combining multiple perspectives to learn virtual human behavior. In AAMAS '10: Proceedings of the 9th international joint conference on Autonomous agents and multiagent systems. International Foundation for Autonomous Agents and Multiagent Systems, 2010.
[9]
A. Kendon. Some uses of the head shake. Gesture, 2:147--182(36), 2002.
[10]
M. Kipp, M. Neff, K. H. Kipp, and I. Albrecht. Towards natural gesture synthesis: Evaluating gesture units in a data-driven approach to gesture synthesis. In Proceedings of the 7th International Conference on Intelligent Virtual Agents, pages 15--28, 2007.
[11]
M. Knapp and J. Hall. Nonverbal Communication in Human Interaction. Harcourt Brace College Publishers, 4th edition, 1997.
[12]
F. L. A. Knoppel, A. S. Tigelaar, D. O. Bos, T. Alofs, and Z. Ruttkay. Trackside DEIRA: a dynamic engaging intelligent reporter agent. In AAMAS '08: Proceedings of the 7th international joint conference on Autonomous agents and multiagent systems, pages 112--119, Richland, SC, 2008. International Foundation for Autonomous Agents and Multiagent Systems.
[13]
S. Kopp, B. Krenn, S. Marsella, A. Marshall, C. Pelachaud, H. Pirker, K. Thorisson, and H. Vilhjálmsson. Towards a common framework for multimodal generation in embodied conversation agents: a behavior markup language. In Proceedings of 6th International Conference on Virtual Agents, pages 205--217, Marina del Rey, CA, USA, 2006.
[14]
J. Lee and S. Marsella. Nonverbal behavior generator for embodied conversational agents. In Proceedings of the 6th International Conference on Intelligent Virtual Agents, pages 243--255. Springer, 2006.
[15]
J. Lee and S. Marsella. Learning a model of speaker head nods using gesture corpora. In AAMAS '09: Proceedings of the 8th international joint conference on Autonomous agents and multiagent systems. International Foundation for Autonomous Agents and Multiagent Systems, 2009.
[16]
J. Lee, H. Prendinger, A. Neviarouskaya, and S. Marsella. Learning models of speaker head nods with affective information. In ACII '09: Proceedings of the 3rd International Conference on Affective Computing and Intelligent Interaction. International Foundation for Autonomous Agents and Multiagent Systems, 2009.
[17]
E. Z. McClave. Linguistic functions of head movements in the context of speech. Journal of Pragmatics, 32:855--878(24), June 2000.
[18]
A. Mignault and A. Chaudhuri. The many faces of a neutral face: Head tilt and perception of dominance and emotion. Journal of Nonverbal Behavior, 2(27):111--132, June 2003.
[19]
L.-P. Morency, I. de Kok, and J. Gratch. Predicting listener backchannels: A probabilistic multimodal approach. In Proceedings of the 8th International Conference on Intelligent Virtual Agents, pages 176--190, 2008.
[20]
K. G. Munhall, J. A. Jones, D. E. Callan, T. Kuratate, and E. Vatikiotis-Bateson. Visual prosody and speech intelligibility: Head movement improves auditory speech perception. Psychological Science, 15:133--137(5), February 2004.
[21]
A. Neviarouskaya, H. Prendinger, and M. Ishizuka. Textual affect sensing for sociable and expressive online communication. In ACII '07: Proceedings of the 2nd international conference on Affective Computing and Intelligent Interaction, pages 218--229, Berlin, Heidelberg, 2007. Springer-Verlag.
[22]
M. Strauss and M. Kipp. Eric: a generic rule-based framework for an affective embodied commentary agent. In L. Padgham, D. C. Parkes, J. Müller, and S. Parsons, editors, AAMAS '08: Proceedings of the 7th international joint conference on Autonomous agents and multiagent systems, pages 97--104, Richland, SC, 2008. International Foundation for Autonomous Agents and Multiagent Systems.
[23]
G. Tom, P. Pettersen, T. Lau, T. Burton, and J. Cook. The role of overt head movement in the formation of affect. Basic and Applied Social Psychology, 12(3):281--289, 1991.

Cited By

View all
  • (2018)Human, Chameleon or Nodding Dog?Proceedings of the 20th ACM International Conference on Multimodal Interaction10.1145/3242969.3242998(428-436)Online publication date: 2-Oct-2018
  • (2016)Teaching Social Communication Skills Through Human-Agent InteractionACM Transactions on Interactive Intelligent Systems10.1145/29377576:2(1-26)Online publication date: 3-Aug-2016

Index Terms

  1. Evaluating models of speaker head nods for virtual agents

    Recommendations

    Comments

    Please enable JavaScript to view thecomments powered by Disqus.

    Information & Contributors

    Information

    Published In

    cover image ACM Other conferences
    AAMAS '10: Proceedings of the 9th International Conference on Autonomous Agents and Multiagent Systems: volume 1 - Volume 1
    May 2010
    1578 pages
    ISBN:9780982657119

    Sponsors

    • IFAAMAS

    In-Cooperation

    Publisher

    International Foundation for Autonomous Agents and Multiagent Systems

    Richland, SC

    Publication History

    Published: 10 May 2010

    Check for updates

    Author Tags

    1. embodied conversational agents
    2. evaluation
    3. head nods
    4. nonverbal behaviors
    5. virtual agents

    Qualifiers

    • Research-article

    Conference

    AAMAS '10
    Sponsor:

    Acceptance Rates

    Overall Acceptance Rate 1,155 of 5,036 submissions, 23%

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)5
    • Downloads (Last 6 weeks)0
    Reflects downloads up to 08 Mar 2025

    Other Metrics

    Citations

    Cited By

    View all
    • (2018)Human, Chameleon or Nodding Dog?Proceedings of the 20th ACM International Conference on Multimodal Interaction10.1145/3242969.3242998(428-436)Online publication date: 2-Oct-2018
    • (2016)Teaching Social Communication Skills Through Human-Agent InteractionACM Transactions on Interactive Intelligent Systems10.1145/29377576:2(1-26)Online publication date: 3-Aug-2016

    View Options

    Login options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Figures

    Tables

    Media

    Share

    Share

    Share this Publication link

    Share on social media