default search action
ICMI 2022: Bengaluru, India
- Raj Tumuluri, Nicu Sebe, Gopal Pingali, Dinesh Babu Jayagopi, Abhinav Dhall, Richa Singh, Lisa Anthony, Albert Ali Salah:
International Conference on Multimodal Interaction, ICMI 2022, Bengaluru, India, November 7-11, 2022. ACM 2022, ISBN 978-1-4503-9390-4
Keynote Talks
- Louis-Philippe Morency:
What is Multimodal? 1 - Deb Roy:
Real Talk, Real Listening, Real Change. 2 - Daniel Gatica-Perez:
Focus on People: Five Questions from Human-Centered Computing. 3 - Justine Cassell:
The Future of the Body in Tomorrow's Workplace. 4
Session 1: Best Paper Nominees
- Yukiko I. Nakano, Eri Hirose, Tatsuya Sakato, Shogo Okada, Jean-Claude Martin:
Detecting Change Talk in Motivational Interviewing using Verbal and Facial Information. 5-14 - David Obremski, Helena Babette Hering, Paula Friedrich, Birgit Lugrin:
Exploratory Study on the Perception of Intelligent Virtual Agents With Non-Native Accents Using Synthetic and Natural Speech in German. 15-24 - Mimi Bocanegra, Mailin Lemke, Roelof Anne Jelle de Vries, Geke D. S. Ludden:
Commensality or Reverie in Eating? Exploring the Solo Dining Experience. 25-35 - Mingcheng Xue, Yu Liu, Kaiping Xu, Haiyang Zhang, Chengyang Yu:
Structured Multimodal Fusion Network for Referring Image Segmentation. 36-47 - Ritvik Agrawal, Shreyank Jyoti, Rohit Girmaji, Sarath Sivaprasad, Vineet Gandhi:
Does Audio help in deep Audio-Visual Saliency prediction models? 48-56
Session 2: Multimodal Machine Learning
- Li Zhou, Yan Luo:
A Spatio-temporal Learning for Music Conditioned Dance Generation. 57-62 - Apostolos Kalatzis, Vishnunarayan Girishan Prabhu, Saidur Rahman, Mike P. Wittie, Laura M. Stanley:
Emotions Matter: Towards Personalizing Human-System Interactions Using a Two-layer Multimodal Approach. 63-72 - Jicheng Li, Anjana Bhat, Roghayeh Barmaki:
Pose Uncertainty Aware Movement Synchrony Estimation via Spatial-Temporal Graph Transformer. 73-82 - Abhinav Joshi, Naman Gupta, Jinang Shah, Binod Bhattarai, Ashutosh Modi, Danail Stoyanov:
Generalized Product-of-Experts for Learning Multimodal Representations in Noisy Environments. 83-93 - Maria Tsfasman, Kristian Fenech, Morita Tarvirdians, András Lörincz, Catholijn M. Jonker, Catharine Oertel:
Towards creating a conversational memory for long-term meeting support: predicting memorable moments in multi-party conversations through eye-gaze. 94-104
Session 3: Health and Wellbeing
- Gauthier Robert Jean Faisandaz, Alix Goguey, Christophe Jouffrais, Laurence Nigay:
Keep in Touch: Combining Touch Interaction with Thumb-to-Finger µGestures for People with Visual Impairment. 105-116 - Jason Raether, Ehsanul Haque Nirjhar, Theodora Chaspari:
Evaluating Just-In-Time Vibrotactile Feedback for Communication Anxiety. 117-127 - Francisca Pessanha, Heysem Kaya, Alkim Almila Akdag Salah, Albert Ali Salah:
Towards using Breathing Features for Multimodal Estimation of Depression Severity. 128-138 - Floris Van Steijn, Gizem Sogancioglu, Heysem Kaya:
Text-based Interpretable Depression Severity Modeling via Symptom Predictions. 139-147 - Abhinandan Jain, Felix Schoeller, Emilie Zhang, Pattie Maes:
Frisson: Leveraging Metasomatic Interactions for Generating Aesthetic Chills. 148-158
Session 4: Conversation and Groups
- Xiang Zhi Tan, Elizabeth Jeanne Carter, Prithu Pareek, Aaron Steinfeld:
Group Formation in Multi-Robot Human Interaction During Service Scenarios. 159-169 - Stephanie Tan, David M. J. Tax, Hayley Hung:
Conversation Group Detection With Spatio-Temporal Context. 170-180 - Ruchen Wen, Brandon Barton, Sebastian Fauré, Tom Williams:
Unpretty Please: Ostensibly Polite Wakewords Discourage Politeness in both Robot-Directed and Human-Directed Communication. 181-190 - Kumar Shubham, Anirban Mukherjee, Dinesh Babu Jayagopi:
Review of realistic behavior and appearance generation in embodied conversational agents: A comparison between traditional and modern approaches. 191-197 - Tiffany D. Do, Mamtaj Akter, Zubin Datta Choudhary, Roger Azevedo, Ryan P. McMahan:
The Effects of an Embodied Pedagogical Agent's Synthetic Speech Accent on Learning Outcomes. 198-206
Session 5: Behavior Analytics and Gaze
- Maria Elena Lechuga Redondo, Radoslaw Niewiadomski, Francesco Rea, Alessandra Sciutti:
Comfortability Recognition from Visual Non-verbal Cues. 207-216 - Satchit Hari, Ajay, Sayan Sarcar, Sougata Sen, Surjya Ghosh:
AffectPro: Towards Constructing Affective Profile Combining Smartphone Typing Interaction and Emotion Self-reporting Pattern. 217-223 - Stephen Hutt, Sidney K. D'Mello:
Evaluating Calibration-free Webcam-based Eye Tracking for Gaze-based User Modeling. 224-235 - Bernd Dudzik, Hayley Hung:
Exploring the Detection of Spontaneous Recollections during Video-viewing In-the-Wild using Facial Behavior Analysis. 236-246 - Yihe Liu, Ziqi Yuan, Huisheng Mao, Zhiyun Liang, Wanqiuyue Yang, Yuanzhe Qiu, Tie Cheng, Xiaoteng Li, Hua Xu, Kai Gao:
Make Acoustic and Visual Cues Matter: CH-SIMS v2.0 Dataset and AV-Mixup Consistent Module. 247-258
Session 6: Interfaces and Interaction
- Shaun Alexander Macdonald, Frank E. Pollick, Stephen Anthony Brewster:
The Impact of Thermal Cues on Affective Responses to Emotionally Resonant Vibrations. 259-269 - Vivian Shen, Chris Harrison:
Pull Gestures with Coordinated Graphics on Dual-Screen Devices. 270-277 - Ramin Hedeshy, Chandan Kumar, Mike Lauer, Steffen Staab:
All Birds Must Fly: The Experience of Multimodal Hands-free Gaming with Gaze and Nonverbal Voice Synchronization. 278-287 - Ali Neshati, Aaron Salo, Shariff A. M. Faleel, Ziming Li, Hai-Ning Liang, Celine Latulipe, Pourang Irani:
EdgeSelect: Smartwatch Data Interaction with Minimal Screen Occlusion. 288-298 - Tiffany C. K. Kwok, Peter Kiefer, Martin Raubal:
Two-Step Gaze Guidance. 299-309
Posters
- Zhe Kong, Xin Wang, Neng Gao, Yifei Zhang, Yuhan Liu, Chenyang Tu:
Multi-level Fusion of Multi-modal Semantic Embeddings for Zero Shot Learning. 310-318 - Yoon Lee, Haoyu Chen, Guoying Zhao, Marcus Specht:
WEDAR: Webcam-based Attention Analysis via Attention Regulator Behavior Recognition with a Novel E-reading Dataset. 319-328 - Riku Arakawa, Mayank Goel, Chris Harrison, Karan Ahuja:
RGBDGaze: Gaze Tracking on Smartphones with RGB and Depth Data. 329-336 - Ayca Aygun, Boyang Lyu, Thuan Nguyen, Zachary Haga, Shuchin Aeron, Matthias Scheutz:
Cognitive Workload Assessment via Eye Gaze and EEG in an Interactive Multi-Modal Driving Task. 337-348 - Shun Katada, Shogo Okada, Kazunori Komatani:
Transformer-Based Physiological Feature Learning for Multimodal Analysis of Self-Reported Sentiment. 349-358 - Wenqing Wei, Sixia Li, Shogo Okada:
Investigating the relationship between dialogue and exchange-level impression. 359-367 - Jing-Xuan Zhang, Genshun Wan, Jia Pan:
Is Lip Region-of-Interest Sufficient for Lipreading? 368-372 - Zahra Vaseqi, Pengnan Fan, James J. Clark, Martin Levine:
A Framework for Video-Text Retrieval with Noisy Supervision. 373-383 - Juhyun Lee, Joosun Yum, Marvin Lee, Ji-Hyun Lee:
A cognitive knowledge-based system for hair and makeup recommendation based on facial features classification. 384-394 - Sandratra Rasendrasoa, Alexandre Pauchet, Julien Saunier, Sébastien Adam:
Real-Time Multimodal Emotion Recognition in Conversation for Multi-Party Interactions. 395-403 - Shreya Sharma, Mukesh K. Mohania:
Comparative Analysis of Entity Identification and Classification of Indian Epics. 404-413 - Gulshan Sharma, Pankaj Pandey, Ramanathan Subramanian, Krishna Prasad Miyapuram, Abhinav Dhall:
Neural Encoding of Songs is Modulated by Their Enjoyment. 414-419 - Francesco Tonini, Cigdem Beyan, Elisa Ricci:
Multimodal Across Domains Gaze Target Detection. 420-431 - Cassandra Scheirer, Chris Harrison:
DynaTags: Low-Cost Fiducial Marker Mechanisms. 432-443 - Marc Fraile, Christine Fawcett, Joakim Lindblad, Natasa Sladoje, Ginevra Castellano:
End-to-End Learning and Analysis of Infant Engagement During Guided Play: Prediction and Explainability. 444-454 - Mengyu Zhong, Vera van Zoest, Ayesha Mae Bilal, Fotios Papadopoulos, Ginevra Castellano:
Unimodal vs. Multimodal Prediction of Antenatal Depression from Smartphone-based Survey Data in a Longitudinal Study. 455-467 - Zhaobo Zheng, Kumar Akash, Teruhisa Misu, Vidya Krishnamoorthy, Miaomiao Dong, Yuni Lee, Gaojian Huang:
Identification of Adaptive Driving Style Preference through Implicit Inputs in SAE L2 Vehicles. 468-475 - Dan Bohus, Sean Andrist, Ashley Feniello, Nick Saw, Eric Horvitz:
Continual Learning about Objects in the Wild: An Interactive Approach. 476-486 - Alexandria K. Vail, Jeffrey M. Girard, Lauren M. Bylsma, Jeffrey F. Cohn, Jay Fournier, Holly Swartz, Louis-Philippe Morency:
Toward Causal Understanding of Therapist-Client Relationships: A Study of Language Modality and Social Entrainment. 487-494 - Ali N. Salman, Carlos Busso:
Privacy Preserving Personalization for Video Facial Expression Recognition Using Federated Learning. 495-503 - Heng Yang, Tao Luo, Yakun Zhang, Mingwu Song, Liang Xie, Ye Yan, Erwei Yin:
Improved Word-level Lipreading with Temporal Shrinkage Network and NetVLAD. 504-508 - Bharat Paudyal, Chris Creed, Ian Williams, Maite Frutos Pascual:
Inclusive Multimodal Voice Interaction for Code Navigation. 509-519 - Chongyang Bai, Maksim Bolonkin, Viney Regunath, V. S. Subrahmanian:
POLLY: A Multimodal Cross-Cultural Context-Sensitive Framework to Predict Political Lying from Videos. 520-530 - Kosmas Pinitas, Konstantinos Makantasis, Antonios Liapis, Georgios N. Yannakakis:
Supervised Contrastive Learning for Affect Modelling. 531-539 - Maha Elgarf, Sahba Zojaji, Gabriel Skantze, Christopher Peters:
CreativeBot: a Creative Storyteller robot to stimulate creativity in children. 540-548 - Radoslaw Niewiadomski, Gabriele De Lucia, Gabriele Grazzi, Maurizio Mancini:
Towards Commensal Activities Recognition. 549-557 - Nadine Wagener, Alex Ackermann, Gian-Luca Savino, Bastian Dänekas, Jasmin Niess, Johannes Schöning:
Influence of Passive Haptic and Auditory Feedback on Presence and Mindfulness in Virtual Reality Environments. 558-569 - Martin T. Schorradt, Douglas W. Cunningham:
Age Regression for Human Voices. 570-578 - Daria Joanna Hemmerling, Maciej Stroinski, Kamil Kwarciak, Krzysztof Trusiak, Maciej Szymkowski, Weronika Celniak, William Frier, Orestis Georgiou, Mykola Maksymenko:
Touchless touch with biosignal transfer for online communication. 579-590 - Marco Porta, Antonino Caminiti, Piercarlo Dondi:
GazeScale: Towards General Gaze-Based Interaction in Public Places. 591-596 - Liu Yang, Catherine Achard, Catherine Pelachaud:
Multimodal classification of interruptions in humans' interaction. 597-604 - Yufeng Yin, Jiashu Xu, Tianxin Zu, Mohammad Soleymani:
X-Norm: Exchanging Normalization Parameters for Bimodal Fusion. 605-614 - Robert G. Moulder, Nicholas D. Duran, Sidney K. D'Mello:
Assessing Multimodal Dynamics in Multi-Party Collaborative Interactions with Multi-Level Vector Autoregression. 615-625 - Neha Tarigopula, Sandrine Tornay, Skanda Muralidhar, Mathew Magimai-Doss:
Towards Accessible Sign Language Assessment and Learning. 626-631 - Vetha Vikashini Chithrra Raghuram, Hanan Salam, Jauwairia Nasir, Barbara Bruno, Oya Çeliktutan:
Personalized Productive Engagement Recognition in Robot-Mediated Collaborative Learning. 632-641 - Shravan Nayak, Christian Schuler, Debjoy Saha, Timo Baumann:
A Deep Dive Into Neural Synchrony Evaluation for Audio-visual Translation. 642-647
Blue Sky Papers
- Lik-Hang Lee, Carlos Bermejo Fernandez, Ahmad Yousef Alhilal, Tristan Braud, Simo Hosio, Esmée Henrieke Anne de Haas, Pan Hui:
Beyond the Blue Sky of Multimodal Interaction: A Centennial Vision of Interplanetary Virtual Spaces in Turn-based Metaverse. 648-652 - Eric Horvitz:
On the Horizon: Interactive and Compositional Deepfakes. 653-661 - Anqi Wang, Ze Gao, Lik Hang Lee, Tristan Braud, Pan Hui:
Decentralized, not Dehumanized in the Metaverse: Bringing Utility to NFTs through Multimodal Interaction. 662-667
Doctoral Consortium Papers
- Francisca Pessanha:
Non-verbal Signals in Oral History Archives. 668-672 - Maia Stiber:
Effective Human-Robot Collaboration via Generalized Robot Error Management Using Natural Human Responses. 673-678 - Tiffany Matej Hrkalovic:
Designing Hybrid Intelligence Techniques for Facilitating Collaboration Informed by Social Science. 679-684 - Nicola Corbellini:
Towards Human-Machine Collaboration: Multimodal Group Potency Estimation. 685-689 - Amr Gomaa:
Adaptive User-Centered Multimodal Interaction towards Reliable and Trusted Automotive Interfaces. 690-695 - Gulshan Sharma:
Physiological Sensing for Media Perception & Activity Recognition. 696-700 - Khalil J. Anderson:
Real-time Feedback for Developing Conversation Literacy. 701-704 - Auriane Boudin:
Interdisciplinary Corpus-based Approach for Exploring Multimodal Conversational Feedback. 705-710 - Soujanya Narayana:
Mood-Emotion Interplay: A Computational Perspective. 711-716 - Abhinav Joshi:
Multimodal Representation Learning For Real-World Applications. 717-723 - Yajing Feng:
Multimodal Representations and Assessments of Emotional Fluctuations of Speakers in Call Centers Conversations. 724-729
Demo and Exhibit Papers
- Masatoshi Hamanaka:
Sound Scope Pad: Controlling a VR Concert with Natural Movement. 730-732 - Denis Ivanko, Alexey M. Kashevnik, Dmitry Ryumin, Andrey Kitenko, Alexandr Axyonov, Igor Lashkov, Alexey Karpov:
MIDriveSafely: Multimodal Interaction for Drive Safely. 733-735
Grand Challenges
- Youngwoo Yoon, Pieter Wolfert, Taras Kucherenko, Carla Viegas, Teodor Nikolov, Mihail Tsakov, Gustav Eje Henter:
The GENEA Challenge 2022: A large evaluation of data-driven co-speech gesture generation. 736-747 - Khaled Saleh:
Hybrid Seq2Seq Architecture for 3D Co-Speech Gesture Generation. 748-752 - Naoshi Kaneko, Yuna Mitsubayashi, Geng Mu:
TransGesture: Autoregressive Gesture Generation with RNN-Transducer. 753-757 - Sicheng Yang, Zhiyong Wu, Minglei Li, Mengchen Zhao, Jiuxin Lin, Liyang Chen, Weihong Bao:
The ReprGesture entry to the GENEA Challenge 2022. 758-763 - Chi Zhou, Tengyue Bian, Kang Chen:
GestureMaster: Graph-based Speech-driven Gesture Generation. 764-770 - Jonathan Windle, David Greenwood, Sarah Taylor:
UEA Digital Humans entry to the GENEA Challenge 2022. 771-777 - Saeed Ghorbani, Ylva Ferstl, Marc-André Carbonneau:
Exemplar-based Stylized Gesture Generation from Speech: An Entry to the GENEA Challenge 2022. 778-783 - Che-Jui Chang, Sen Zhang, Mubbasir Kapadia:
The IVI Lab entry to the GENEA Challenge 2022 - A Tacotron2 Based Method for Co-Speech Gesture Generation With Locality-Constraint Attention Mechanism. 784-789 - Shuhong Lu, Andrew Feng:
The DeepMotion entry to the GENEA Challenge 2022. 790-796
Workshop Summaries
- Theodoros Kostoulas, Michal Muszynski, Leimin Tian, Edgar Roman-Rangel, Theodora Chaspari, Panos Amelidis:
Multimodal Affect and Aesthetic Experience. 797-798 - Pieter Wolfert, Taras Kucherenko, Carla Viegas, Zerrin Yumak, Youngwoo Yoon, Gustav Eje Henter:
GENEA Workshop 2022: The 3rd Workshop on Generation and Evaluation of Non-verbal Behaviour for Embodied Agents. 799-800 - Keith Curtis, George Awad, Shahzad Rajput, Ian Soboroff:
Second International Workshop on Deep Video Understanding. 801-802 - Bernd Dudzik, Dennis Küster, David St-Onge, Felix Putze:
The 4th Workshop on Modeling Socio-Emotional and Cognitive Processes from Multimodal Data In-the-Wild (MSECP-Wild). 803-804 - Hiroki Tanaka, Satoshi Nakamura, Kazuhiro Shidara, Jean-Claude Martin, Catherine Pelachaud:
3rd Workshop on Social Affective Multimodal Interaction for Health (SAMIH). 805-806 - Anika van der Klis, Heysem Kaya, Maryam Najafian, Saeid Safavi:
3rd ICMI Workshop on Bridging Social Sciences and AI for Understanding Child Behaviour. 807-809
manage site settings
To protect your privacy, all features that rely on external API calls from your browser are turned off by default. You need to opt-in for them to become active. All settings here will be stored as cookies with your web browser. For more information see our F.A.Q.