[go: up one dir, main page]
More Web Proxy on the site http://driver.im/ skip to main content
10.5555/3237383.3237831acmconferencesArticle/Chapter ViewAbstractPublication PagesaamasConference Proceedingsconference-collections
research-article

Towards a Robust Interactive and Learning Social Robot

Published: 09 July 2018 Publication History

Abstract

Pepper is a humanoid robot, specifically designed for social interaction, that has been deployed in a variety of public environments. A programmable version of Pepper is also available, enabling our focused research on perception and behavior robustness and capabilities of an interactive social robot. We address Pepper perception by integrating state-of-the-art vision and speech recognition systems and experimentally analyzing their effectiveness. As we recognize limitations of the individual perceptual modalities, we introduce a multi-modality approach to increase the robustness of human social interaction with the robot. We combine vision, gesture, speech, and input from an onboard tablet, a remote mobile phone, and external microphones. Our approach includes the proactive seeking of input from a different modality, adding robustness to the failures of the separate components. We also introduce a learning algorithm to improve communication capabilities over time, updating speech recognition through social interactions. Finally, we realize the rich robot body-sensory data and introduce both a nearest-neighbor and a deep learning approach to enable Pepper to classify and speak up a variety of its own body motions. We view the contributions of our work to be relevant both to Pepper specifically and to other general social robots.

References

[1]
Marko Bjelonic. 2017. YOLO ROS: Real-Time Object Detection for ROS. (2017). Retrieved October 30, 2017 from https://github.com/leggedrobotics/darknet_ros
[2]
J. Bohren, R.b. Rusu, E. Gil Jones, E. Marder-Eppstein, C. Pantofaru, M. Wise, L. Mösenlechner, W. Meeussen, and S. Holzer. 2011. Towards autonomous robotic butlers: Lessons learned with the PR2 2011 IEEE International Conference on Robotics and Automation. 5568--5575.
[3]
Zhe Cao, Tomas Simon, Shih-En Wei, and Yaser Sheikh. 2017. Realtime Multi-Person 2D Pose Estimation using Part Affinity Fields CVPR.
[4]
Dragos Datcu, M Richert, T Roberti, W De Vries, and LJM Rothkrantz. 2004. AIBO Robot as a soccer and rescue game player. Proceedings of GAME-ON 2004 (2004), 45--49.
[5]
Can Erdogan and Manuela Veloso. 2011. Action Selection via Learning Behavior Patterns in Multi-Robot Systems Proceedings of the International Joint Conference on Artificial Intelligence. Barcelona, Spain.
[6]
Jonathan Gerbscheid, Thomas Groot, and Arnoud Visser. 2016. UvA@ Home 2017 standard platform proposal. simulation Vol. 8 (2016), 9.
[7]
Google. 2017. Google Cloud Speech. (2017). Retrieved November 9, 2017 from https://cloud.google.com/speech/
[8]
Nick Hawes, Chris Burbridge, Ferdian Jovan, Lars Kunze, Bruno Lacerda, Lenka Mudrová, Jay Young, Jeremy L. Wyatt, Denise Hebesberger, Tobias Körtner, Rares Ambrus, Nils Bore, John Folkesson, Patric Jensfelt, Lucas Beyer, Alexander Hermans, Bastian Leibe, Aitor Aldoma, Thomas Faulhammer, Michael Zillich, Markus Vincze, Muhannad Al-Omari, Eris Chinellato, Paul Duckworth, Yiannis Gatsoulis, David C. Hogg, Anthony G. Cohn, Christian Dondrup, Jaime Pulido Fentanes, Tomás Krajn'ık, Jo ao M. Santos, Tom Duckett, and Marc Hanheide. To Appear. The STRANDS Project: Long-Term Autonomy in Everyday Environments. IEEE Robotics and Automation Magazine (To Appear).
[9]
Sepp Hochreiter and Jürgen Schmidhuber. 1997. Long short-term memory. Neural computation, Vol. 9, 8 (1997), 1735--1780.
[10]
Thomas Kollar, Vittorio Perera, Daniele Nardi, and Manuela Veloso. 2013. Learning environmental knowledge from task-based human-robot dialog Robotics and Automation (ICRA), 2013 IEEE International Conference on. IEEE, 4304--4309.
[11]
Tsung-Yi Lin, Michael Maire, Serge Belongie, James Hays, Pietro Perona, Deva Ramanan, Piotr Dollár, and C Lawrence Zitnick. 2014. Microsoft coco: Common objects in context. In European conference on computer vision. Springer, 740--755.
[12]
Sharon Oviatt. 1999. Mutual Disambiguation of Recognition Errors in a Multimodel Architecture Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (CHI '99). ACM, New York, NY, USA, 576--583.
[13]
Vittorio Perera, Sai P Selveraj, Stephanie Rosenthal, and Manuela Veloso. 2016. Dynamic generation and refinement of robot verbalization Robot and Human Interactive Communication (RO-MAN), 2016 25th IEEE International Symposium on. IEEE, 212--218.
[14]
Vittorio Perera and Manuela Veloso. 2017. Learning to Understand Questions on the Task History of a Service Robot. (2017).
[15]
Luis A Pineda, Caleb Rascon, Gibran Fuentes, Arturo Rodrıguez, Hernando Ortega, Mauricio Reyes, Noé Hernández, Ricardo Cruz, Ivette Vélez, and Marco Ramırez. 2017. The Golem Team, RoboCup@ Home 2017. (2017).
[16]
Lerrel Pinto, Dhiraj Gandhi, Yuanfeng Han, Yong-Lae Park, and Abhinav Gupta. 2016. The Curious Robot: Learning Visual Representations via Physical Interactions. CoRR Vol. abs/1604.01360 (2016). {arxiv}1604.01360 http://arxiv.org/abs/1604.01360
[17]
Agence France Presse. 2016. Robot receptionists introduced at hospitals in Belgium. (2016). Retrieved November 15, 2017 from https://www.theguardian.com/technology/2016/jun/14/
[18]
Emil Protalinski. 2017. Google's speech recognition technology now has a 4.9% word-error-rate. (2017). Retrieved November 9, 2017 from https://venturebeat.com/2017/05/17/googles-speech-recognition-technology-now-has-a-4--9
[19]
Joseph Redmon, Santosh Divvala, Ross Girshick, and Ali Farhadi. 2016. You only look once: Unified, real-time object detection Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 779--788.
[20]
Joseph Redmon and Ali Farhadi. 2016. YOLO9000: Better, Faster, Stronger. arXiv preprint arXiv:1612.08242 (2016).
[21]
Shaoqing Ren, Kaiming He, Ross Girshick, and Jian Sun. 2015. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. arXiv preprint arXiv:1506.01497 (2015).
[22]
RoboCup. 2017. RoboCup@Home Leagues. (2017). Retrieved November 13, 2017 from https://www.robocup2017.org/eng/leagues_home.html
[23]
SoftBank Robotics. 2017a. NAOqi API and Pepper documentation. (2017). Retrieved October 30, 2017 from http://doc.aldebaran.com/2--5
[24]
SoftBank Robotics. 2017b. Who is Pepper? (2017). Retrieved November 1, 2017 from https://www.ald.softbankrobotics.com/en/robots/pepper
[25]
Stephanie Rosenthal, Sai P Selvaraj, and Manuela M Veloso. 2016. Verbalization: Narration of Autonomous Robot Experience. IJCAI. 862--868.
[26]
Michael Sipser. 2006. Introduction to the Theory of Computation. Vol. Vol. 2. Thomson Course Technology Boston.
[27]
Jörg Stückler, Ishrat Badami, David Droeschel, Kathrin Gr"ave, Dirk Holz, Manus McElhone, Matthias Nieuwenhuisen, Michael Schreiber, Max Schwarz, and Sven Behnke. 2013b. Nimbro@ home: Winning team of the robocup@ home competition 2012. RoboCup 2012: Robot Soccer World Cup XVI. Springer, 94--105.
[28]
Jörg Stückler, Ishrat Badami, David Droeschel, Kathrin Gr"ave, Dirk Holz, Manus McElhone, Matthias Nieuwenhuisen, Michael Schreiber, Max Schwarz, and Sven Behnke. 2013a. NimbRo@Home: Winning Team of the RoboCup@Home Competition 2012. Springer Berlin Heidelberg, Berlin, Heidelberg, 94--105.
[29]
Manuela Veloso, Nicholas Armstrong-Crews, Sonia Chernova, Elisabeth Crawford, Colin McMillen, Maayan Roth, Douglas Vail, and Stefan Zickler. 2008. A team of humanoid game commentators. International Journal of Humanoid Robotics Vol. 5, 03 (2008), 457--480.
[30]
Manuela M Veloso, Joydeep Biswas, Brian Coltin, and Stephanie Rosenthal. 2015. CoBots: Robust Symbiotic Autonomous Mobile Service Robots. IJCAI. 4423.
[31]
Sam Byford (The Verge). 2014. SoftBank announces emotional robots to staff its stores and watch your baby. (2014). Retrieved November 1, 2017 from https://www.theverge.com/2014/6/5/5781628/softbank-announces-pepper-robot

Cited By

View all

Recommendations

Comments

Please enable JavaScript to view thecomments powered by Disqus.

Information & Contributors

Information

Published In

cover image ACM Conferences
AAMAS '18: Proceedings of the 17th International Conference on Autonomous Agents and MultiAgent Systems
July 2018
2312 pages

Sponsors

In-Cooperation

Publisher

International Foundation for Autonomous Agents and Multiagent Systems

Richland, SC

Publication History

Published: 09 July 2018

Check for updates

Author Tags

  1. human-robot interaction
  2. pepper
  3. robot autonomy
  4. robot machine learning
  5. service robot
  6. social robot

Qualifiers

  • Research-article

Conference

AAMAS '18
Sponsor:
AAMAS '18: Autonomous Agents and MultiAgent Systems
July 10 - 15, 2018
Stockholm, Sweden

Acceptance Rates

AAMAS '18 Paper Acceptance Rate 149 of 607 submissions, 25%;
Overall Acceptance Rate 1,155 of 5,036 submissions, 23%

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • 0
    Total Citations
  • 276
    Total Downloads
  • Downloads (Last 12 months)7
  • Downloads (Last 6 weeks)0
Reflects downloads up to 10 Dec 2024

Other Metrics

Citations

Cited By

View all

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media