R2-D2 Reloaded: A flexible sound synthesis system for sonic human-robot interaction design

A key skill for social robots is the ability to communicate their inner state to humans. In this paper, we explore abstracted robot-specific ways of interaction as an alternative to human-like or animal-like social cues. In particular, we present a sound system as a novel modality that extends a robot's ability for non-verbal communication. Unlike prior work which used pre-recorded audio samples to this end, we propose a flexible architecture with a generalized sound synthesizer that uses the principle of modulation to shape the sound in real-time by external and internal stimuli from the robot or the interaction. This allows for almost unlimited possibilities in the design of an expressive auditory social cue for human-robot interaction. We instantiate the architecture and report on example design choices for the sound synthesis principle, the real-time synthesizer, the sound modulation routings, and a sound sequence composer. We then demonstrate the system's ability for affect communication of primary and secondary emotions on a social robot.

[1]  M. Seif El-Nasr,et al.  A fuzzy emotional agent for decision-making in a mobile robot , 1998, 1998 IEEE International Conference on Fuzzy Systems Proceedings. IEEE World Congress on Computational Intelligence (Cat. No.98CH36228).

[2]  Cynthia Breazeal,et al.  Designing sociable robots , 2002 .

[3]  Brian R. Duffy,et al.  Anthropomorphism and the social robot , 2003, Robotics Auton. Syst..

[4]  S. L. See,et al.  The influence of voice pitch on the evaluation of a social robot receptionist , 2011, 2011 International Conference on User Science and Engineering (i-USEr ).

[5]  John Lasseter,et al.  Principles of traditional animation applied to 3D computer animation , 1987, SIGGRAPH.

[6]  Federico Fontana Preserving the Structure of the Moog VCF in the Digital Domain , 2007, ICMC.

[7]  Tetsuya Ogata,et al.  Voice quality manipulation for humanoid robots consistent with their head movements , 2009, 2009 9th IEEE-RAS International Conference on Humanoid Robots.

[8]  Leila Takayama,et al.  Judging a bot by its cover: An experiment on expectation setting for personal robots , 2010, 2010 5th ACM/IEEE International Conference on Human-Robot Interaction (HRI).

[9]  Jodi Forlizzi,et al.  The Snackbot: Documenting the design of a robot for long-term Human-Robot Interaction , 2009, 2009 4th ACM/IEEE International Conference on Human-Robot Interaction (HRI).

[10]  Elisabeth André,et al.  Creation and Evaluation of emotion expression with body movement, sound and eye color for humanoid robots , 2011, 2011 RO-MAN.

[11]  Scott S. Snibbe,et al.  Experiences with Sparky, a Social Robot , 2002 .

[12]  Nick Collins,et al.  The SuperCollider Book , 2011 .

[13]  Arika Okrent,et al.  In the Land of Invented Languages: Esperanto Rock Stars, Klingon Poets, Loglan Lovers, and the Mad Dreamers Who Tried to Build A Perfect Language , 2009 .

[14]  Tony Belpaeme,et al.  Situational Context Directs How People Affectively Interpret Robotic Non-Linguistic Utterances , 2014, 2014 9th ACM/IEEE International Conference on Human-Robot Interaction (HRI).

[15]  Kai Oliver Arras,et al.  Place-dependent people tracking , 2011, Int. J. Robotics Res..

[16]  Robin R. Murphy,et al.  Non-facial/non-verbal methods of affective expression as applied to robot-assisted victim assessment , 2007, 2007 2nd ACM/IEEE International Conference on Human-Robot Interaction (HRI).

[17]  Tony Belpaeme,et al.  People Interpret Robotic Non-linguistic Utterances Categorically , 2013, 2013 8th ACM/IEEE International Conference on Human-Robot Interaction (HRI).

[18]  Kai Oliver Arras,et al.  Robot-specific social cues in emotional body language , 2012, 2012 IEEE RO-MAN: The 21st IEEE International Symposium on Robot and Human Interactive Communication.

[19]  S. Wada,et al.  Time-frequency approach for analysis and synthesis of particular emotional voice , 2005, Proceedings of the Fifth IEEE International Symposium on Signal Processing and Information Technology, 2005..