Detecting perceived quality of interaction with a robot using contextual features

This work aims to advance the state of the art in exploring the role of task, social context and their interdependencies in the automatic prediction of affective and social dimensions in human–robot interaction. We explored several SVMs-based models with different features extracted from a set of context logs collected in a human–robot interaction experiment where children play a chess game with a social robot. The features include information about the game and the social context at the interaction level (overall features) and at the game turn level (turn-based features). While overall features capture game and social context at the interaction level, turn-based features attempt to encode the dependencies of game and social context at each turn of the game. Results showed that game and social context-based features can be successfully used to predict dimensions of quality of interaction with the robot. In particular, overall features proved to perform equally or better than turn-based features, and game context-based features more effective than social context-based features. Our results show that the interplay between game and social context-based features, combined with features encoding their dependencies, lead to higher recognition performances for a subset of dimensions.

[1]  Radoslaw Niewiadomski,et al.  Smiling virtual agent in social context , 2012, Cognitive Processing.

[2]  Arvid Kappas,et al.  Mixing Implicit and Explicit Probes: Finding a Ground Truth for Engagement in Social Human-Robot Interactions , 2014, 2014 9th ACM/IEEE International Conference on Human-Robot Interaction (HRI).

[3]  Ana Paiva,et al.  Modelling empathic behaviour in a robotic game companion for children: An ethnographic study in real-world settings , 2012, 2012 7th ACM/IEEE International Conference on Human-Robot Interaction (HRI).

[4]  Paul Baxter,et al.  Child-robot interaction in the wild: advice to the aspiring experimenter , 2011, ICMI '11.

[5]  Caifeng Shan,et al.  Bodily Expression for Automatic Affect Recognition , 2015 .

[6]  Brian Scassellati,et al.  Active vision for sociable robots , 2001, IEEE Trans. Syst. Man Cybern. Part A.

[7]  Nicole C. Krämer,et al.  Living with a robot companion: empirical study on the interaction with an artificial health advisor , 2011, ICMI '11.

[8]  Ana Paiva,et al.  Modelling Empathy in Social Robotic Companions , 2011, UMAP Workshops.

[9]  Ana Paiva,et al.  Automatic analysis of affective postures and body motion to detect engagement with a game companion , 2011, 2011 6th ACM/IEEE International Conference on Human-Robot Interaction (HRI).

[10]  F. Biocca,et al.  The cyborg's dilemma: embodiment in virtual environments , 1997, Proceedings Second International Conference on Cognitive Technology Humanizing the Information Age.

[11]  I. Poggi Mind, hands, face and body. A goal and belief view of multimodal communication , 2007 .

[12]  Chih-Jen Lin,et al.  LIBSVM: A library for support vector machines , 2011, TIST.

[13]  James C. Lester,et al.  Modeling Learner Affect with Theoretically Grounded Dynamic Bayesian Networks , 2011, ACII.

[14]  Kostas Karpouzis,et al.  Investigating shared attention with a virtual agent using a gaze-based interface , 2010, Journal on Multimodal User Interfaces.

[15]  Ana Paiva,et al.  Empathic Robots for Long-term Interaction , 2014, Int. J. Soc. Robotics.

[16]  Iolanda Leite,et al.  Long-term interactions with empathic social robots , 2015, SIGAI.

[17]  Ana Paiva,et al.  Context-Sensitive Affect Recognition for a Robotic Game Companion , 2014, ACM Trans. Interact. Intell. Syst..

[18]  Ashish Kapoor,et al.  Multimodal affect recognition in learning environments , 2005, ACM Multimedia.

[19]  Ana Paiva,et al.  Inter-ACT: an affective and contextually rich multimodal video corpus for studying interaction with robots , 2010, ACM Multimedia.

[20]  Christoph Bartneck,et al.  Expressive robots in education: varying the degree of social supportive behavior of a robotic tutor , 2010, CHI.

[21]  Chih-Jen Lin,et al.  Combining SVMs with Various Feature Selection Strategies , 2006, Feature Extraction.

[22]  Nikolaos Mavridis,et al.  Opinions and attitudes toward humanoid robots in the Middle East , 2011, AI & SOCIETY.

[23]  Ana Paiva,et al.  Detecting user engagement with a robot companion using task and social interaction-based features , 2009, ICMI-MLMI '09.

[24]  Ana Paiva,et al.  Multimodal Affect Modeling and Recognition for Empathic Robot Companions , 2013, Int. J. Humanoid Robotics.

[25]  Bogdan Raducanu,et al.  Long-term socially perceptive and interactive robot companions: challenges and future perspectives , 2011, ICMI '11.

[26]  Ana Paiva,et al.  It's all in the game: Towards an affect sensitive and context aware game companion , 2009, 2009 3rd International Conference on Affective Computing and Intelligent Interaction and Workshops.

[27]  Caridakis,et al.  Body gesture and facial expression analysis for automatic affect recognition , 2010 .

[28]  Anton Nijholt,et al.  Brain computer interfaces as intelligent sensors for enhancing human-computer interaction , 2012, ICMI '12.

[29]  Jeffrey F. Cohn,et al.  Automatic detection of pain intensity , 2012, ICMI '12.

[30]  Zhihong Zeng,et al.  A Survey of Affect Recognition Methods: Audio, Visual, and Spontaneous Expressions , 2007, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[31]  C. Breazeal Role of expressive behaviour for robots that learn from people , 2009, Philosophical Transactions of the Royal Society B: Biological Sciences.

[32]  Peter Robinson,et al.  Challenges and Opportunities in Building Socially Intelligent Machines [Social Sciences] , 2011, IEEE Signal Processing Magazine.

[33]  Maurizio Mancini,et al.  Expressive Copying Behavior for Social Agents: A Perceptual Analysis , 2012, IEEE Transactions on Systems, Man, and Cybernetics - Part A: Systems and Humans.

[34]  Ana Paiva,et al.  Affect recognition for interactive companions: challenges and design in real world scenarios , 2009, Journal on Multimodal User Interfaces.

[35]  M. Mendelson,et al.  Measuring friendship quality in late adolescents and young adults: McGill Friendship Questionnaires. , 1999 .

[36]  Laurel D. Riek,et al.  Ibn Sina Steps Out: Exploring Arabic Attitudes Toward Humanoid Robots , 2010 .

[37]  Regan L. Mandryk,et al.  Using psychophysiological techniques to measure user experience with entertainment technologies , 2006, Behav. Inf. Technol..

[38]  B. Scassellati,et al.  Robots for use in autism research. , 2012, Annual review of biomedical engineering.

[39]  Maja J. Mataric,et al.  Recognition of spatial dynamics for predicting social interaction , 2011, 2011 6th ACM/IEEE International Conference on Human-Robot Interaction (HRI).

[40]  Georgios N. Yannakakis,et al.  Mining multimodal sequential patterns: a case study on affect detection , 2011, ICMI '11.

[41]  Nicu Sebe,et al.  Please, tell me about yourself: automatic personality assessment using short self-presentations , 2011, ICMI '11.

[42]  Vanessa Evers,et al.  The influence of social presence on acceptance of a companion robot by older people , 2008 .

[43]  Lætitia Gros,et al.  Combination of physiological and subjective measures to assess quality of experience for audiovisual technologies , 2011, 2011 Third International Workshop on Quality of Multimedia Experience.

[44]  Xue Yan,et al.  iCat: an animated user-interface robot with personality , 2005, AAMAS '05.

[45]  Nilanjan Sarkar,et al.  Operator Engagement Detection and Robot Behavior Adaptation in Human-Robot Interaction , 2005, Proceedings of the 2005 IEEE International Conference on Robotics and Automation.

[46]  Candace L. Sidner,et al.  Recognizing engagement in human-robot interaction , 2010, 2010 5th ACM/IEEE International Conference on Human-Robot Interaction (HRI).

[47]  Maja J. Mataric,et al.  Automated detection and classification of positive vs. negative robot interactions with children with autism using distance-based features , 2011, 2011 6th ACM/IEEE International Conference on Human-Robot Interaction (HRI).

[48]  Peter Robinson,et al.  Challenges and Opportunities in Building Socially Intelligent Machines , 2011 .

[49]  Tatsuya Nomura,et al.  The influence of people’s culture and prior experiences with Aibo on their attitude towards robots , 2006, AI & SOCIETY.

[50]  Nikolaos Mavridis,et al.  Towards an Empathizing and Adaptive Storyteller System , 2021, INT@AIIDE.

[51]  N. Borgers,et al.  Children as Respondents in Survey Research: Cognitive Development and Response Quality 1 , 2000 .

[52]  Ana Paiva,et al.  Are emotional robots more fun to play with? , 2008, RO-MAN 2008 - The 17th IEEE International Symposium on Robot and Human Interactive Communication.

[53]  Chih-Jen Lin,et al.  A Practical Guide to Support Vector Classication , 2008 .

[54]  Candace L. Sidner,et al.  Where to look: a study of human-robot engagement , 2004, IUI '04.

[55]  Ana Paiva,et al.  Using Anticipation to Create Believable Behaviour , 2006, AAAI.