Using Social and Physiological Signals for User Adaptation in Conversational Agents

In face-to-face communication, humans subconsciously emit social signals which are picked up and used by their interlocutors as feedback for how well the previously communicated messages have been received. The feedback is then used in order to adapt the way the coming messages are being produced and sent to the interlocutor, leading to the communication to become as efficient and enjoyable as possible. Currently however, it is rare to find conversational agents utilizing this feedback channel for altering how the multimodal output is produced during interactions with users, largely due to the complex nature of the problem. In most regards, humans have a significant advantage over conversational agents in interpreting and acting on social signals. Humans are however restricted to a limited set of sensors, "the five senses", which conversational agents are not. This makes it possible for conversational agents to use specialized sensors to pick up physiological signals, such as skin temperature, respiratory rate or pupil dilation, which carry valuable information about the user with respect to the conversation. This thesis work aims at developing methods for utilizing both social and physiological signals emitted by humans in order to adapt the output of the conversational agent, allowing for an increase in conversation quality. These methods will primarily be based on automatically learning adaptive behavior from examples of real human interactions using machine learning methods.

[1]  Patrick G. T. Healey,et al.  Human, Chameleon or Nodding Dog? , 2018, ICMI.

[2]  Stephen D. Laycock,et al.  Predicting Head Pose in Dyadic Conversation , 2017, IVA.

[3]  Ben J. A. Kröse,et al.  The Influence of Social Presence on Acceptance of an Assistive Social Robot and Screen Agent by Elderly Users , 2009 .

[4]  Mamun Bin Ibne Reaz,et al.  A Review of Smart Homes—Past, Present, and Future , 2012, IEEE Transactions on Systems, Man, and Cybernetics, Part C (Applications and Reviews).

[5]  Natalie Sommer,et al.  Workload-driven modulation of mixed-reality robot-human communication , 2018, MCPMD@ICMI.

[6]  L. O. Svaasand,et al.  Remote plethysmographic imaging using ambient light. , 2008, Optics express.

[7]  J. Broekens,et al.  Assistive social robots in elderly care: a review , 2009 .

[8]  Dimosthenis Kontogiorgos,et al.  FARMI: A FrAmework for Recording Multi-Modal Interactions , 2018, LREC.

[9]  S. Kopp,et al.  Towards Adaptive Social Behavior Generation for Assistive Robots Using Reinforcement Learning , 2017, 2017 12th ACM/IEEE International Conference on Human-Robot Interaction (HRI.

[10]  Elisabeth André,et al.  How to Shape the Humor of a Robot - Social Behavior Adaptation Based on Reinforcement Learning , 2018, ICMI.

[11]  Helen F. Hastie,et al.  Symptoms of cognitive load in interactions with a dialogue system , 2018, MCPMD@ICMI.

[12]  Stefan Wermter,et al.  Learning Empathy-Driven Emotion Expressions using Affective Modulations , 2018, 2018 International Joint Conference on Neural Networks (IJCNN).

[13]  Maja Pantic,et al.  Social signal processing: Survey of an emerging domain , 2009, Image Vis. Comput..

[14]  F. Vetere,et al.  Cognitive Heat , 2017 .

[15]  Brian Scassellati,et al.  A thermal emotion classifier for improved human-robot interaction , 2016, 2016 25th IEEE International Symposium on Robot and Human Interactive Communication (RO-MAN).

[16]  Dimosthenis Kontogiorgos,et al.  Crowdsourced Multimodal Corpora Collection Tool , 2018, LREC.

[17]  Carlos Busso,et al.  Novel Realizations of Speech-Driven Head Movements with Generative Adversarial Networks , 2018, 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).

[18]  Heather Pon-Barry,et al.  Effects of voice-adaptation and social dialogue on perceptions of a robotic learning companion , 2016, 2016 11th ACM/IEEE International Conference on Human-Robot Interaction (HRI).

[19]  Omar Mubin,et al.  Adaptive Social Robot for Sustaining Social Engagement during Long-Term Children–Robot Interaction , 2017, Int. J. Hum. Comput. Interact..

[20]  Hai Xuan Pham,et al.  End-to-end Learning for 3D Facial Animation from Speech , 2018, ICMI.

[21]  Ana Paiva,et al.  Modelling Empathy in Social Robotic Companions , 2011, UMAP Workshops.

[22]  J. Bailenson,et al.  Digital Chameleons , 2005, Psychological science.

[23]  Jaap Ham,et al.  Trusting Digital Chameleons: The Effect of Mimicry by a Virtual Social Agent on User Trust , 2013, PERSUASIVE.

[24]  Ionut Damian,et al.  Natural interaction with culturally adaptive virtual characters , 2012, Journal on Multimodal User Interfaces.

[25]  Susan R. Fussell,et al.  Effects of adaptive robot dialogue on information exchange and social relations , 2006, HRI '06.

[26]  Barbara Deml,et al.  Estimating mental load in passive and active tasks from pupil and gaze changes using bayesian surprise , 2018, MCPMD@ICMI.

[27]  Taku Komura,et al.  A Deep Learning Framework for Character Motion Synthesis and Editing , 2016, ACM Trans. Graph..

[28]  Joakim Gustafson,et al.  Machine Learning and Social Robotics for Detecting Early Signs of Dementia , 2017, ArXiv.

[29]  Jorge Dias,et al.  User-Adaptive Interaction in Social Robots: A Survey Focusing on Non-physical Interaction , 2018, Int. J. Soc. Robotics.

[30]  Yuko Ohno,et al.  Measurement of Respiration Rate and Depth Through Difference in Temperature Between Skin Surface and Nostril by Using Thermal Image , 2017, MedInfo.

[31]  Jaakko Lehtinen,et al.  Audio-driven facial animation by joint end-to-end learning of pose and emotion , 2017, ACM Trans. Graph..

[32]  Jennifer Healey,et al.  Toward Machine Emotional Intelligence: Analysis of Affective Physiological State , 2001, IEEE Trans. Pattern Anal. Mach. Intell..

[33]  Ginevra Castellano,et al.  Adaptive Robotic Tutors that Support Self-Regulated Learning: A Longer-Term Investigation with Primary School Children , 2018, Int. J. Soc. Robotics.

[34]  Fang Chen,et al.  Cognitive Load Measurement from User's Linguistic Speech Features for Adaptive Interaction Design , 2009, INTERACT.