Estimating user's engagement from eye-gaze behaviors in human-agent conversations

In face-to-face conversations, speakers are continuously checking whether the listener is engaged in the conversation and change the conversational strategy if the listener is not fully engaged in the conversation. With the goal of building a conversational agent that can adaptively control conversations with the user, this study analyzes the user's gaze behaviors and proposes a method for estimating whether the user is engaged in the conversation based on gaze transition 3-gram patterns. First, we conduct a Wizard-of-Oz experiment to collect the user's gaze behaviors. Based on the analysis of the gaze data, we propose an engagement estimation method that detects the user's disengagement gaze patterns. The algorithm is implemented as a real-time engagement-judgment mechanism and is incorporated into a multimodal dialogue manager in a conversational agent. The agent estimates the user's conversational engagement and generates probing questions when the user is distracted from the conversation. Finally, we conduct an evaluation experiment using the proposed engagement-sensitive agent and demonstrate that the engagement estimation function improves the user's impression of the agent and the interaction with the agent. In addition, probing performed with proper timing was also found to have a positive effect on user's verbal/nonverbal behaviors in communication with the conversational agent.

[1]  E. Bard,et al.  Limited visual control of the intelligibility of speech in face-to-face dialogue , 1997, Perception & psychophysics.

[2]  M. Argyle,et al.  Gaze and Mutual Gaze , 1994, British Journal of Psychiatry.

[3]  David R. Traum,et al.  Modelling Grounding and Discourse Obligations Using Update Rules , 2000, ANLP.

[4]  David G. Novick,et al.  Coordinating turn-taking with gaze , 1996, Proceeding of Fourth International Conference on Spoken Language Processing. ICSLP '96.

[5]  Mitsuru Ishizuka,et al.  Attentive Presentation Agents , 2007, IVA.

[6]  A. Kendon Some functions of gaze-direction in social interaction. , 1967, Acta psychologica.

[7]  Christopher E. Peters Direction of Attention Perception for Conversation Initiation in Virtual Environments , 2005, IVA.

[8]  Shumin Zhai,et al.  Conversing with the user based on eye-gaze patterns , 2005, CHI.

[9]  Joyce Yue Chai,et al.  What's in a gaze?: the role of eye-gaze in reference resolution in multimodal conversational interfaces , 2008, IUI '08.

[10]  Yukiko I. Nakano,et al.  Towards a Model of Face-to-Face Grounding , 2003, ACL.

[11]  S. Duncan,et al.  On the structure of speaker–auditor interaction during speaking turns , 1974, Language in Society.

[12]  Eric Horvitz,et al.  Learning to Predict Engagement with a Spoken Dialog System in Open-World Settings , 2009, SIGDIAL Conference.

[13]  H. H. Clark,et al.  Understanding by addressees and overhearers , 1989, Cognitive Psychology.

[14]  Candace L. Sidner,et al.  Explorations in engagement for humans and robots , 2005, Artif. Intell..

[15]  M. Argyle,et al.  The Different Functions of Gaze , 1973 .

[16]  Michael Argyle,et al.  The central Europe experiment: Looking at persons and looking at objects , 1976 .

[17]  Louis-Philippe Morency,et al.  Predicting Listener Backchannels: A Probabilistic Multimodal Approach , 2008, IVA.

[18]  Catherine Pelachaud,et al.  Modelling Gaze Behaviour for Conversational Agents , 2003, IVA.

[19]  Toyoaki Nishida,et al.  Attentional Behaviors as Nonverbal Communicative Signals in Situated Interactions with Conversational Agents , 2007 .

[20]  Trevor Darrell,et al.  Head gestures for perceptual interfaces: The role of context in improving recognition , 2007, Artif. Intell..

[21]  Toyoaki Nishida,et al.  Converting Text into Agent Animations: Assigning Gestures to Text , 2004, HLT-NAACL.

[22]  A. Kendon Conducting Interaction: Patterns of Behavior in Focused Encounters , 1990 .

[23]  Candace L. Sidner,et al.  Where to look: a study of human-robot engagement , 2004, IUI '04.

[24]  Stacy Marsella,et al.  Virtual Rapport , 2006, IVA.

[25]  Johan Bos,et al.  TrindiKit 1.0 manual , 1999 .

[26]  Yukiko I. Nakano,et al.  Estimating User's Conversational Engagement Based on Gaze Behaviors , 2008, IVA.