The roles of haptic-ostensive referring expressions in cooperative, task-based human-robot dialogue

Generating referring expressions is a task that has received a great deal of attention in the natural-language generation community, with an increasing amount of recent effort targeted at the generation of multimodal referring expressions. However, most implemented systems tend to assume very little shared knowledge between the speaker and the hearer, and therefore must generate fully-elaborated linguistic references. Some systems do include a representation of the physical context or the dialogue context; however, other sources of contextual information are not normally used. Also, the generated references normally consist only of language and, possibly, deictic pointing gestures. When referring to objects in the context of a task-based interaction involving jointly manipulating objects, a much richer notion of context is available, which permits a wider range of referring options. In particular, when conversational partners cooperate on a mutual task in a shared environment, objects can be made accessible simply by manipulating them as part of the task. We demonstrate that such expressions are common in a corpus of human-human dialogues based on constructing virtual objects, and then describe how this type of reference can be incorporated into the output of a humanoid robot that engages in similar joint construction dialogues with a human partner.

[1]  Paul Piwek,et al.  Modality Choice for Generation of Referring Acts: Pointing versus Describing , 2007 .

[2]  Mira Ariel The function of accessibility in a theory of grammar , 1991 .

[3]  Harry Bunt,et al.  Multimodal referece. Studies in automatic generation of multimodal referring expressions , 2000 .

[4]  Laurent Romary,et al.  Referring to objects with spoken and haptic modalities , 2002, Proceedings. Fourth IEEE International Conference on Multimodal Interfaces.

[5]  Christopher Gauker Thinking Out Loud , 1994 .

[6]  Alois Knoll,et al.  Integrating Language, Vision and Action for Human Robot Dialog Systems , 2007, HCI.

[7]  John D. Kelleher,et al.  Incremental Generation of Spatial Referring Expressions in Situated Dialog , 2006, ACL.

[8]  Daniel C. Richardson,et al.  The Art of Conversation Is Coordination , 2007, Psychological science.

[9]  Alois Knoll,et al.  Human-Robot dialogue for joint construction tasks , 2006, ICMI '06.

[10]  Ielka van der Sluis,et al.  Building a Semantically Transparent Corpus for the Generation of Referring Expressions. , 2006, INLG.

[11]  Mary Ellen Foster,et al.  Following Assembly Plans in Cooperative, Task-Based Human-Robot Dialogue , 2008 .

[12]  Robbert-Jan Beun,et al.  Object reference in a shared domain of conversation , 1998 .

[13]  Robert Dale,et al.  Computational Interpretations of the Gricean Maxims in the Generation of Referring Expressions , 1995, Cogn. Sci..

[14]  Carolyn Penstein Rosé,et al.  Modeling the impact of shared visual information on collaborative reference , 2007, CHI.

[15]  Jean Carletta,et al.  Eyetracking for two-person tasks with manipulation of a virtual world , 2010, Behavior research methods.

[16]  D. Byron Understanding Referring Expressions in Situated Language Some Challenges for Real-World Agents Donna , 2003 .

[17]  Matthew P. Aylett,et al.  Referential form, word duration, and modelling the listener in spoken dialogue , 2004 .

[18]  Ipke Wachsmuth,et al.  Incremental Generation of Multimodal Deixis Referring to Objects , 2005, ENLG.

[19]  Ipke Wachsmuth,et al.  Deictic object reference in task-oriented dialogue , 2006 .

[20]  M. Tanenhaus,et al.  Approaches to studying world-situated language use : bridging the language-as-product and language-as-action traditions , 2005 .