Using listener gaze to augment speech generation in a virtual 3D environment

Using listener gaze to augment speech generation in a virtual 3D environment Maria Staudte Saarland University Alexander Koller University of Potsdam Abstract Listeners tend to gaze at objects to which they resolve referring expressions. We show that this remains true even when these objects are presented in a virtual 3D environment in which lis- teners can move freely. We further show that an automated speech generation system that uses eyetracking information to monitor listener’s understanding of referring expressions outperforms comparable systems that do not draw on listener gaze. Introduction In situated spoken interaction, there is evidence that the gaze of interlocutors can augment both language comprehension and production processes. For example, speaker gaze to ob- jects that are about to be mentioned (Griffin & Bock, 2000) has been shown to benefit listener comprehension by direct- ing listener gaze to the intended visual referents (Hanna & Brennan, 2007; Staudte & Crocker, 2011; Kreysa & Knoe- ferle, 2011). Even when speaker gaze is not visible to the listener, however, listeners are known to rapidly attend to mentioned objects (Tanenhaus, Spivey-Knowlton, Eberhard, & Sedivy, 1995). This gaze behavior on the part of listeners potentially provides speakers with useful feedback regarding the communicative success of their utterances: By monitor- ing listener gaze to objects in the environment, the speaker can determine whether or not a referring expression (RE) they have just produced was correctly understood or not, and po- tentially use this information to adjust subsequent production. In this paper we investigate the hypothesis that speaker use of listener gaze can potentially enhance interaction, even when situated in complex and dynamic scenes that simulate physical environments. In order to examine this hypothesis in a controlled and consistent manner, we monitor listener per- formance in the context of a computer system that generates spoken instructions to direct the listener through a 3D virtual environment with the goal of finding a trophy. Successful completion of the task requires listeners to press specific but- tons. Our experiment manipulated whether or not the com- puter system could follow up its original RE with feedback based on the listener’s gaze or movement behavior, with the aim of shedding light on the following two questions: • Do listener eye movements provide a consistent and useful indication of referential understanding, on a per-utterance basis, and when embedded in a dynamic and complex, goal-driven scenario? • What effect does gaze-based feedback have on listeners’ (gaze-)behavior and does it increase the more general ef- fectiveness of an interaction? We show that the listeners’ eye movements are a reliable predictor of referential understanding in our virtual environ- Konstantina Garoufi University of Potsdam Matthew Crocker Saarland University ments. A natural language generation (NLG) system, that ex- ploited this information to provide direct feedback, commu- nicated its intended referent to the listener more effectively than similar systems that did not draw on listener gaze. Gaze- based feedback was further shown to increase listener atten- tion to potential target objects in a scene, indicating a gen- erally more focused and task-oriented listener behavior. This system is, to our knowledge, the first NLG system that adjusts its referring expressions to listener gaze. Related work Previous research has shown that listeners align with speak- ers by visually attending to mentioned objects (Tanenhaus et al., 1995) and, if possible, to what the speaker attends to (Richardson & Dale, 2005; Hanna & Brennan, 2007; Staudte & Crocker, 2011). Little is known, however, about speaker adaptation to the listener’s (gaze) behavior, in particular when this occurs in dynamic and goal-oriented situations. Typi- cally, Visual World experiments have used simple and static visual scenes and disembodied utterances and have analyzed the recorded listener gaze off-line (e.g., Altmann & Kamide, 1999; Knoeferle, Crocker, Pickering, & Scheepers, 2005). Although studies involving an embodied speaker inherently include some dynamics in their stimuli, this is normally con- strained to speaker head and eye movements (Hanna & Bren- nan, 2007; Staudte & Crocker, 2011). Besides simplifying the physical environment to a static visual scene, none of these approaches can capture the reciprocal nature of interac- tion. That is, they do not take into account that the listeners’ eye movements may, as a signal of referential understanding to the speaker, change the speaker’s behavior and utterances on-line and, as such, affect the listener again. One study that emphasized interactive communication in a dynamic environment was conducted by Clark and Krych (2004). In this experiment, two partners assembled Lego models: The directing participant advised the building par- ticipant on how to achieve that goal. It was manipulated whether or not the director could see the builder’s workspace and, thus, use the builder’s visual attention as feedback for directions. Clark and Krych found, for instance, that the vis- ibility of the listener’s workspace led to significantly more deictic expressions by the speaker and to shorter task com- pletion times. However, the experimental setting introduced large variability in the dependent and independent variables, making controlled manipulation and fine-grained observa- tions difficult. In fact, we are not aware of any previ- ous work that has successfully integrated features of natu- ral environments—realistic, complex and dynamic scenes in which the visual salience of objects can change as a result of the listener’s moves in the environment—with the reciprocal

[1]  Takenobu Tokunaga,et al.  Multi-modal Reference Resolution in Situated Dialogue by Integrating Linguistic and Extra-Linguistic Clues , 2011, IJCNLP.

[2]  Alexander Koller,et al.  The Potsdam NLG systems at the GIVE-2.5 Challenge , 2011, ENLG.

[3]  M. Crocker,et al.  Investigating joint attention mechanisms through spoken human–robot interaction , 2011, Cognition.

[4]  Matthew W. Crocker,et al.  The influence of the immediate visual context on incremental thematic role-assignment: evidence from eye-movements in depicted events , 2005, Cognition.

[5]  Alexandre Denis Generating Referring Expressions with Reference Domain Theory , 2010, INLG.

[6]  Johanna D. Moore,et al.  The First Challenge on Generating Instructions in Virtual Environments , 2010, Empirical Methods in Natural Language Generation.

[7]  G. Altmann,et al.  Incremental interpretation at verbs: restricting the domain of subsequent reference , 1999, Cognition.

[8]  Mariët Theune,et al.  Report on the Second Second Challenge on Generating Instructions in Virtual Environments (GIVE-2.5) , 2011, ENLG.

[9]  Pia Knoeferle,et al.  Peripheral speaker gaze facilitates spoken language comprehension: syntactic structuring and thematic role asssignment in German , 2011 .

[10]  Alexander Koller,et al.  The GIVE-2 Corpus of Giving Instructions in Virtual Environments , 2010, LREC.

[11]  Mary Ellen Foster Enhancing Human-Computer Interaction with Embodied Conversational Agents , 2007, HCI.

[12]  Felix Hülsmann,et al.  Comparing gaze-based and manual interaction in a fast-paced gaming task in Virtual Reality , 2011 .

[13]  Zenzi M. Griffin,et al.  What the Eyes Say About Speaking , 2000, Psychological science.

[14]  H. H. Clark,et al.  Speaking while monitoring addressees for understanding , 2004 .

[15]  Marc Schröder,et al.  The German Text-to-Speech Synthesis System MARY: A Tool for Research, Development and Teaching , 2003, Int. J. Speech Technol..

[16]  K Naoko,et al.  Gaze matching of referring expressions in collaborative problem solving-Analysis based on the attributes of referring expressions - , 2011 .

[17]  Daniel C. Richardson,et al.  Looking To Understand: The Coupling Between Speakers' and Listeners' Eye Movements and Its Relationship to Discourse Comprehension , 2005, Cogn. Sci..

[18]  Julie C. Sedivy,et al.  Subject Terms: Linguistics Language Eyes & eyesight Cognition & reasoning , 1995 .

[19]  R. Baayen,et al.  Mixed-effects modeling with crossed random effects for subjects and items , 2008 .

[20]  S. Brennan,et al.  Speakers' eye gaze disambiguates referring expressions early during face-to-face conversation , 2007 .