Co-expressivity of Speech and Gesture: Lessons for Models of Aligned Speech and Gesture Production

When people combine language and gesture to convey their intended information, both modalities are characterized by an intriguing degree of coherence and consistency. For developing an account how speech and gesture are aligned to each other, one question of major importance is how meaning is distributed across the two channels. In this paper, we start from recent empirical findings indicating a flexible interaction between both systems and show that psycholinguistic models of speech and gesture production in literature cannot account for this interplay equally well. Based on a discussion of these theories as well as current computational approaches, we point out conclusions as to how a production model must be designed in order to simulate aligned, human-like multimodal behavior in virtual expressive agents.

[1]  Stefan Kopp,et al.  Synthesizing multimodal utterances for conversational agents , 2004, Comput. Animat. Virtual Worlds.

[2]  Stefan Kopp,et al.  Towards integrated microplanning of language and iconic gesture for multimodal output , 2004, ICMI '04.

[3]  B. Butterworth,et al.  Iconic gestures, imagery, and word retrieval in speech , 1997 .

[4]  Hao Yan,et al.  Coordination and context-dependence in the generation of embodied conversation , 2000, INLG.

[5]  Sotaro Kita,et al.  What does cross-linguistic variation in semantic coordination of speech and gesture reveal? Evidence for an interface representation of spatial thinking and speaking , 2003 .

[6]  Maurizio Mancini,et al.  Implementing Expressive Gesture Synthesis for Embodied Conversational Agents , 2005, Gesture Workshop.

[7]  Janet Beavin Bavelas,et al.  An experimental study of when and how speakers use gestures to communicate , 2002 .

[8]  Judith Holler,et al.  How iconic gestures and speech interact in the representation of meaning: Are both aspects really integral to the process? , 2003 .

[9]  Stefan Kopp,et al.  Trading Spaces: How Humans and Humanoids Use Speech and Gesture to Give Directions , 2007 .

[10]  M. Denis The description of routes : A cognitive approach to the production of spatial discourse , 1997 .

[11]  M. Studdert-Kennedy Hand and Mind: What Gestures Reveal About Thought. , 1994 .

[12]  Stefan Kopp,et al.  Verbal or Visual? How Information is Distributed across Speech and Gesture in Spatial Dialog , 2006 .

[13]  Justine Cassell,et al.  BEAT: the Behavior Expression Animation Toolkit , 2001, Life-like characters.

[14]  Yihsiu Chen,et al.  Language and Gesture: Lexical gestures and lexical access: a process model , 2000 .

[15]  W. Levelt,et al.  Speaking: From Intention to Articulation , 1990 .

[16]  S. Nobe Language and Gesture: Where do most spontaneous representational gestures actually occur with respect to speech? , 2000 .

[17]  De Ruiter,et al.  Gesture and speech production , 1998 .

[18]  M. Pickering,et al.  Toward a mechanistic psychology of dialogue , 2004, Behavioral and Brain Sciences.

[19]  J. D. Ruiter The production of gesture and speech , 2000 .