Crafting the illusion of meaning: template-based specification of embodied conversational behavior

Templates are a widespread natural language technology that achieves believability within a narrow range of interaction and coverage. We consider templates for embodied conversational behavior. Such templates combine a specific pattern of marked-up text, specifying prosody and conversational signals as well as words, with similarly annotated gaps that can be filled in by rule to yield a coherent contribution to a dialogue with a user. In this paper we argue that templates can give a designer substantial freedom to realize specific combinations of behaviors in interactions with users and thereby to explore the relationships among such factors as emotion, personality, individuality and social role.

[1]  Justine Cassell,et al.  Embodiment in conversational characters: Rea , 1999, CHI 1999.

[2]  Justine Cassell,et al.  BEAT: the Behavior Expression Animation Toolkit , 2001, Life-like characters.

[3]  Stefan Kopp,et al.  Model-based animation of co-verbal gesture , 2002, Proceedings of Computer Animation 2002 (CA 2002).

[4]  Alexander Kain,et al.  OGIresLPC: Diphone synthesizer using residual-excited linear prediction , 1997 .

[5]  Nicole Chovil Discourse‐oriented facial displays in conversation , 1991 .

[6]  Paul Taylor,et al.  Festival Speech Synthesis System , 1998 .

[7]  Stefan Kopp,et al.  A Communicative Mediator in a Virtual Environment: Processing of Multimodal Input and Output , 2001 .

[8]  Michael White,et al.  EXEMPLARS: A Practical, Extensible Framework For Dynamic Text Generation , 1998, INLG.

[9]  Catherine Pelachaud,et al.  Eye Communication in a Conversational 3D Synthetic Agent , 2000, AI Commun..

[10]  Mari Ostendorf,et al.  TOBI: a standard for labeling English prosody , 1992, ICSLP.

[11]  Ehud Reiter,et al.  Book Reviews: Building Natural Language Generation Systems , 2000, CL.

[12]  Ken Perlin Noise, hypertexture, antialiasing, and gesture , 2003 .

[13]  Joseph Weizenbaum,et al.  and Machine , 1977 .

[14]  C. Creider Hand and Mind: What Gestures Reveal about Thought , 1994 .

[15]  Matthew Stone,et al.  Making discourse visible: coding and animating conversational facial displays , 2002, Proceedings of Computer Animation 2002 (CA 2002).

[16]  Ken Perlin,et al.  Layered compositing of facial expression , 1997, SIGGRAPH '97.

[17]  Catherine Pelachaud,et al.  Subtleties of facial expressions in embodied agents , 2002, Comput. Animat. Virtual Worlds.

[18]  Terry Winograd,et al.  Understanding computers and cognition - a new foundation for design , 1987 .

[19]  J. Proudfoot,et al.  Noise , 1931, The Indian medical gazette.

[20]  Mariët Theune,et al.  From data to speech : language generation in context , 2000 .

[21]  Ehud Reiter,et al.  NLG vs. Templates , 1995, ArXiv.

[22]  Jakob Nielsen,et al.  Usability engineering , 1997, The Computer Science and Engineering Handbook.

[23]  Scott A. King,et al.  A facial model and animation techniques for animated speech , 2001 .

[24]  Igor S. Pandzic,et al.  Conversational virtual character for the Web , 2002, Proceedings of Computer Animation 2002 (CA 2002).

[25]  André Vellino T. Winograd and F. Flores, Understanding Computers and Cognition: A New Foundation for Design , 1987, Artif. Intell..

[26]  Justine Cassell,et al.  Embodied conversational interface agents , 2000, CACM.

[27]  Phoebe Sengers,et al.  Designing Comprehensible Agents , 1999, IJCAI.

[28]  Norman I. Badler,et al.  Representing and parameterizing agent behaviors , 2002, Proceedings of Computer Animation 2002 (CA 2002).

[29]  Joseph Bates,et al.  Virtual Reality, Art, and Entertainment , 1992, Presence: Teleoperators & Virtual Environments.

[30]  J. Cassell,et al.  Towards a model of technology and literacy development: Story listening systems , 2004 .

[31]  Mark Steedman,et al.  Generating Facial Expressions for Speech , 1996, Cogn. Sci..

[32]  Carl Machover,et al.  Virtual reality , 1994, IEEE Computer Graphics and Applications.

[33]  K. Chang,et al.  Embodiment in conversational interfaces: Rea , 1999, CHI '99.

[34]  Michael M. Cohen,et al.  Modeling Coarticulation in Synthetic Visual Speech , 1993 .

[35]  G. Ayers,et al.  Guidelines for ToBI labelling , 1994 .

[36]  Justine Cassell,et al.  Negotiated Collusion: Modeling Social Language and its Relationship Effects in Intelligent Agents , 2003, User Modeling and User-Adapted Interaction.

[37]  Michael D. McNeese,et al.  Assessment of User Affective and Belief States for Interface Adaptation: Application to an Air Force Pilot Task , 2002, User Modeling and User-Adapted Interaction.

[38]  W. Lewis Johnson,et al.  Animated Agents for Procedural Training in Virtual Reality: Perception, Cognition, and Motor Control , 1999, Appl. Artif. Intell..