Online Behavior Evaluation with the Switching Wizard of Oz

Advances in animation and sensor technology allow us to engage in face-to-face conversations with virtual agents [1]. One major challenge is to generate the virtual agent's appropriate, human-like behavior contingent with that of the human conversational partner. Models of (nonverbal) behavior are pre-dominantly learned from corpora of dialogs between human subjects [2], or based on simple observations from literature (e.g. [3,4,5,6])

[1]  J. Beskow,et al.  MushyPeek: A Framework for Online Investigation of Audiovisual Dialogue Phenomena , 2009, Language and speech.

[2]  C. Pelachaud,et al.  Generating Listening Behaviour , 2011 .

[3]  Fabio Pianesi,et al.  Introduction to the special issue on multimodal corpora for modeling human multimodal behavior , 2008, Lang. Resour. Evaluation.

[4]  Dirk Heylen,et al.  Backchannel Strategies for Artificial Listeners , 2010, IVA.

[5]  Louis-Philippe Morency,et al.  Virtual Rapport 2.0 , 2011, IVA.

[6]  Dirk Heylen,et al.  Backchannels: Quantity, Type and Timing Matters , 2011, IVA.

[7]  Dirk Heylen,et al.  A rule-based backchannel prediction model using pitch and pause information , 2010, INTERSPEECH.

[8]  Jessica K. Hodgins,et al.  The saliency of anomalies in animated human characters , 2010, TAP.

[9]  Jeremy N. Bailenson,et al.  Detecting digital chameleons , 2008, Comput. Hum. Behav..

[10]  Louis-Philippe Morency,et al.  A probabilistic multimodal approach for predicting listener backchannels , 2009, Autonomous Agents and Multi-Agent Systems.

[11]  Carol O'Sullivan,et al.  Talking bodies: Sensitivity to desynchronization of conversations , 2009, TAP.

[12]  Jean Carletta,et al.  A shallow model of backchannel continuers in spoken dialogue , 2003 .

[13]  Nigel G. Ward,et al.  Prosodic features which cue back-channel responses in English and Japanese , 2000 .

[14]  Dirk Heylen,et al.  Iterative perceptual learning for social behavior synthesis , 2013, Journal on Multimodal User Interfaces.