Towards modeling user behavior in human-machine interactions: Effect of Errors and Emotions

Data-driven approaches to spoken dialog strategy design rely on a sound understanding and modeling of user behavior in their interaction with machines. The spoken language usermachine communication channel is inherently noisy; noise in the channel may be due to errors in machine speech recognition, language understanding or other machine/user communication uncertainty and errors. Hence, annotation of human-machine dialogs needs to pay special attention to user behavior under errors and uncertainty. Another key aspect of user behavior is the dynamics of affect or emotions during an interaction relating to "how", contra "what", information is being conveyed. The goal of our work is on developing an account of user behavior working with annotated data from real human-machine mixed initiative dialogs. We illustrate the details of preparing the data for these needs using two case studies. First, we consider the DARPA Communicator dialog corpus to examine categories of error perception, user behavior under error, effect of user strategies on error recovery, and the role of user initiative in error situations. A conditional probability model smoothed by weighted ASR error rate is proposed. Second, we consider data and tagging needs for tracking the “emotional” aspects of human-machine interaction. Toward that end, we used data from a commercially deployed airline information systems and a WoZ study of child-machine interactions. Issues related to capturing emotionally salient information at the lexical and discourse level are highlighted in the context of automatic emotion recognition.

[1]  Laura M. Gonnerman,et al.  Cross-linguistic evidence for the early acquisition of discourse markers as register variables , 1999 .

[2]  Shrikanth Narayanan,et al.  Recognition of negative emotions from the speech signal , 2001, IEEE Workshop on Automatic Speech Recognition and Understanding, 2001. ASRU '01..

[3]  Shrikanth S. Narayanan,et al.  Combining acoustic and language information for emotion recognition , 2002, INTERSPEECH.

[4]  Jean Carletta,et al.  Assessing Agreement on Classification Tasks: The Kappa Statistic , 1996, CL.

[5]  Shrikanth S. Narayanan,et al.  Spoken dialog systems for children , 1998, Proceedings of the 1998 IEEE International Conference on Acoustics, Speech and Signal Processing, ICASSP '98 (Cat. No.98CH36181).

[6]  Victor Zue,et al.  JUPlTER: a telephone-based conversational interface for weather information , 2000, IEEE Trans. Speech Audio Process..

[7]  Shrikanth S. Narayanan,et al.  Politeness and frustration language in child-machine interactions , 2001, INTERSPEECH.

[8]  Andreas Stolcke,et al.  Prosody-based automatic detection of annoyance and frustration in human-computer dialog , 2002, INTERSPEECH.

[9]  Roberto Pieraccini,et al.  A stochastic model of human-machine interaction for learning dialog strategies , 2000, IEEE Trans. Speech Audio Process..

[10]  George N. Votsis,et al.  Emotion recognition in human-computer interaction , 2001, IEEE Signal Process. Mag..

[11]  K. Fischer,et al.  DESPERATELY SEEKING EMOTIONS OR: ACTORS, WIZARDS, AND HUMAN BEINGS , 2000 .

[12]  Andrew Ortony,et al.  The Cognitive Structure of Emotions , 1988 .

[13]  Marilyn A. Walker,et al.  AUTOMATIC PREDICTION OF PROBLEMATIC HUMAN-COMPUTER DIALO GUES IN 'HOW MAY I HELP YOU? , 1999 .

[14]  Gina-Anne Levow,et al.  Characterizing and Recognizing Spoken Corrections in Human-Computer Dialogue , 1998, ACL.

[15]  Marilyn A. Walker,et al.  The AT&t-DARPA communicator mixed-initiative spoken dialog system , 2000, INTERSPEECH.

[16]  Wayne H. Ward,et al.  THE CU COMMUNICATOR SYSTEM 1 , 1999 .

[17]  Sharon L. Oviatt Talking to thimble jellies: children²s conversational speech with animated characters , 2000, INTERSPEECH.

[18]  Shrikanth S. Narayanan,et al.  Creating conversational interfaces for children , 2002, IEEE Trans. Speech Audio Process..

[19]  Dani Byrd,et al.  Relations of sex and dialect to reduction , 1994, Speech Communication.

[20]  R. Plutchik The psychology and biology of emotion , 1994 .

[21]  Lynette Hirschman,et al.  Finding Errors Automatically in Semantically Tagged Dialogues , 2001, HLT.

[22]  Julia Hirschberg,et al.  Corrections in spoken dialogue systems , 2000, INTERSPEECH.