Combining Verbal and Nonverbal Features to Overcome the “Information Gap” in Task-Oriented Dialogue

Dialogue act modeling in task-oriented dialogue poses significant challenges. It is particularly challenging for corpora consisting of two interleaved communication streams: a dialogue stream and a task stream. In such corpora, information can be conveyed implicitly by the task stream, yielding a dialogue stream with seemingly missing information. A promising approach leverages rich resources from both the dialog and the task streams, combining verbal and non-verbal features. This paper presents work on dialogue act modeling that leverages body posture, which may be indicative of particular dialogue acts. Combining three information sources (dialogue exchanges, task context, and users' posture), three types of machine learning frameworks were compared. The results indicate that some models better preserve the structure of task-oriented dialogue than others, and that automatically recognized postural features may help to disambiguate user dialogue moves.

[1]  Kristy Elizabeth Boyer,et al.  Characterizing the Effectiveness of Tutorial Dialogue with Hidden Markov Models , 2010, Intelligent Tutoring Systems.

[2]  Barbara Di Eugenio,et al.  Dialogue Act Classification, Instance-Based Learning, and Higher Order Dialogue Structure , 2010, Dialogue Discourse.

[3]  Shafiq R. Joty,et al.  Unsupervised Modeling of Dialog Acts in Asynchronous Conversations , 2011, IJCAI.

[4]  J. Burgoon,et al.  Nonverbal Communication , 2018, Encyclopedia of Evolutionary Psychological Science.

[5]  Andreas Stolcke,et al.  Dialogue act modeling for automatic tagging and recognition of conversational speech , 2000, CL.

[6]  M. Knapp,et al.  Nonverbal communication in human interaction , 1972 .

[7]  Kristy Elizabeth Boyer,et al.  Analyzing Posture and Affect in Task-Oriented Tutoring , 2012, FLAIRS Conference.

[8]  William M. Pottenger,et al.  Posting Act Tagging Using Transformation-Based Learning , 2005, Foundations of Data Mining and knowledge Discovery.

[9]  Kristy Elizabeth Boyer,et al.  Dialogue Act Modeling in a Complex Task-Oriented Domain , 2010, SIGDIAL Conference.

[10]  Federica Cavicchio,et al.  The Modulation of Cooperation and Emotion in Dialogue: The REC Corpus , 2009, ACL.

[11]  J. Russell,et al.  Facial and vocal expressions of emotion. , 2003, Annual review of psychology.

[12]  Shrikanth S. Narayanan,et al.  Combining lexical, syntactic and prosodic cues for improved online dialog act tagging , 2009, Comput. Speech Lang..

[13]  Edward Ivanovic,et al.  Automatic instant messaging dialogue using statistical models and dialogue acts , 2008 .

[14]  Eric N. Forsyth Improving automated lexical and discourse analysis of online chat dialog , 2007 .

[15]  Alan Ritter,et al.  Unsupervised Modeling of Twitter Conversations , 2010, NAACL.

[16]  Gina-Anne Levow,et al.  Dialog act tagging with support vector machines and hidden Markov models , 2006, INTERSPEECH.

[17]  Jacob Cohen A Coefficient of Agreement for Nominal Scales , 1960 .

[18]  Timothy Baldwin,et al.  Classifying Dialogue Acts in One-on-One Live Chats , 2010, EMNLP.

[19]  Arthur C. Graesser,et al.  Mining Bodily Patterns of Affective Experience during Learning , 2010, EDM.

[20]  Ana Paiva,et al.  Automatic analysis of affective postures and body motion to detect engagement with a game companion , 2011, 2011 6th ACM/IEEE International Conference on Human-Robot Interaction (HRI).

[21]  Kristy Elizabeth Boyer,et al.  An Affect-Enriched Dialogue Act Classification Model for Task-Oriented Dialogue , 2011, ACL.

[22]  Gwyneth Doherty-Sneddon,et al.  The Reliability of a Dialogue Structure Coding Scheme , 1997, CL.

[23]  Ashish Kapoor,et al.  Automatic prediction of frustration , 2007, Int. J. Hum. Comput. Stud..

[24]  Beverly Park Woolf,et al.  Affect-aware tutors: recognising and responding to student affect , 2009, Int. J. Learn. Technol..