A Methodology for the Automatic Extraction and Generation of Non-Verbal Signals Sequences Conveying Interpersonal Attitudes

In many applications, Embodied Conversational Agents (ECAs) must be able to express various affects such as emotions or social attitudes. Non-verbal signals, such as smiles or gestures, contribute to the expression of attitudes. Social attitudes affect the whole behavior of a person: they are “characteristic of an affective style that colors the entire interaction” [1] . Moreover, recent findings have demonstrated that non-verbal signals are not interpreted in isolation but along with surrounding signals. Non-verbal behavior planning models designed to allow ECAs to express attitudes should thus consider complete sequences of non-verbal signals and not only signals independently of one another. However, existing models do not take this into account, or in a limited manner. The contribution of this paper is a methodology for the automatic extraction of sequences of non-verbal signals characteristic of a social phenomenon from a multimodal corpus, and a non-verbal behavior planning model that takes into account sequences of non-verbal signals rather than signals independently. This methodology is applied to design a virtual recruiter capable of expressing social attitudes, which is then evaluated in and out of an interaction context.

[1]  T. Leary Interpersonal diagnosis of personality , 1957 .

[2]  F. Atelsek,et al.  Firo: A Three-Dimensional Theory of Interpersonal Behavior , 1959 .

[3]  A. Kendon Some functions of gaze-direction in social interaction. , 1967, Acta psychologica.

[4]  P. Ekman,et al.  The Repertoire of Nonverbal Behavior: Categories, Origins, Usage, and Coding , 1969 .

[5]  M. Knapp,et al.  Nonverbal communication in human interaction , 1972 .

[6]  S. Goeppert [Speech behavior and neurotic personality structure (author's transl)]. , 1976, MMW, Munchener medizinische Wochenschrift.

[7]  Allan Mazur,et al.  Culture and the perception of social dominance from facial expression. , 1981 .

[8]  Jerold L. Hale,et al.  The fundamental topoi of relational communication , 1984 .

[9]  Jerold L. Hale,et al.  Relational Messages Associated with Nonverbal Behaviors. , 1984 .

[10]  R. Rosenthal,et al.  Rapport expressed through nonverbal behavior , 1985 .

[11]  J. Aronoff,et al.  The recognition of threatening facial stimuli. , 1988, Journal of personality and social psychology.

[12]  L. Smith-Lovin,et al.  INTERRUPTIONS IN GROUP DISCUSSIONS: THE EFFECTS OF GENDER AND GROUP COMPOSITION* , 1989 .

[13]  R. Gifford Mapping Nonverbal Behavior on the Interpersonal Circle , 1991 .

[14]  M. Argyle,et al.  Gaze and Mutual Gaze , 1994, British Journal of Psychiatry.

[15]  Ramakrishnan Srikant,et al.  Fast Algorithms for Mining Association Rules in Large Databases , 1994, VLDB.

[16]  E. Otta,et al.  The effect of smiling and of head tilting on person perception. , 1994, The Journal of psychology.

[17]  Donald W. Hine,et al.  The Role of Verbal Behavior in the Encoding and Decoding of Interpersonal Dispositions , 1994 .

[18]  M. Studdert-Kennedy Hand and Mind: What Gestures Reveal About Thought. , 1994 .

[19]  Robert Gifford,et al.  A lens-mapping framework for understanding the encoding and decoding of interpersonal dispositions in nonverbal behavior. , 1994 .

[20]  Judith A. Hall,et al.  Beliefs about female and male nonverbal communication , 1995 .

[21]  D. Keltner Signs of appeasement: evidence for the distinct displays of embarrassment, amusement, and shame , 1995 .

[22]  B. Knutson Facial expressions of emotion influence interpersonal trait inferences , 1996 .

[23]  Ramakrishnan Srikant,et al.  Mining Sequential Patterns: Generalizations and Performance Improvements , 1996, EDBT.

[24]  Mark Steedman,et al.  Generating Facial Expressions for Speech , 1996, Cogn. Sci..

[25]  Catherine Pelachaud,et al.  Performative faces , 1998, Speech Commun..

[26]  D. Keltner,et al.  Social Functions of Emotions at Four Levels of Analysis , 1999 .

[27]  J. Burgoon,et al.  Nonverbal cues and interpersonal judgments: Participant and observer perceptions of intimacy, dominance, composure, and formality , 1999 .

[28]  J. Dillard,et al.  The sounds of dominance: Vocal precursors of perceived dominance during interpersonal influence , 2000 .

[29]  U. Hess,et al.  The Influence of Facial Emotion Displays, Gender, and Ethnicity on Judgments of Dominance and Affiliation , 2000 .

[30]  Marco Costa,et al.  Head Canting in Paintings: An Historical Study , 2001 .

[31]  A Meta-Analysis,et al.  Dominance as Expressed and Inferred Through Speaking Time , 2002 .

[32]  Norihiro Hagita,et al.  Messages embedded in gaze of interface agents --- impression management with agent's gaze , 2002, CHI.

[33]  Paul Boersma,et al.  Praat, a system for doing phonetics by computer , 2002 .

[34]  A. Chaudhuri,et al.  The Many Faces of a Neutral Face: Head Tilt and Perception of Dominance and Emotion , 2003 .

[35]  I. Poggi MIND MARKERS , 2003 .

[36]  Justine Cassell,et al.  BEAT: the Behavior Expression Animation Toolkit , 2001, Life-like characters.

[37]  D. Ballin,et al.  A framework for interpersonal attitude and non-verbal communication in improvisational visual media production , 2004 .

[38]  K. Scherer What are emotions? And how can they be measured? , 2005 .

[39]  Thomas V. Merluzzi,et al.  Cognitive assessment of social anxiety: Development and validation of a self-statement questionnaire , 1982, Cognitive Therapy and Research.

[40]  Norah E. Dunbar,et al.  Perceptions of power and interactional dominance in interpersonal relationships , 2005 .

[41]  Judith A. Hall,et al.  Beliefs about the nonverbal expression of social power , 2005 .

[42]  Timothy W. Bickmore,et al.  Establishing and maintaining long-term human-computer relationships , 2005, TCHI.

[43]  Paulo J. Azevedo,et al.  Protein Sequence Classification Through Relevant Sequence Mining and Bayes Classifiers , 2005, EPIA.

[44]  Dirk Heylen,et al.  Listening Heads , 2006, ZiF Workshop.

[45]  Maurizio Mancini,et al.  Du corpus vidéo à l'agent expressif. Utilisation des différents niveaux de représentation multimodale et émotionnelle , 2006, Rev. d'Intelligence Artif..

[46]  Anne Laurent,et al.  Sequential patterns for text categorization , 2006, Intell. Data Anal..

[47]  Hatice Gunes,et al.  Creating and Annotating Affect Databases from Face and Body Display: A Contemporary Survey , 2006, 2006 IEEE International Conference on Systems, Man and Cybernetics.

[48]  Stacy Marsella,et al.  Virtual Rapport , 2006, IVA.

[49]  Peter Wittenburg,et al.  ELAN: a Professional Framework for Multimodality Research , 2006, LREC.

[50]  Céline Loscos,et al.  Expressing Complex Mental States Through Facial Expressions , 2007, ACII.

[51]  John W. Du Bois The stance triangle , 2007 .

[52]  Reginald B. Adams,et al.  Looking at You or Looking Elsewhere: The Influence of Head Orientation on the Signal Value of Emotional Facial Expressions , 2007 .

[53]  Ursula Hess,et al.  Display of Empathy and Perception of Out-Group Members , 2007 .

[54]  Zhigang Deng,et al.  Data-Driven 3D Facial Animation , 2007 .

[55]  K. Scherer,et al.  Are facial expressions of emotion produced by categorical affect programs or dynamically driven by appraisal? , 2007, Emotion.

[56]  Maurizio Mancini,et al.  Dynamic Behavior Qualifiers for Conversational Agents , 2007, IVA.

[57]  A. Manstead,et al.  Temporal Aspects of Facial Displays in Person and Expression Perception: The Effects of Smile Dynamics, Head-tilt, and Gender , 2007 .

[58]  Stefan Kopp,et al.  The Behavior Markup Language: Recent Developments and Challenges , 2007, IVA.

[59]  Stefan Kopp,et al.  The Next Step towards a Function Markup Language , 2008, IVA.

[60]  Maja Pantic,et al.  Social signal processing: Survey of an emerging domain , 2009, Image Vis. Comput..

[61]  Daniel Gatica-Perez,et al.  Automatic nonverbal analysis of social interaction in small groups: A review , 2009, Image Vis. Comput..

[62]  Russell Beale,et al.  Affective interaction: How emotional agents affect users , 2009, Int. J. Hum. Comput. Stud..

[63]  Ian H. Witten,et al.  The WEKA data mining software: an update , 2009, SKDD.

[64]  Radoslaw Niewiadomski,et al.  Greta: an interactive expressive ECA system , 2009, AAMAS.

[65]  Marilyn A. Walker,et al.  Bossy or Wimpy: Expressing Social Dominance by Combining Gaze and Linguistic Behaviors , 2010, IVA.

[66]  Sebastian Möller,et al.  Quality of talking heads in different interaction and media contexts , 2010, Speech Commun..

[67]  Nicole C. Krämer,et al.  "It doesn't matter what you are!" Explaining social effects of agents and avatars , 2010, Comput. Hum. Behav..

[68]  Dirk Heylen,et al.  How Turn-Taking Strategies Influence Users' Impressions of an Agent , 2010, IVA.

[69]  L. Devillers,et al.  Issues in Data Collection , 2011 .

[70]  Radoslaw Niewiadomski,et al.  Constraint-Based Model for Synthesis of Multimodal Sequential Expressions of Emotions , 2011, IEEE Transactions on Affective Computing.

[71]  Geneviève Calbris,et al.  Elements of Meaning in Gesture , 2011 .

[72]  Susanne Kaiser,et al.  Sequential Patterning of Facial Actions in the Production and Perception of Emotional Expressions , 2011 .

[73]  Stacy Marsella,et al.  Modeling Side Participants and Bystanders: The Importance of Being a Laugh Track , 2011, IVA.

[74]  Georgios N. Yannakakis,et al.  Mining multimodal sequential patterns: a case study on affect detection , 2011, ICMI '11.

[75]  Chen Yu,et al.  Sequential pattern mining of multimodal data streams in dyadic interactions , 2011, 2011 IEEE International Conference on Development and Learning (ICDL).

[76]  Dawn Knight,et al.  The future of multimodal corpora , 2011 .

[77]  Alan Cienki,et al.  Some Uses of Head Tilts and Shoulder Shrugs during Human Interaction, and Their Relation to Stancetaking , 2012, 2012 International Conference on Privacy, Security, Risk and Trust and 2012 International Confernece on Social Computing.

[78]  Mathieu Guillame-Bert,et al.  Learning Temporal Association Rules on Symbolic Time Sequences , 2012, ACML.

[79]  Etienne de Sevin,et al.  A listener model: introducing personality traits , 2012, Journal on Multimodal User Interfaces.

[80]  I. Szekrenyes,et al.  Annotation of spoken syntax in relation to prosody and multimodal pragmatics , 2012, 2012 IEEE 3rd International Conference on Cognitive Infocommunications (CogInfoCom).

[81]  Elisabeth Ahlsén,et al.  Some Suggestions for the Study of Stance in Communication , 2012, 2012 International Conference on Privacy, Security, Risk and Trust and 2012 International Confernece on Social Computing.

[82]  Roddy Cowie,et al.  Tracing Emotion: An Overview , 2012, Int. J. Synth. Emot..

[83]  Nick Campbell,et al.  How Do We React to Context? Annotation of Individual and Group Engagement in a Video Corpus , 2012, 2012 International Conference on Privacy, Security, Risk and Trust and 2012 International Confernece on Social Computing.

[84]  Tobias Baur,et al.  The social signal interpretation (SSI) framework: multimodal signal processing and recognition in real-time , 2013, ACM Multimedia.

[85]  Catherine Pelachaud,et al.  A Multimodal Corpus Approach to the Design of Virtual Recruiters , 2013, 2013 Humaine Association Conference on Affective Computing and Intelligent Interaction.

[86]  Catherine Pelachaud,et al.  From a User-created Corpus of Virtual Agent's Non-verbal Behavior to a Computational Model of Interpersonal Attitudes , 2013, IVA.

[87]  Mark G. Core,et al.  Learning intercultural communication skills with virtual humans: Feedback and fidelity. , 2013 .

[88]  Dirk Heylen,et al.  Taking things at face value: how stance informs politeness of virtual agents , 2013 .

[89]  Catherine Pelachaud,et al.  A computational model of social attitudes for a virtual recruiter , 2014, AAMAS.

[90]  Catherine Pelachaud,et al.  Laughter animation synthesis , 2014, AAMAS.

[91]  Jean-Claude Martin,et al.  Modeling facial signs of appraisal during interaction: impact on users' perception and behavior , 2014, AAMAS.

[92]  Oliver G. B. Garrod,et al.  Dynamic Facial Expressions of Emotion Transmit an Evolving Hierarchy of Signals over Time , 2014, Current Biology.

[93]  Catherine Pelachaud,et al.  From Non-verbal Signals Sequence Mining to Bayesian Networks for Interpersonal Attitudes Expression , 2014, IVA.

[94]  Patrick Gebhard,et al.  Exploring interaction strategies for virtual characters to induce stress in simulated job interviews , 2014, AAMAS.

[95]  Yuyu Xu,et al.  Compound Gesture Generation: A Model Based on Ideational Units , 2014, IVA.

[96]  Patrick Gebhard,et al.  Games are Better than Books: In-Situ Comparison of an Interactive Job Interview Game with Conventional Training , 2015, AIED.

[97]  Catherine Pelachaud,et al.  Towards a Socially Adaptive Virtual Agent , 2015, IVA.

[98]  Torsten Wörtwein,et al.  Exploring feedback strategies to improve public speaking: an interactive virtual audience framework , 2015, UbiComp.

[99]  Timothy W. Bickmore,et al.  First Impressions in Human--Agent Virtual Encounters , 2016, ACM Trans. Comput. Hum. Interact..

[100]  S. Adolphs,et al.  Multimodal Corpora , 2020, A Practical Handbook of Corpus Linguistics.