The POTUS Corpus, a Database of Weekly Addresses for the Study of Stance in Politics and Virtual Agents

One of the main challenges in the field of Embodied Conversational Agent (ECA) is to generate socially believable agents. The common strategy for agent behaviour synthesis is to rely on dedicated corpus analysis. Such a corpus is composed of multimedia files of socio-emotional behaviors which have been annotated by external observers. The underlying idea is to identify interaction information for the agent’s socio-emotional behavior by checking whether the intended socio-emotional behavior is actually perceived by humans. Then, the annotations can be used as learning classes for machine learning algorithms applied to the social signals. This paper introduces the POTUS Corpus composed of high-quality audio-video files of political addresses to the American people. Two protagonists are present in this database. First, it includes speeches of former president Barack Obama to the American people. Secondly, it provides videos of these same speeches given by a virtual agent named Rodrigue. The ECA reproduces the original address as closely as possible using social signals automatically extracted from the original one. Both are annotated for social attitudes, providing information about the stance observed in each file. It also provides the social signals automatically extracted from Obama’s addresses used to generate Rodrigue’s ones.

[1]  Ran Zhao,et al.  Socially-Aware Virtual Agents: Automatically Assessing Dyadic Rapport from Temporal Patterns of Behavior , 2016, IVA.

[2]  Nora A. Murphy,et al.  Using Thin Slices for Behavioral Coding , 2005 .

[3]  Carlos Busso,et al.  IEMOCAP: interactive emotional dyadic motion capture database , 2008, Lang. Resour. Evaluation.

[4]  J. S. Wiggins,et al.  A psychological taxonomy of trait-descriptive terms: The interpersonal domain. , 1979 .

[5]  Isabella Poggi,et al.  The Perception of Charisma from Voice: A Cross-Cultural Study , 2013, 2013 Humaine Association Conference on Affective Computing and Intelligent Interaction.

[6]  Elisabeth André,et al.  From Annotated Multimodal Corpora to Simulated Human-Like Behaviors , 2006, ZiF Workshop.

[7]  D. Kiesler From communications to interpersonal theory: a personal odyssey. , 1996, Journal of personality assessment.

[8]  Karl F. MacDorman,et al.  The Uncanny Valley [From the Field] , 2012, IEEE Robotics Autom. Mag..

[9]  D. Keltner Signs of appeasement: evidence for the distinct displays of embarrassment, amusement, and shame , 1995 .

[10]  Alessandro Vinciarelli,et al.  Canal9: A database of political debates for analysis of social interactions , 2009, 2009 3rd International Conference on Affective Computing and Intelligent Interaction and Workshops.

[11]  Fabien Ringeval,et al.  Introducing the RECOLA multimodal corpus of remote collaborative and affective interactions , 2013, 2013 10th IEEE International Conference and Workshops on Automatic Face and Gesture Recognition (FG).

[12]  Fernando De la Torre,et al.  Supervised Descent Method and Its Applications to Face Alignment , 2013, 2013 IEEE Conference on Computer Vision and Pattern Recognition.

[13]  Catherine Pelachaud,et al.  A multimodal corpus for the study of non-verbal behavior expressing interpersonal stances , 2013 .

[14]  Vladimir Pavlovic,et al.  1 Machine Learning Methods for Social Signal Processing , 2014 .

[15]  K. Scherer What are emotions? And how can they be measured? , 2005 .

[16]  K. Locke Interpersonal circumplex measures. , 2006 .

[17]  K. Scherer,et al.  Introducing the Geneva Multimodal Emotion Portrayal (GEMEP) corpus , 2010 .

[18]  Klaus Krippendorff,et al.  Computing Krippendorff's Alpha-Reliability , 2011 .

[19]  Timothy W. Bickmore,et al.  First Impressions in Human--Agent Virtual Encounters , 2016, ACM Trans. Comput. Hum. Interact..

[20]  Vladimir Pavlovic,et al.  Machine Learning Methods for Social Signal Processing , 2017, Social Signal Processing.

[21]  P. Ekman,et al.  Facial action coding system: a technique for the measurement of facial movement , 1978 .

[22]  Dirk Heylen,et al.  Bridging the Gap between Social Animal and Unsocial Machine: A Survey of Social Signal Processing , 2012, IEEE Transactions on Affective Computing.

[23]  Catherine Pelachaud,et al.  Engagement Modeling in Dyadic Interaction , 2019, ICMI.

[24]  Michael J. Constantino,et al.  How Interpersonal Motives Clarify the Meaning of Interpersonal Behavior: A Revised Circumplex Model , 2006, Personality and social psychology review : an official journal of the Society for Personality and Social Psychology, Inc.

[25]  Mohamed Chetouani,et al.  Real-time facial action unit intensity prediction with regularized metric learning , 2016, Image Vis. Comput..

[26]  Heloir,et al.  The Uncanny Valley , 2019, The Animation Studies Reader.

[27]  Maja Pantic,et al.  AFEW-VA database for valence and arousal estimation in-the-wild , 2017, Image Vis. Comput..

[28]  Maja Pantic,et al.  Social signal processing: Survey of an emerging domain , 2009, Image Vis. Comput..

[29]  Andrei Popescu-Belis,et al.  Combining content with user preferences for TED lecture recommendation , 2013, 2013 11th International Workshop on Content-Based Multimedia Indexing (CBMI).

[30]  M. Lorr,et al.  Differentiating Normal and Abnormal Personality , 1994 .

[31]  E. Brunswik Perception and the Representative Design of Psychological Experiments , 1957 .

[32]  A. Pentland Social Dynamics: Signals and Behavior , 2004 .

[33]  Chloé Clavel,et al.  Using Temporal Association Rules for the Synthesis of Embodied Conversational Agents with a Specific Stance , 2016, IVA.

[34]  J. Cassell 17. Body Language: Lessons from the Near-Human , 2019 .

[35]  N. Ambady,et al.  Thin slices of expressive behavior as predictors of interpersonal consequences: A meta-analysis. , 1992 .

[36]  Louis-Philippe Morency,et al.  Multimodal Machine Learning: A Survey and Taxonomy , 2017, IEEE Transactions on Pattern Analysis and Machine Intelligence.