Video and Text-Based Affect Analysis of Children in Play Therapy

Play therapy is an approach to psychotherapy where a child is engaging in play activities. Because of the strong affective component of play, it provides a natural setting to analyze feelings and coping strategies of the child. In this paper, we investigate an approach to track the affective state of a child during a play therapy session. We assume a simple, camera-based sensor setup, and describe the challenges of this application scenario. We use fine-tuned off-the-shelf deep convolutional neural networks for the processing of the child’s face during sessions to automatically extract valence and arousal dimensions of affect, as well as basic emotional expressions. We further investigate text-based and body-movement based affect analysis. We evaluate these modalities separately and in conjunction with play therapy videos in natural sessions, discussing the results of such analysis and how it aligns with the professional clinicians’ assessments.

[1]  Cha Zhang,et al.  Image based Static Facial Expression Recognition with Multiple Deep Network Learning , 2015, ICMI.

[2]  S. Droit-Volet,et al.  Synchrony and emotion in children and adults , 2018, International journal of psychology : Journal international de psychologie.

[3]  Heysem Kaya,et al.  Efficient and effective strategies for cross-corpus acoustic emotion recognition , 2018, Neurocomputing.

[4]  Mohammad H. Mahoor,et al.  AffectNet: A Database for Facial Expression, Valence, and Arousal Computing in the Wild , 2017, IEEE Transactions on Affective Computing.

[5]  Dee C. Ray,et al.  Intensive Short-Term Child-Centered Play Therapy and Externalizing Behaviors in Children , 2017 .

[6]  Yaser Sheikh,et al.  OpenPose: Realtime Multi-Person 2D Pose Estimation Using Part Affinity Fields , 2018, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[7]  Albert Ali Salah,et al.  EmoChildRu: Emotional Child Russian Speech Corpus , 2015, SPECOM.

[8]  Björn W. Schuller,et al.  The INTERSPEECH 2009 emotion challenge , 2009, INTERSPEECH.

[9]  B. Mesquita,et al.  Context in Emotion Perception , 2011 .

[10]  Amy Beth Warriner,et al.  Norms of valence, arousal, and dominance for 13,915 English lemmas , 2013, Behavior Research Methods.

[11]  Hiroshi G. Okuno,et al.  The MEI Robot: Towards Using Motherese to Develop Multimodal Emotional Intelligence , 2014, IEEE Transactions on Autonomous Mental Development.

[12]  Mika Mäntylä,et al.  Mining Valence, Arousal, and Dominance - Possibilities for Detecting Burnout and Productivity? , 2016, 2016 IEEE/ACM 13th Working Conference on Mining Software Repositories (MSR).

[13]  Emmanuel Dellandréa,et al.  Affective Video Content Analysis: A Multidisciplinary Insight , 2018, IEEE Transactions on Affective Computing.

[14]  P F Kernberg,et al.  The Children's Play Therapy Instrument (CPTI). Description, development, and reliability studies. , 1998, The Journal of psychotherapy practice and research.

[15]  Tanaya Guha,et al.  A Computational Study of Expressive Facial Dynamics in Children with Autism , 2018, IEEE Transactions on Affective Computing.

[16]  Kemal Oflazer,et al.  SentiTurkNet: a Turkish polarity lexicon for sentiment analysis , 2016, Lang. Resour. Evaluation.

[17]  Hong Zhang,et al.  Facial expression recognition via learning deep sparse autoencoders , 2018, Neurocomputing.

[18]  Yaser Sheikh,et al.  OpenPose: Realtime Multi-Person 2D Pose Estimation Using Part Affinity Fields , 2018, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[19]  Matthew S. Goodwin,et al.  Automatic Recognition of Posed Facial Expression of Emotion in Individuals with Autism Spectrum Disorder , 2018, Journal of Autism and Developmental Disorders.

[20]  Xiaoou Tang,et al.  Facial Landmark Detection by Deep Multi-task Learning , 2014, ECCV.

[21]  R. Adolphs,et al.  Emotion Perception from Face, Voice, and Touch: Comparisons and Convergence , 2017, Trends in Cognitive Sciences.

[22]  Kemal Oflazer,et al.  Turkish and Its Challenges for Language and Speech Processing , 2018 .

[23]  Rheta LeAnne Steen Emerging Research in Play Therapy, Child Counseling, and Consultation , 2017 .

[24]  Saïda Bouakaz,et al.  A novel database of Children's Spontaneous Facial Expressions (LIRIS-CSE) , 2018, Image Vis. Comput..

[25]  E. Nöth,et al.  Releasing a thoroughly annotated and processed spontaneous emotional database : the FAU Aibo Emotion Corpus , 2008 .

[26]  Albert Ali Salah,et al.  Automatic Assessment of Dimensional Affective Content in Turkish Multi-party Chat Messages , 2015, ERM4CT@ICMI.

[27]  Albert Ali Salah,et al.  Protocol and baseline for experiments on Bogazici University Turkish emotional speech corpus , 2014, 2014 22nd Signal Processing and Communications Applications Conference (SIU).

[28]  Albert Ali Salah,et al.  Assessing Affective Dimensions of Play in Psychodynamic Child Psychotherapy via Text Analysis , 2016, HBU.

[29]  Saif M. Mohammad,et al.  Sentiment Analysis: Detecting Valence, Emotions, and Other Affectual States from Text , 2016, ArXiv.

[30]  Björn W. Schuller,et al.  Personalized machine learning for robot perception of affect and engagement in autism therapy , 2018, Science Robotics.

[31]  James M. Rehg,et al.  Detecting Gaze Towards Eyes in Natural Social Interactions and Its Use in Child Assessment , 2017, Proc. ACM Interact. Mob. Wearable Ubiquitous Technol..

[32]  Nicholas Scurich,et al.  Detecting Deception in Children: A Meta-Analysis , 2017, Law and human behavior.

[33]  Martin Wegrzyn,et al.  Mapping the emotional face. How individual face parts contribute to successful emotion recognition , 2017, PloS one.

[34]  Erik Cambria,et al.  A review of affective computing: From unimodal analysis to multimodal fusion , 2017, Inf. Fusion.

[35]  Timothy R. Rice,et al.  Manual of Regulation-Focused Psychotherapy for Children (RFP-C) with Externalizing Behaviors: A Psychodynamic Approach , 2015 .

[36]  M. Chethik Techniques of Child Therapy: Psychodynamic Strategies , 1989 .