Deep Learning for Classification of Peak Emotions within Virtual Reality Systems

Research has demonstrated well-being benefits from positive, ‘peak’ emotions such as awe and wonder, prompting the HCI community to utilize affective computing and AI modelling for elicitation and measurement of those target emotional states. The immersive nature of virtual reality (VR) content and systems can lead to feelings of awe and wonder, especially with a responsive, personalized environment based on biosignals. However, an accurate model is required to differentiate between emotional states that have similar biosignal input, such as awe and fear. Deep learning may provide a solution since the subtleties of these emotional states and affect may be recognized, with biosignal data viewed in a time series so that researchers and designers can understand which features of the system may have influenced target emotions. The proposed deep learning fusion system in this paper will use data collected from a corpus, created through collection of physiological biosignals and ranked qualitative data, and will classify these multimodal signals into target outputs of affect. This model will be real-time for the evaluation of VR system features which influence awe/wonder, using a bio-responsive environment. Since biosignal data will be collected through wireless, wearable sensor technology, and modelled through the same computer powering the VR system, it can be used in field research and studios.

[1]  Bernhard E. Riecke,et al.  Awestruck: Natural interaction with virtual reality on eliciting awe , 2017, 2017 IEEE Symposium on 3D User Interfaces (3DUI).

[2]  Romain Tavenard,et al.  Data Augmentation for Time Series Classification using Convolutional Neural Networks , 2016 .

[3]  Sung Hee Kim,et al.  Exploring the social aspects of goose bumps and their role in awe and envy , 2012 .

[4]  Stefan Koelsch,et al.  Effects of Aesthetic Chills on a Cardiac Signature of Emotionality , 2015, PloS one.

[5]  Georgios N. Yannakakis,et al.  Deep Multimodal Fusion: Combining Discrete Events and Continuous Signals , 2014, ICMI.

[6]  Georgios N. Yannakakis,et al.  Correlation between heart rate, electrodermal activity and player experience in first-person shooter games , 2010, Sandbox '10.

[7]  Wioleta Szwoch,et al.  Emotion Recognition Using Physiological Signals , 2015, MIDI '15.

[8]  D. Keltner,et al.  Approaching awe, a moral, spiritual, and aesthetic emotion , 2003, Cognition & emotion.

[9]  Lan Li,et al.  Emotion Recognition Using Physiological Signals from Multiple Subjects , 2006, 2006 International Conference on Intelligent Information Hiding and Multimedia.

[10]  Mathias Benedek,et al.  Objective and continuous measurement of piloerection. , 2010, Psychophysiology.

[11]  R. Baños,et al.  Transforming Experience: The Potential of Augmented Reality and Virtual Reality for Enhancing Personal and Clinical Change , 2016, Front. Psychiatry.

[12]  Michael Giering,et al.  Wearable EEG-based Activity Recognition in PHM-related Service Environment via Deep Learning , 2020, International Journal of Prognostics and Health Management.

[13]  Andrea Gaggioli,et al.  The Potential of Virtual Reality for the Investigation of Awe , 2016, Front. Psychol..

[14]  Stephanie E. Moser,et al.  Feeling good: autonomic nervous system responding in five positive emotions. , 2011, Emotion.

[15]  Dacher Keltner,et al.  Positive affect and markers of inflammation: discrete positive emotions predict lower levels of inflammatory cytokines. , 2015, Emotion.

[16]  E. Altenmüller,et al.  Chills in different sensory domains: Frisson elicited by acoustical, visual, tactile and gustatory stimuli , 2011 .

[17]  Lorraine Johnston,et al.  Evaluation using cued-recall debrief to elicit information about a user's affective experiences , 2005, OZCHI.

[18]  Jennifer Aaker,et al.  Awe Expands People’s Perception of Time, Alters Decision Making, and Enhances Well-Being , 2012, Psychological science.

[19]  Amani El-Alayli,et al.  Getting aesthetic chills from music: The connection between openness to experience and frisson , 2016 .

[20]  Yoshua Bengio,et al.  Learning deep physiological models of affect , 2013, IEEE Computational Intelligence Magazine.

[21]  Roger E. Beaty,et al.  Openness to experience and awe in response to nature and music: Personality and profound aesthetic experiences. , 2015 .

[22]  Reinhard Kopiez,et al.  The Chill Parameter: Goose Bumps and Shivers as Promising Measures in Emotion Research , 2009 .

[23]  Thomas S. Huang,et al.  Multi-metric learning for multi-sensor fusion based classification , 2013, Inf. Fusion.

[24]  Lauren Reinerman-Jones,et al.  A Neurophenomenology of Awe and Wonder , 2015 .

[25]  Pia Dietze,et al.  Awe, the small self, and prosocial behavior. , 2015, Journal of personality and social psychology.

[26]  Artur Lugmayr,et al.  Emotive media: a review of emotional interfaces and media in human-computer-interaction , 2016, OZCHI.

[27]  Carlos Agón,et al.  Time-series data mining , 2012, CSUR.

[28]  Mathias Benedek,et al.  Physiological correlates and emotional specificity of human piloerection , 2011, Biological Psychology.

[29]  Xiaojuan Ma,et al.  Sensing and Handling Engagement Dynamics in Human-Robot Interaction Involving Peripheral Computing Devices , 2017, CHI.

[30]  Tsutomu Miyasato,et al.  Multimodal human emotion/expression recognition , 1998, Proceedings Third IEEE International Conference on Automatic Face and Gesture Recognition.

[31]  Dacher Keltner,et al.  The nature of awe: Elicitors, appraisals, and effects on self-concept , 2007 .

[32]  Jun Li,et al.  A convolutional neural network based method for event classification in event-driven multi-sensor network , 2017, Comput. Electr. Eng..