A Fractal Approach to Characterize Emotions in Audio and Visual Domain: A Study on Cross-Modal Interaction

It is already known that both auditory and visual stimulus is able to convey emotions in human mind to different extent. The strength or intensity of the emotional arousal vary depending on the type of stimulus chosen. In this study, we try to investigate the emotional arousal in a cross-modal scenario involving both auditory and visual stimulus while studying their source characteristics. A robust fractal analytic technique called Detrended Fluctuation Analysis (DFA) and its 2D analogue has been used to characterize three (3) standardized audio and video signals quantifying their scaling exponent corresponding to positive and negative valence. It was found that there is significant difference in scaling exponents corresponding to the two different modalities. Detrended Cross Correlation Analysis (DCCA) has also been applied to decipher degree of cross-correlation among the individual audio and visual stimulus. This is the first of its kind study which proposes a novel algorithm with which emotional arousal can be classified in cross-modal scenario using only the source audio and visual signals while also attempting a correlation between them.

[1]  K. Scherer,et al.  EMOTIONAL EFFECTS OF MUSIC: PRODUCTION RULES , 2001 .

[2]  Kevin J. Grimm,et al.  Music-evoked nostalgia: affect, memory, and personality. , 2010, Emotion.

[3]  K. Scherer,et al.  Emotions evoked by the sound of music: characterization, classification, and measurement. , 2008, Emotion.

[4]  Sayan Nag,et al.  Music of brain and music on brain: a novel EEG sonification approach , 2017, Cognitive Neurodynamics.

[5]  Sayan Nag,et al.  Can musical emotion be quantified with neural jitter or shimmer? A novel EEG based study with Hindustani classical music , 2017, 2017 4th International Conference on Signal Processing and Integrated Networks (SPIN).

[6]  D. Ghosh,et al.  A novel study on perception–cognition scenario in music using deterministic and non-deterministic approach , 2021 .

[7]  T. Eerola,et al.  Can sad music really make you sad? Indirect measures of affective states induced by music and autobiographical memories. , 2012 .

[8]  Sayan Nag,et al.  A Simultaneous EEG and EMG Study to Quantify Emotions from Hindustani Classical Music , 2021 .

[9]  Sayan Nag,et al.  Tagore and neuroscience: A non-linear multifractal study to encapsulate the evolution of Tagore songs over a century , 2021, Entertain. Comput..

[10]  Sayan Nag,et al.  Emotion specification from musical stimuli: An EEG study with AFA and DFA , 2017, 2017 4th International Conference on Signal Processing and Integrated Networks (SPIN).

[11]  Ranjan Sengupta,et al.  A Non Linear Approach towards Automated Emotion Analysis in Hindustani Music , 2016, ArXiv.

[12]  Petri Toiviainen,et al.  Prediction of Multidimensional Emotional Ratings in Music from Audio Using Multivariate Regression Models , 2009, ISMIR.

[13]  Min Peng,et al.  Non-Linear Analysis: Music and Human Emotions , 2016 .

[14]  C. Peng,et al.  Mosaic organization of DNA nucleotides. , 1994, Physical review. E, Statistical physics, plasmas, fluids, and related interdisciplinary topics.

[15]  H. Stanley,et al.  Detrended cross-correlation analysis: a new method for analyzing two nonstationary time series. , 2007, Physical review letters.

[16]  Sayan Nag,et al.  Gestalt Phenomenon in Music? A Neurocognitive Physics Study with EEG , 2017, ArXiv.

[17]  Sayan Nag,et al.  Speaker Recognition in Bengali Language from Nonlinear Features , 2020, ArXiv.

[18]  D. Ghosh,et al.  Neural (EEG) Response during Creation and Appreciation: A Novel Study with Hindustani Raga Music , 2017, 1704.05687.

[19]  D. Keltner,et al.  Approaching awe, a moral, spiritual, and aesthetic emotion , 2003, Cognition & emotion.

[20]  Sayan Nag,et al.  From Speech to Recital - A case of phase transition? A non-linear study , 2020, ArXiv.