Comparative Analysis of Three Different Modalities for Perception of Artifacts in Videos

This study compares three popular modalities for analyzing perceived video quality; user ratings, eye tracking, and EEG. We contrast these three modalities for a given video sequence to determine if there is a gap between what humans consciously see and what we implicitly perceive. Participants are shown a video sequence with different artifacts appearing at specific distances in their field of vision; near foveal, middle peripheral, and far peripheral. Our results show distinct differences between what we saccade to (eye tracking), how we consciously rate video quality, and our neural responses (EEG data). Our findings indicate that the measurement of perceived quality depends on the specific modality used.

[1]  David Robinson,et al.  The oculomotor control system: A review , 1968 .

[2]  Cristina Hava Muntean,et al.  User-centered EEG-based multimedia quality assessment , 2013, 2013 IEEE International Symposium on Broadband Multimedia Systems and Broadcasting (BMSB).

[3]  Marcus A. Magnor,et al.  Assessing the quality of compressed images using EEG , 2011, 2011 18th IEEE International Conference on Image Processing.

[4]  Deep Medhi,et al.  Measurement of Quality of Experience of Video-on-Demand Services: A Survey , 2016, IEEE Communications Surveys & Tutorials.

[5]  Benjamin Blankertz,et al.  Classification of Eye Fixation Related Potentials for Variable Stimulus Saliency , 2016, Front. Neurosci..

[6]  S. Somani The P 300 wave of Event-Related-Potential , 2014 .

[7]  Sebastian Möller,et al.  A Physiological Approach to Determine Video Quality , 2011, 2011 IEEE International Symposium on Multimedia.

[8]  Byron Nakos,et al.  EyeMMV toolbox: An eye movement post-analysis tool based on a two-step spatial dispersion threshold for fixation identification , 2014 .

[9]  Laurent Itti,et al.  Visual attention guided bit allocation in video compression , 2011, Image Vis. Comput..

[10]  Ingrid Heynderickx,et al.  Visual Attention in Objective Image Quality Assessment: Based on Eye-Tracking Data , 2011, IEEE Transactions on Circuits and Systems for Video Technology.

[11]  Sebastian Möller,et al.  Using Electroencephalography to Measure Perceived Video Quality , 2014, IEEE Journal of Selected Topics in Signal Processing.

[12]  T W Picton,et al.  The P300 Wave of the Human Event‐Related Potential , 1992, Journal of clinical neurophysiology : official publication of the American Electroencephalographic Society.

[13]  Marcus A. Magnor,et al.  Evaluation of video artifact perception using event-related potentials , 2011, APGV '11.

[14]  Sebastian Bosse,et al.  EEG-based classification of video quality perception using steady state visual evoked potentials (SSVEPs) , 2015, Journal of neural engineering.

[15]  Sebastian Bosse,et al.  Psychophysiology-Based QoE Assessment: A Survey , 2017, IEEE Journal of Selected Topics in Signal Processing.

[16]  Yi Zhu,et al.  Understanding the role of social context and user factors in video Quality of Experience , 2015, Comput. Hum. Behav..

[17]  Sebastian Bosse,et al.  Neurally informed assessment of perceived natural texture image quality , 2014, 2014 IEEE International Conference on Image Processing (ICIP).

[18]  Seungjin Choi,et al.  SSVEP response on Oculus Rift , 2015, The 3rd International Winter Conference on Brain-Computer Interface.

[19]  J. Fermaglich Electric Fields of the Brain: The Neurophysics of EEG , 1982 .

[20]  Wei Zhang,et al.  Study of Saliency in Objective Video Quality Assessment , 2017, IEEE Transactions on Image Processing.

[21]  Stephen R. Gulliver,et al.  Stars in their eyes: what eye-tracking reveals about multimedia perceptual quality , 2004, IEEE Transactions on Systems, Man, and Cybernetics - Part A: Systems and Humans.

[22]  Touradj Ebrahimi,et al.  EEG correlates during video quality perception , 2014, 2014 22nd European Signal Processing Conference (EUSIPCO).

[23]  J. Enderle,et al.  Relationships between manual reaction time and saccade latency in response to visual and auditory stimuli. , 1991, Aviation, space, and environmental medicine.

[24]  Marcus A. Magnor,et al.  Single-trial EEG classification of artifacts in videos , 2012, TAP.

[25]  G. Horridge The evolution of visual processing and the construction of seeing systems , 1987, Proceedings of the Royal Society of London. Series B. Biological Sciences.

[26]  Rajiv Soundararajan,et al.  Study of Subjective and Objective Quality Assessment of Video , 2010, IEEE Transactions on Image Processing.

[27]  Hans-Peter Seidel,et al.  Saccade landing position prediction for gaze-contingent rendering , 2017, ACM Trans. Graph..

[28]  Jinglei Lv,et al.  What Makes a Good Movie Trailer?: Interpretation from Simultaneous EEG and Eyetracker Recording , 2016, ACM Multimedia.

[29]  C. C. Duncan,et al.  Event-related potentials in clinical research: Guidelines for eliciting, recording, and quantifying mismatch negativity, P300, and N400 , 2009, Clinical Neurophysiology.

[30]  David H. Laidlaw,et al.  Thoughts on User Studies: Why, How and When , 1993 .

[31]  Nikolay N. Ponomarenko,et al.  TID2008 – A database for evaluation of full-reference visual quality assessment metrics , 2004 .

[32]  Patrick Le Callet,et al.  Overt visual attention for free-viewing and quality assessment tasks Impact of the regions of interest on a video quality metric , 2010 .

[33]  C. C. Duncan,et al.  Event-related potentials in clinical research : Guidelines for eliciting , recording , and quantifying mismatch negativity , P 300 , and N 400 , 2017 .

[34]  Sebastian Möller,et al.  Perception of low-quality videos analyzed by means of electroencephalography , 2012, 2012 Fourth International Workshop on Quality of Multimedia Experience.

[35]  Peter Gena,et al.  A Physiological Approach to DNA Music , 2001 .

[36]  Victoria Interrante,et al.  User Studies: Why, How, and When? , 2003, IEEE Computer Graphics and Applications.

[37]  Marina Schmid,et al.  An Introduction To The Event Related Potential Technique , 2016 .

[38]  Gary R. Bradski,et al.  Peripheral-Foveal Vision for Real-time Object Recognition and Tracking in Video , 2007, IJCAI.

[39]  J. Dowling,et al.  Response characteristics of different Butterworth low-pass digital filters , 1993 .

[40]  R. Brothers Aviation, Space, and Environmental Medicine , 2012 .