Cognitive and Affective Assessment of Navigation and Mobility Tasks for the Visually Impaired via Electroencephalography and Behavioral Signals

This paper presented the assessment of cognitive load (as an effective real-time index of task difficulty) and the level of brain activation during an experiment in which eight visually impaired subjects performed two types of tasks while using the white cane and the Sound of Vision assistive device with three types of sensory input—audio, haptic, and multimodal (audio and haptic simultaneously). The first task was to identify object properties and the second to navigate and avoid obstacles in both the virtual environment and real-world settings. The results showed that the haptic stimuli were less intuitive than the audio ones and that the navigation with the Sound of Vision device increased cognitive load and working memory. Visual cortex asymmetry was lower in the case of multimodal stimulation than in the case of separate stimulation (audio or haptic). There was no correlation between visual cortical activity and the number of collisions during navigation, regardless of the type of navigation or sensory input. The visual cortex was activated when using the device, but only for the late-blind users. For all the subjects, the navigation with the Sound of Vision device induced a low negative valence, in contrast with the white cane navigation.

[1]  Raymond M. Fish An Audio Display for the Blind , 1976, IEEE Transactions on Biomedical Engineering.

[2]  Md. Milon Islam,et al.  Developing Walking Assistants for Visually Impaired People: A Review , 2019, IEEE Sensors Journal.

[3]  R. Pekrun The impact of emotions on learning and achievement : towards a theory of cognitive/motivational mediators , 1992 .

[4]  Florica Moldoveanu,et al.  From neuromotor command to feedback: A survey of techniques for rehabilitation through altered perception , 2015, 2015 E-Health and Bioengineering Conference (EHB).

[5]  J. A. Leonard,et al.  The use of heart rate as an index of stress in blind pedestrians. , 1971, Ergonomics.

[6]  René J. Huster,et al.  Auditory Event-Related Response in Visual Cortex Modulates Subsequent Visual Responses in Humans , 2011, The Journal of Neuroscience.

[7]  K. Petrini,et al.  Efficiency of Sensory Substitution Devices Alone and in Combination With Self-Motion for Spatial Navigation in Sighted and Visually Impaired , 2020, Frontiers in Psychology.

[8]  J. Hattie,et al.  Well-Being as a Cognitive Load Reducing Agent: A Review of the Literature , 2019, Front. Educ..

[9]  Kyriaki Kalimeri,et al.  Exploring multimodal biosignal features for stress detection during indoor mobility , 2016, ICMI.

[10]  Mohammad Soleymani,et al.  Multimodal emotion recognition in response to videos (Extended abstract) , 2015, 2015 International Conference on Affective Computing and Intelligent Interaction (ACII).

[11]  Antonio Barrientos,et al.  Tactile-Sight: A Sensory Substitution Device Based on Distance-Related Vibrotactile Flow , 2013 .

[12]  Florica Moldoveanu,et al.  The TRAVEE System for a Multimodal Neuromotor Rehabilitation , 2019, IEEE Access.

[13]  Maria Bianca Amadeo,et al.  Stronger responses in the visual cortex of sighted compared to blind individuals during auditory space representation , 2019, Scientific Reports.

[14]  Kyriaki Kalimeri,et al.  Multimodal Classification of Stressful Environments in Visually Impaired Mobility Using EEG and Peripheral Biosignals , 2018, IEEE Transactions on Affective Computing.

[15]  Woodrow Barfield,et al.  Virtual environments and advanced interface design , 1995 .

[16]  Richard E. Mayer,et al.  The effects of graphic organizers giving cues to the structure of a hypertext document on users' navigation strategies and performance , 2002, Int. J. Hum. Comput. Stud..

[17]  Sunil Kumar Jena,et al.  Examination stress and its effect on EEG , 2015 .

[18]  R. Wycherley,et al.  The heart rate of blind and sighted pedestrians on a town route. , 1970, Ergonomics.

[19]  S. A. Hosseini,et al.  Classification of Emotional Stress Using Brain Activity , 2011 .

[20]  Hiroshi Ishiguro,et al.  EEG theta and Mu oscillations during perception of human and robot actions , 2013, Front. Neurorobot..

[21]  Shahzadi Tayyaba,et al.  Fuzzy-Based Approach Using IoT Devices for Smart Home to Assist Blind People for Navigation , 2020, Sensors.

[22]  S. Lederman,et al.  TACTUAL PERCEPTION , 2003 .

[23]  Michelle N. Lumicao,et al.  EEG correlates of task engagement and mental workload in vigilance, learning, and memory tasks. , 2007, Aviation, space, and environmental medicine.

[24]  Liang-Bi Chen,et al.  Design and Implementation of an Intelligent Assistive System for Visually Impaired People for Aerial Obstacle Avoidance and Fall Detection , 2020, IEEE Sensors Journal.

[25]  D. O. Bos,et al.  EEG-based Emotion Recognition The Influence of Visual and Auditory Stimuli , 2007 .

[26]  Kyriaki Kalimeri,et al.  Cognitive Load Assessment from EEG and Peripheral Biosignals for the Design of Visually Impaired Mobility Aids , 2018, Wirel. Commun. Mob. Comput..

[27]  Kyriaki Kalimeri,et al.  Identifying Urban Mobility Challenges for the Visually Impaired with Mobile Monitoring of Multimodal Biosignals , 2016, HCI.

[28]  Hazel Roddam,et al.  The Voice of Technology , 1994 .

[29]  Florica Moldoveanu,et al.  VIRTUAL MINI-GAMES - A SERIOUS LEARNING TOOL FOR SENSORY SUBSTITUTION DEVICES , 2017 .

[30]  David Dewhurst,et al.  CREATING AND ACCESSING AUDIOTACTILE IMAGES WITH "HFVE" VISION SUBSTITUTION SOFTWARE , 2010 .

[31]  Soraia M. Alarcão,et al.  Emotions Recognition Using EEG Signals: A Survey , 2019, IEEE Transactions on Affective Computing.

[32]  M. Hallett,et al.  Activation of the primary visual cortex by Braille reading in blind subjects , 1996, Nature.

[33]  Gerhard Tröster,et al.  Discriminating Stress From Cognitive Load Using a Wearable EDA Device , 2010, IEEE Transactions on Information Technology in Biomedicine.

[34]  Pavlo D. Antonenko,et al.  Using Electroencephalography to Measure Cognitive Load , 2010 .

[35]  J. Russell Core affect and the psychological construction of emotion. , 2003, Psychological review.

[36]  P. Pietrini,et al.  Imagery and spatial processes in blindness and visual impairment , 2008, Neuroscience & Biobehavioral Reviews.

[37]  Pablo Revuelta Sanz,et al.  A sonification proposal for safe travels of blind people , 2012 .

[38]  S. Stergiopoulos,et al.  Sonification of range information for 3-D space perception , 2003, IEEE Transactions on Neural Systems and Rehabilitation Engineering.

[39]  William M. Stern,et al.  Shape conveyed by visual-to-auditory sensory substitution activates the lateral occipital complex , 2007, Nature Neuroscience.

[40]  Jennifer Healey,et al.  Detecting stress during real-world driving tasks using physiological sensors , 2005, IEEE Transactions on Intelligent Transportation Systems.

[41]  E. Peper,et al.  Is There More to Blood Volume Pulse Than Heart Rate Variability , Respiratory Sinus Arrhythmia , and Cardiorespiratory Synchrony ? , 2007 .

[42]  H. Burton Visual Cortex Activity in Early and Late Blind People , 2003, The Journal of Neuroscience.

[43]  Pawel Strumillo,et al.  Development of a versatile assistive system for the visually impaired based on sensor fusion , 2017, 2017 21st International Conference on System Theory, Control and Computing (ICSTCC).

[44]  Stephen A. Brewster,et al.  Sensory substitution using tactile pin arrays: Human factors, technology and applications , 2006, Signal Process..

[45]  Guido Bologna,et al.  Image and Video Processing for Visually Handicapped People , 2007, EURASIP J. Image Video Process..

[46]  Florina Ungureanu,et al.  Usability assessment of assistive technology for blind and visually impaired , 2017, 2017 E-Health and Bioengineering Conference (EHB).

[47]  Mendel Kleiner,et al.  Spatial sound in auditory vision substitution systems , 2006 .

[48]  György Wersényi,et al.  Current Use and Future Perspectives of Spatial Audio Technologies in Electronic Travel Aids , 2018, Wirel. Commun. Mob. Comput..

[49]  Gabriela Moise,et al.  Emotion Classification Based on Biophysical Signals and Machine Learning Techniques , 2019, Symmetry.

[50]  Ómar I. Jóhannesson,et al.  Designing sensory-substitution devices: Principles, pitfalls and potential , 2016, Restorative neurology and neuroscience.

[51]  Brianna Scott,et al.  Navigational spatial displays: The role of metacognition as cognitive load * , 2007 .

[52]  E. E. FOURNIER D'ALBE,et al.  The Optophone: An Instrument for Reading by Ear. , 1920, Nature.

[53]  Sankari Subbiah,et al.  Navigation gadget for visually impaired based on IoT , 2017, 2017 2nd International Conference on Computing and Communications Technologies (ICCCT).

[54]  Adrian Burlacu,et al.  Computer Vision for the Visually Impaired: the Sound of Vision System , 2017, 2017 IEEE International Conference on Computer Vision Workshops (ICCVW).

[55]  永福 智志 The Organization of Learning , 2005, Journal of Cognitive Neuroscience.

[56]  Florina Ungureanu,et al.  Mastering an advanced sensory substitution device for visually impaired through innovative virtual training , 2017, 2017 IEEE 7th International Conference on Consumer Electronics - Berlin (ICCE-Berlin).

[57]  Paul Bach-y-Rita,et al.  Tactile displays , 1995 .

[58]  Yu Cao,et al.  ReliefF-Based EEG Sensor Selection Methods for Emotion Recognition , 2016, Sensors.