Reinforcement Learning Motion Planning for an EOG-centered Robot Assisted Navigation in a Virtual Environment

This paper presents a new collaborative approach for robot motion planning of an assistive robotic platform that takes into account the intentions of the user provided through Electrooculographic (EOG) signals, as well as obstacles surrounding the robotic platform. In order to increase human confidence in the operation of robotic platforms with some degree of navigational autonomy, the intent of the user must be included in the decision process. In our system, the humanrobot interface works through ocular movements (saccades and blinks), which are acquired as EOG signals and classified using a Convolutional Neural Network. In our proposed approach, a model-free Reinforcement Learning (RL) layer is used to provide commands to a virtual robotic platform. The RL layer is constantly being updated with the inputs from the user’s intent, environment perception and previous machine-based decisions. In order to prevent collisions, machine-based perception using the proposed RL motion planning approach will assist the user by selecting suitable actions while learning from prior driving behaviors. The approach was validated by a set of tests that consisted of driving a robotic platform in an in-house 3D virtual model of our Research Center (ISR-UC). The experimental results show a better performance of the proposed approach with RL when compared to the version without the RL-based motion planning component. Results show that the approach is a promising step in the concept put forward for collaborative Human-Robotic Interaction (HRI), and opens a path for future research.

[1]  Luis Garrote,et al.  Measuring the impact of reinforcement learning on an electrooculography-only computer game , 2018, 2018 IEEE 6th International Conference on Serious Games and Applications for Health (SeGAH).

[2]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[3]  Xiaoping Chen,et al.  EOG-based drowsiness detection using convolutional neural networks , 2014, 2014 International Joint Conference on Neural Networks (IJCNN).

[4]  Doru Talaba,et al.  EOG-based visual navigation interface development , 2012, Expert Syst. Appl..

[5]  Michael S. Bernstein,et al.  ImageNet Large Scale Visual Recognition Challenge , 2014, International Journal of Computer Vision.

[6]  Nguyen Truong-Thinh,et al.  Using Electrooculogram and Electromyogram for powered wheelchair , 2011, 2011 IEEE International Conference on Robotics and Biomimetics.

[7]  K. Latha,et al.  Feature Selection in Classification of Eye Movements Using Electrooculography for Activity Recognition , 2014, Comput. Math. Methods Medicine.

[8]  Andreas Bulling,et al.  End-to-End Eye Movement Detection Using Convolutional Neural Networks , 2016, ArXiv.

[9]  Yael Edan,et al.  A Human-Robot Collaborative Reinforcement Learning Algorithm , 2010, J. Intell. Robotic Syst..

[10]  João Paulo,et al.  Robot-Assisted Navigation for a Robotic Walker with Aided User Intent , 2018, 2018 27th IEEE International Symposium on Robot and Human Interactive Communication (RO-MAN).

[11]  Jessie Y. C. Chen,et al.  A Meta-Analysis of Factors Affecting Trust in Human-Robot Interaction , 2011, Hum. Factors.

[12]  Sabalan Daneshvar,et al.  A low-cost real-time wheelchair navigation system using electrooculography , 2014, 2014 22nd Iranian Conference on Electrical Engineering (ICEE).

[13]  Luis Garrote,et al.  Modular software architecture for human-robot interaction applied to the InterBot mobile robot , 2018, 2018 IEEE International Conference on Autonomous Robot Systems and Competitions (ICARSC).

[14]  Geoffrey E. Hinton,et al.  ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.

[15]  James M. Conrad,et al.  Human-robot collaboration: A survey , 2015, SoutheastCon 2015.

[16]  Gerhard Tröster,et al.  Eye Movement Analysis for Activity Recognition Using Electrooculography , 2011, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[17]  Yuanqing Li,et al.  An EOG-Based Human–Machine Interface for Wheelchair Control , 2018, IEEE Transactions on Biomedical Engineering.

[18]  B. E. Maki,et al.  Assistive devices for balance and mobility: benefits, demands, and adverse consequences. , 2005, Archives of physical medicine and rehabilitation.

[19]  José Luis Lázaro,et al.  Integral system for assisted mobility , 2000, Inf. Sci..

[20]  Min Peng,et al.  NIRFaceNet: A Convolutional Neural Network for Near-Infrared Face Identification , 2016, Inf..

[21]  Manuel Mazo,et al.  Wheelchair Guidance Strategies Using EOG , 2002, J. Intell. Robotic Syst..

[22]  Manuel Mazo,et al.  An integral system for assisted mobility [automated wheelchair] , 2001, IEEE Robotics Autom. Mag..