Safety Criteria Analysis for Negotiating Blind Corners in Personal Mobility Vehicles Based on Driver’s Attention Simulation on 3D Map

In this study, we attempt to establish the numerical safety criteria for negotiating blind corners in personal mobility vehicles (PMVs). Safety should be the most important consideration in designing autonomous PMVs. However, determining the suitable trade-off between safety and speed is a weighty concern because speed is significantly compromised when performing overly safe navigation. We analyze the driving behavior of a robotic PMV operated by a human driver. The robotic PMV can measure the driver’s gaze, and allows us to recognize both the pose of the PMV and the driver’s visual attention on a 3D map. As a result, the occluded areas for the driver can be estimated. Then, potential colliding hazard obstacles (PCHOs) are simulated based on the occlusion. PCHOs refer to occluded obstacles that the driver encounters suddenly with which he cannot avoid collision. The participants of our experiments were one skillful and three non-skilled ones. Experimental results demonstrate that similar PCHOs are observed even when the driving styles of the participants are different. Additionally, the existence of a boundary that distinguishes expected and unexpected obstacles is indicated by investigating the parameters of the PCHOs. Finally, we conclude that the boundary could be utilized as a numerical criterion for ensuring safety while negotiating blind corners.

[1]  T. Eames THE VISUAL FIELDS , 1955 .

[2]  J. D. Rains SIGNAL LUMINANCE AND POSITION EFFECTS IN HUMAN REACTION TIME. , 1963, Vision research.

[3]  R. Mansfield,et al.  Latency functions in human vision. , 1973, Vision research.

[4]  Mitsuo Ikeda,et al.  Influence of foveal load on the functional visual field , 1975 .

[5]  Jens Rasmussen,et al.  Skills, rules, and knowledge; signals, signs, and symbols, and other distinctions in human performance models , 1983, IEEE Transactions on Systems, Man, and Cybernetics.

[6]  John A. Michon,et al.  A critical view of driver behavior models: What do we know , 1985 .

[7]  Yiannis Aloimonos,et al.  Active vision , 2004, International Journal of Computer Vision.

[8]  T Miura VISUAL SEARCH IN INTERSECTIONS - AN UNDERLYING MECHANISM , 1992 .

[9]  Michael F. Land,et al.  Predictable eye-head coordination during driving , 1992, Nature.

[10]  M. Ishihara,et al.  Lateralized effects of target location on reaction times when preparing for manual aiming at a visual target. , 2002, Human movement science.

[11]  Qiang Ji,et al.  Real-Time Eye, Gaze, and Face Pose Tracking for Monitoring Driver Vigilance , 2002, Real Time Imaging.

[12]  Wolfram Burgard,et al.  Improved Techniques for Grid Mapping With Rao-Blackwellized Particle Filters , 2007, IEEE Transactions on Robotics.

[13]  Mohan Trivedi,et al.  Investigating the relationships between gaze patterns, dynamic vehicle surround analysis, and driver intentions , 2009, 2009 IEEE Intelligent Vehicles Symposium.

[14]  Kurt Konolige,et al.  Double window optimisation for constant time visual SLAM , 2011, 2011 International Conference on Computer Vision.

[15]  Horst Bischof,et al.  GPSlam: Marrying Sparse Geometric and Dense Probabilistic Visual Mapping , 2011, BMVC.

[16]  Martial Hebert,et al.  Activity Forecasting , 2012, ECCV.

[17]  Timo Milbich,et al.  May I enter the roundabout? A time-to-contact computation based on stereo-vision , 2012, 2012 IEEE Intelligent Vehicles Symposium.

[18]  Wolfram Burgard,et al.  OctoMap: an efficient probabilistic 3D mapping framework based on octrees , 2013, Autonomous Robots.

[19]  Johann Schrammel,et al.  3D attention: measurement of visual saliency using eye tracking glasses , 2013, CHI Extended Abstracts.

[20]  Takatsugu Hirayama,et al.  Analysis of Temporal Relationships between Eye Gaze and Peripheral Vehicle Behavior for Detecting Driver Distraction , 2013 .

[21]  Mohan M. Trivedi,et al.  Where is the driver looking: Analysis of head, eye and iris for robust gaze zone estimation , 2014, 17th International IEEE Conference on Intelligent Transportation Systems (ITSC).

[22]  Alan Yuille,et al.  Active Vision , 2014, Computer Vision, A Reference Guide.

[23]  Norihiro Hagita,et al.  Visibility analysis for autonomous vehicle comfortable navigation , 2014, 2014 IEEE International Conference on Robotics and Automation (ICRA).

[24]  Wolfram Burgard,et al.  Learning driving styles for autonomous vehicles from demonstration , 2015, 2015 IEEE International Conference on Robotics and Automation (ICRA).

[25]  Tatsuya Suzuki,et al.  Analyzing driver gaze behavior and consistency of decision making during automated driving , 2015, 2015 IEEE Intelligent Vehicles Symposium (IV).

[26]  Eijiro Takeuchi,et al.  Blind Area Traffic Prediction Using High Definition Maps and LiDAR for Safe Driving Assist , 2015, 2015 IEEE 18th International Conference on Intelligent Transportation Systems.

[27]  Morales Saiki Luis Yoichi,et al.  Comfortable Autonomous Navigation Based on Collision Prediction in Blind Occluded Regions. , 2015 .

[28]  Tatsuya Suzuki,et al.  Realization of different driving characteristics for autonomous vehicle by using model predictive control , 2016, 2016 IEEE Intelligent Vehicles Symposium (IV).

[29]  Takatsugu Hirayama,et al.  Classification of Driver's Neutral and Cognitive Distraction States Based on Peripheral Vehicle Behavior in Driver's Gaze Transition , 2016, IEEE Transactions on Intelligent Vehicles.

[30]  Eijiro Takeuchi,et al.  Proactive driving modeling in blind intersections based on expert driver data , 2017, 2017 IEEE Intelligent Vehicles Symposium (IV).

[31]  Eijiro Takeuchi,et al.  Autonomous predictive driving for blind intersections , 2017, 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[32]  Nicholas Rhinehart,et al.  First-Person Activity Forecasting with Online Inverse Reinforcement Learning , 2016, 2017 IEEE International Conference on Computer Vision (ICCV).

[33]  Charles Richter,et al.  Safe Visual Navigation via Deep Learning and Novelty Detection , 2017, Robotics: Science and Systems.

[34]  Eijiro Takeuchi,et al.  Robust localization using 3D NDT scan matching with experimentally determined uncertainty and road marker matching , 2017, 2017 IEEE Intelligent Vehicles Symposium (IV).

[35]  Eijiro Takeuchi,et al.  Autonomous driving based on accurate localization using multilayer LiDAR and dead reckoning , 2017, 2017 IEEE 20th International Conference on Intelligent Transportation Systems (ITSC).

[36]  Simon K. Rushton,et al.  Peripheral Visual Cues Contribute to the Perception of Object Movement During Self-Movement , 2017, i-Perception.

[37]  Klaus C. J. Dietmayer,et al.  Entering crossroads with blind corners. A safe strategy for autonomous vehicles , 2017, 2017 IEEE Intelligent Vehicles Symposium (IV).

[38]  Eijiro Takeuchi,et al.  Driving Feature Extraction and Behavior Classification Using an Autoencoder to Reproduce the Velocity Styles of Experts , 2018, 2018 21st International Conference on Intelligent Transportation Systems (ITSC).

[39]  Hiroshi Murase,et al.  Towards Predictive Driving through Blind Intersections , 2018, 2018 21st International Conference on Intelligent Transportation Systems (ITSC).

[40]  Shaojie Shen,et al.  SLAM-based localization of 3D gaze using a mobile eye tracker , 2018, ETRA.

[41]  Hiroshi Murase,et al.  Mobile Robot Localization Considering Class of Sensor Observations , 2018, 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[42]  Norihiro Hagita,et al.  Comfortable Intelligence for Evaluating Passenger Characteristics in Autonomous Wheelchairs , 2018, IEICE Trans. Fundam. Electron. Commun. Comput. Sci..

[43]  Ali Farhadi,et al.  YOLOv3: An Incremental Improvement , 2018, ArXiv.

[44]  Hiroshi Murase,et al.  Driving Behavior Modeling Based on Hidden Markov Models with Driver's Eye-Gaze Measurement and Ego-Vehicle Localization , 2019, 2019 IEEE Intelligent Vehicles Symposium (IV).