End-to-End Deep Learning Applied in Autonomous Navigation using Multi-Cameras System with RGB and Depth Images

The present work demonstrates how an autonomous navigation system of ‘End-to-End’ deep learning principles is directly improved in its response process, depending on the information obtained by different input images configurations. For this, a methodology was developed to allow working with RGB and depth images, which were obtained through a Microsoft Kinect V2 sensor device. Three cameras were used for this experiment. The images of the different cameras were concatenated or grouped, generating new and different input configurations from the vision system. To develop the presented methodology, two support and validation systems were implemented. Through the process of computer simulation, it was able to test the first approaches and define the most important ones. In order to validate the proposed methodology and solutions in real world situations, a 1/4 scale automotive vehicle was prototyped. Finally, the experiments shows the importance of the use of multi-cameras systems for a better performance of autonomous navigation systems based on End-to-End learning approach, heaving an average error of 2.41 degrees in the best configuration tested, with three RGB cameras.

[1]  Roland Siegwart,et al.  Using multi-camera systems in robotics: Efficient solutions to the NPnP problem , 2013, 2013 IEEE International Conference on Robotics and Automation.

[2]  Rachid Belaroussi,et al.  Accurate lateral positioning from map data and road marking detection , 2016, Expert Syst. Appl..

[3]  Dean Pomerleau,et al.  ALVINN, an autonomous land vehicle in a neural network , 2015 .

[4]  Torsten Sattler,et al.  Project AutoVision: Localization and 3D Scene Perception for an Autonomous Vehicle with a Multi-Camera System , 2018, 2019 International Conference on Robotics and Automation (ICRA).

[5]  Madjid Tavana,et al.  Autonomous vehicles: challenges, opportunities, and future implications for transportation policies , 2016, Journal of Modern Transportation.

[6]  Steven F. Barrett Arduino Microcontroller: Processing for Everyone! Second Edition , 2012, Arduino Microcontroller: Processing for Everyone! Second Edition.

[7]  Morgan Quigley,et al.  ROS: an open-source Robot Operating System , 2009, ICRA 2009.

[8]  Andrzej Skalski,et al.  Accuracy assessment of Kinect for Xbox One in point-based tracking applications , 2015, International Conference on Machine Vision.

[9]  R. Kozak,et al.  Virtual Reality Educational Pathfinders (VREP) , 2013, 2013 3rd Interdisciplinary Engineering Design Education Conference.

[10]  Xin Zhang,et al.  End to End Learning for Self-Driving Cars , 2016, ArXiv.

[11]  Guy Rosman,et al.  Variational End-to-End Navigation and Localization , 2018, 2019 International Conference on Robotics and Automation (ICRA).

[12]  Daniel J. Fagnant,et al.  Preparing a Nation for Autonomous Vehicles: Opportunities, Barriers and Policy Recommendations , 2015 .

[13]  Luc Van Gool,et al.  End-to-End Learning of Driving Models with Surround-View Cameras and Route Planners , 2018, ECCV.

[14]  Steven F. Barrett,et al.  Arduino Microcontroller Processing for Everyone! Third Edition , 2013, Arduino Microcontroller Processing for Everyone! Third Edition.

[15]  Marc Pollefeys,et al.  Motion Estimation for Self-Driving Cars with a Generalized Camera , 2013, 2013 IEEE Conference on Computer Vision and Pattern Recognition.

[16]  Xinming Huang,et al.  End-to-end learning for lane keeping of self-driving cars , 2017, 2017 IEEE Intelligent Vehicles Symposium (IV).

[17]  J. Krejsa,et al.  Ackermann mobile robot chassis with independent rear wheel drives , 2010, Proceedings of 14th International Power Electronics and Motion Control Conference EPE-PEMC 2010.

[18]  Pascal Vasseur,et al.  Visual odometry with unsynchronized multi-cameras setup for intelligent vehicle application , 2014, 2014 IEEE Intelligent Vehicles Symposium Proceedings.

[19]  Roland Siegwart,et al.  Introduction to Autonomous Mobile Robots , 2004 .