A Simulation-based End-to-End Learning Framework for Evidential Occupancy Grid Mapping

Evidential occupancy grid maps (OGMs) are a popular representation of the environment of automated vehicles. Inverse sensor models (ISMs) are used to compute OGMs from sensor data such as lidar point clouds. Geometric ISMs show a limited performance when estimating states in unobserved but inferable areas and have difficulties dealing with ambiguous input. Deep learning-based ISMs face the challenge of limited training data and they often cannot handle uncertainty quantification yet. We propose a deep learning-based framework for learning an OGM algorithm which is both capable of quantifying first- and second-order uncertainty and which does not rely on manually labeled data. Results on synthetic and on real-world data show superiority over other approaches. Source code and datasets are available at https://github.com/ika-rwth-aachen/EviLOG.

[1]  Alberto Elfes,et al.  Using occupancy grids for mobile robot perception and navigation , 1989, Computer.

[2]  Lutz Eckstein,et al.  Reducing Uncertainty by Fusing Dynamic Occupancy Grid Maps in a Cloud-based Collective Environment Model , 2020, 2020 IEEE Intelligent Vehicles Symposium (IV).

[3]  P. Alam ‘S’ , 2021, Composites Engineering: An A–Z Guide.

[4]  P. Alam,et al.  R , 1823, The Herodotus Encyclopedia.

[5]  Shaul Oron,et al.  Road Scene Understanding by Occupancy Grid Learning from Sparse Radar Clusters using Semantic Segmentation , 2019, 2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW).

[6]  Murat Sensoy,et al.  Evidential Deep Learning to Quantify Classification Uncertainty , 2018, NeurIPS.

[7]  Sebastian Thrun,et al.  Probabilistic robotics , 2002, CACM.

[8]  Lutz Eckstein,et al.  Deep Inverse Sensor Models as Priors for evidential Occupancy Mapping , 2020, ArXiv.

[9]  Gebräuchliche Fertigarzneimittel,et al.  V , 1893, Therapielexikon Neurologie.

[10]  Yin Zhou,et al.  VoxelNet: End-to-End Learning for Point Cloud Based 3D Object Detection , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[11]  Klaus C. J. Dietmayer,et al.  A random finite set approach for dynamic occupancy grid maps with real-time application , 2016, Int. J. Robotics Res..

[12]  Wolfram Burgard,et al.  Probabilistic Robotics (Intelligent Robotics and Autonomous Agents) , 2005 .

[13]  Miss A.O. Penney (b) , 1974, The New Yale Book of Quotations.

[14]  Kilian von Neumann-Cosel,et al.  Virtual test drive - provision of a consistent tool-set for [D,H,S,V]-in-the-loop , 2009 .

[15]  Audun Jøsang,et al.  Subjective Logic , 2016, Artificial Intelligence: Foundations, Theory, and Algorithms.

[16]  Eduardo J. Molinos,et al.  Automation of the UNICARagil Vehicles , 2020 .

[17]  Hermann Winner,et al.  UNICARagil - Disruptive Modular Architectures for Agile, Automated Vehicle Concepts , 2018 .

[18]  Qiang Xu,et al.  nuScenes: A Multimodal Dataset for Autonomous Driving , 2019, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[19]  Michael Aeberhard,et al.  Object-level fusion for surround environment perception in automated driving applications , 2017 .

[20]  Lutz Eckstein,et al.  Deep, spatially coherent Inverse Sensor Models with Uncertainty Incorporation using the evidential Framework , 2019, 2019 IEEE Intelligent Vehicles Symposium (IV).

[21]  P. Barry Deep , 2018, 2018 11th International Conference on Human System Interaction (HSI).

[22]  Jiong Yang,et al.  PointPillars: Fast Encoders for Object Detection From Point Clouds , 2018, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[23]  Arthur P. Dempster,et al.  A Generalization of Bayesian Inference , 1968, Classic Works of the Dempster-Shafer Theory of Belief Functions.

[24]  Dushyant Rao,et al.  Deep tracking in the wild: End-to-end tracking using recurrent neural networks , 2018, Int. J. Robotics Res..

[25]  Glenn Shafer,et al.  A Mathematical Theory of Evidence , 2020, A Mathematical Theory of Evidence.

[26]  Dimitris N. Metaxas,et al.  MotionNet: Joint Perception and Motion Prediction for Autonomous Driving Based on Bird’s Eye View Maps , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).