A LiDAR Point Cloud Generator: from a Virtual World to Autonomous Driving

3D LiDAR scanners are playing an increasingly important role in autonomous driving as they can generate depth information of the environment. However, creating large 3D LiDAR point cloud datasets with point-level labels requires a significant amount of manual annotation. This jeopardizes the efficient development of supervised deep learning algorithms which are often data-hungry. We present a framework to rapidly create point clouds with accurate point-level labels from a computer game. To our best knowledge, this is the first publication on LiDAR point cloud simulation framework for autonomous driving. The framework supports data collection from both auto-driving scenes and user-configured scenes. Point clouds from auto-driving scenes can be used as training data for deep learning algorithms, while point clouds from user-configured scenes can be used to systematically test the vulnerability of a neural network, and use the falsifying examples to make the neural network more robust through retraining. In addition, the scene images can be captured simultaneously in order for sensor fusion tasks, with a method proposed to do automatic registration between the point clouds and captured scene images. We show a significant improvement in accuracy (+9%) in point cloud segmentation by augmenting the training dataset with the generated synthesized data. Our experiments also show by testing and retraining the network using point clouds from user-configured scenes, the weakness/blind spots of the neural network can be fixed.

[1]  Trevor Darrell,et al.  Active Learning with Gaussian Processes for Object Categorization , 2007, 2007 IEEE 11th International Conference on Computer Vision.

[2]  Christoph Stiller,et al.  Segmentation of 3D lidar data in non-flat urban environments using a local convexity criterion , 2009, 2009 IEEE Intelligent Vehicles Symposium.

[3]  Pietro Perona,et al.  Online crowdsourcing: Rating annotators and obtaining cost-effective labels , 2010, 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition - Workshops.

[4]  Doug A. Bowman,et al.  Rapid and accurate 3D selection by progressive refinement , 2011, 2011 IEEE Symposium on 3D User Interfaces (3DUI).

[5]  Ghassan Hamarneh,et al.  Active Learning for Interactive 3D Image Segmentation , 2011, MICCAI.

[6]  Yue Gao,et al.  3-D Object Retrieval and Recognition With Hypergraph Analysis , 2012, IEEE Transactions on Image Processing.

[7]  Kun Zhou,et al.  An interactive approach to semantic modeling of indoor scenes with an RGBD camera , 2012, ACM Trans. Graph..

[8]  Derek Hoiem,et al.  Indoor Segmentation and Support Inference from RGBD Images , 2012, ECCV.

[9]  Andreas Geiger,et al.  Vision meets robotics: The KITTI dataset , 2013, Int. J. Robotics Res..

[10]  Slobodan Ilic,et al.  Framework for Generation of Synthetic Ground Truth Data for Driver Assistance Applications , 2013, GCPR.

[11]  Antonio Capobianco,et al.  Go'Then'Tag: A 3-D point cloud annotation technique , 2014, 2014 IEEE Symposium on 3D User Interfaces (3DUI).

[12]  P. Peixoto,et al.  Road Detection Using High Resolution LIDAR , 2014, 2014 IEEE Vehicle Power and Propulsion Conference (VPPC).

[13]  Aleksey Boyko,et al.  Cheaper by the dozen: group annotation of 3D data , 2014, UIST.

[14]  Wen Gao,et al.  Mining Compact Bag-of-Patterns for Low Bit Rate Mobile Visual Search , 2014, IEEE Transactions on Image Processing.

[15]  Niloy J. Mitra,et al.  SmartAnnotator: An Interactive Tool for Annotating RGBD Indoor Images , 2014, ArXiv.

[16]  Thomas A. Funkhouser,et al.  Learning Hierarchical Semantic Segmentations of LIDAR Data , 2015, 2015 International Conference on 3D Vision.

[17]  Hongxun Yao,et al.  View-based 3D object retrieval via multi-modal graph learning , 2015, Signal Process..

[18]  Niloy J. Mitra,et al.  SmartAnnotator An Interactive Tool for Annotating Indoor RGBD Images , 2015, Comput. Graph. Forum.

[19]  Hongxun Yao,et al.  Strategy for dynamic 3D depth data matching towards robust action retrieval , 2015, Neurocomputing.

[20]  Daniel H. Biedermann,et al.  Evaluating visual ADAS components on the COnGRATS dataset , 2016, 2016 IEEE Intelligent Vehicles Symposium (IV).

[21]  Vladlen Koltun,et al.  Playing for Data: Ground Truth from Computer Games , 2016, ECCV.

[22]  Alberto L. Sangiovanni-Vincentelli,et al.  Systematic Testing of Convolutional Neural Networks for Autonomous Driving , 2017, ArXiv.

[23]  Germán Ros,et al.  CARLA: An Open Urban Driving Simulator , 2017, CoRL.

[24]  Matthew Johnson-Roberson,et al.  Driving in the Matrix: Can virtual worlds replace human-generated annotations for real world tasks? , 2016, 2017 IEEE International Conference on Robotics and Automation (ICRA).

[25]  Wolfram Burgard,et al.  Deep semantic classification for 3D LiDAR data , 2017, 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[26]  Vladlen Koltun,et al.  Playing for Benchmarks , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).

[27]  Nanning Zheng,et al.  Detecting Drivable Area for Self-driving Cars: An Unsupervised Approach , 2017, ArXiv.

[28]  Sanjit A. Seshia,et al.  Compositional Falsification of Cyber-Physical Systems with Machine Learning Components , 2017, NFM.

[29]  Nanning Zheng,et al.  A Co-Point Mapping-Based Approach to Drivable Area Detection for Self-Driving Cars , 2018, Engineering.

[30]  Kurt Keutzer,et al.  SqueezeSeg: Convolutional Neural Nets with Recurrent CRF for Real-Time Road-Object Segmentation from 3D LiDAR Point Cloud , 2017, 2018 IEEE International Conference on Robotics and Automation (ICRA).