RSS-Net: Weakly-Supervised Multi-Class Semantic Segmentation with FMCW Radar

This paper presents an efficient annotation procedure and an application thereof to end-to-end, rich semantic segmentation of the sensed environment using Frequency-Modulated Continuous-Wave scanning radar. We advocate radar over the traditional sensors used for this task as it operates at longer ranges and is substantially more robust to adverse weather and illumination conditions. We avoid laborious manual labelling by exploiting the largest radar-focused urban autonomy dataset collected to date, correlating radar scans with RGB cameras and LiDAR sensors, for which semantic segmentation is an already consolidated procedure. The training procedure leverages a state-of-the-art natural image segmentation system which is publicly available and as such, in contrast to previous approaches, allows for the production of copious labels for the radar stream by incorporating four camera and two LiDAR streams. Additionally, the losses are computed taking into account labels to the radar sensor horizon by accumulating LiDAR returns along a pose-chain ahead and behind of the current vehicle position. Finally, we present the network with multi-channel radar scan inputs in order to deal with ephemeral and dynamic scene objects.

[1]  Paul Newman,et al.  Radar-only ego-motion estimation in difficult settings via graph matching , 2019, 2019 International Conference on Robotics and Automation (ICRA).

[2]  Paul Newman,et al.  RSL-Net: Localising in Satellite Images From a Radar on the Ground , 2020, IEEE Robotics and Automation Letters.

[3]  Paul Newman,et al.  Probably Unknown: Deep Inverse Sensor Modelling Radar , 2018, 2019 International Conference on Robotics and Automation (ICRA).

[4]  Paul Newman,et al.  The Oxford Radar RobotCar Dataset: A Radar Extension to the Oxford RobotCar Dataset , 2020, 2020 IEEE International Conference on Robotics and Automation (ICRA).

[5]  Jürgen Dickmann,et al.  Semantic Segmentation on Radar Point Clouds , 2018, 2018 21st International Conference on Information Fusion (FUSION).

[6]  George Papandreou,et al.  Searching for Efficient Multi-Scale Architectures for Dense Image Prediction , 2018, NeurIPS.

[7]  Ingmar Posner,et al.  Driven to Distraction: Self-Supervised Distractor Learning for Robust Monocular Visual Odometry in Urban Environments , 2017, 2018 IEEE International Conference on Robotics and Automation (ICRA).

[8]  Ingmar Posner,et al.  Masking by Moving: Learning Distraction-Free Radar Odometry from Pose Information , 2019, CoRL.

[9]  George Papandreou,et al.  Encoder-Decoder with Atrous Separable Convolution for Semantic Image Segmentation , 2018, ECCV.

[10]  Zhi-Hua Zhou,et al.  A brief introduction to weakly supervised learning , 2018 .

[11]  Paul Newman,et al.  FAB-MAP: Probabilistic Localization and Mapping in the Space of Appearance , 2008, Int. J. Robotics Res..

[12]  Jürgen Dickmann,et al.  Semantic radar grids , 2017, 2017 IEEE Intelligent Vehicles Symposium (IV).

[13]  Paul Newman,et al.  Fast Radar Motion Estimation with a Learnt Focus of Attention using Weak Supervision , 2019, 2019 International Conference on Robotics and Automation (ICRA).

[14]  Jinyong Jeong,et al.  MulRan: Multimodal Range Dataset for Urban Place Recognition , 2020, 2020 IEEE International Conference on Robotics and Automation (ICRA).

[15]  Sebastian Ramos,et al.  The Cityscapes Dataset for Semantic Urban Scene Understanding , 2016, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[16]  Paul Newman,et al.  1 year, 1000 km: The Oxford RobotCar dataset , 2017, Int. J. Robotics Res..

[17]  Matthew Gadd,et al.  Kidnapped Radar: Topological Radar Localisation using Rotationally-Invariant Metric Learning , 2020, 2020 IEEE International Conference on Robotics and Automation (ICRA).

[18]  Bo Chen,et al.  MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications , 2017, ArXiv.

[19]  George Papandreou,et al.  Rethinking Atrous Convolution for Semantic Image Segmentation , 2017, ArXiv.

[20]  Ken Shoemake,et al.  Animating rotation with quaternion curves , 1985, SIGGRAPH.

[21]  Matthew Gadd,et al.  Look Around You: Sequence-based Radar Place Recognition with Learned Rotational Invariance , 2020, 2020 IEEE/ION Position, Location and Navigation Symposium (PLANS).

[22]  Thomas Brox,et al.  U-Net: Convolutional Networks for Biomedical Image Segmentation , 2015, MICCAI.

[23]  Ba-Ngu Vo,et al.  Robotic Navigation and Mapping with Radar , 2012 .

[24]  Nils Appenrodt,et al.  Radar-based Road User Classification and Novelty Detection with Recurrent Neural Network Ensembles , 2019, 2019 IEEE Intelligent Vehicles Symposium (IV).

[25]  Jürgen Dickmann,et al.  Radar-based Feature Design and Multiclass Classification for Road User Recognition , 2018, 2018 IEEE Intelligent Vehicles Symposium (IV).

[26]  Matthew Gadd,et al.  The Hulk: Design and Development of a Weather-Proof Vehicle for Long-Term Autonomy in Outdoor Environments , 2019, FSR.

[27]  Paul Newman,et al.  What Could Go Wrong? Introspective Radar Odometry in Challenging Environments , 2019, 2019 IEEE Intelligent Transportation Systems Conference (ITSC).

[28]  Frédéric Jurie,et al.  Recent Advances in Object Detection in the Age of Deep Convolutional Neural Networks , 2018, ArXiv.

[29]  Paul Newman,et al.  Precise Ego-Motion Estimation with Millimeter-Wave Radar Under Diverse and Challenging Conditions , 2018, 2018 IEEE International Conference on Robotics and Automation (ICRA).

[30]  Klaus Dietmayer,et al.  Dynamic Occupancy Grid Prediction for Urban Autonomous Driving: A Deep Learning Approach with Fully Automatic Labeling , 2017, 2018 IEEE International Conference on Robotics and Automation (ICRA).