SemAlign: Annotation-Free Camera-LiDAR Calibration with Semantic Alignment Loss

Multi-sensor solution has been widely adopted in real-world robotics systems (e.g., self-driving vehicles) due to its better robustness. However, its performance is highly dependent on the accurate calibration between different sensors, which is very time-consuming (i.e., hours of human efforts) to acquire. Recent learning-based solutions partially address this yet still require costly ground-truth annotations as supervision. In this paper, we introduce a novel self-supervised semantic alignment loss to quantitatively measure the quality of a given calibration. It is well correlated with conventional evaluation metrics while it does not require ground-truth calibration annotations as the reference. Based on this loss, we further propose an annotation-free optimization-based calibration algorithm (SemAlign) that first estimates a coarse calibration with loss-guided initialization and then refines it with gradient-based optimization. SemAlign reduces the calibration time from hours of human efforts to only seconds of GPU computation. It not only achieves comparable performance with existing supervised learning frameworks but also demonstrates a much better generalization capability when transferred to a different dataset.

[1]  K. S. Nagla,et al.  Improved auto-extrinsic calibration between stereo vision camera and laser range finder , 2020 .

[2]  K. Madhava Krishna,et al.  CalibNet: Geometrically Supervised Extrinsic Calibration using 3D Spatial Transformer Networks , 2018, 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[3]  Shawn D. Newsam,et al.  Improving Semantic Segmentation via Video Propagation and Label Relaxation , 2018, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[4]  Adam Herout,et al.  Calibration of RGB camera with velodyne LiDAR , 2014 .

[5]  Levente Hajder,et al.  Accurate Calibration of Multi-LiDAR-Multi-Camera Systems , 2018, Sensors.

[6]  Srikanth Saripalli,et al.  Extrinsic Calibration of a 3D-LIDAR and a Camera , 2020, 2020 IEEE Intelligent Vehicles Symposium (IV).

[7]  Csaba Benedek,et al.  On-the-Fly Camera and Lidar Calibration , 2020, Remote. Sens..

[8]  Silvio Savarese,et al.  Automatic Extrinsic Calibration of Vision and Lidar by Maximizing Mutual Information , 2015, J. Field Robotics.

[9]  Michael Buchholz,et al.  Online Extrinsic Calibration Based on Per-Sensor Ego-Motion Using Dual Quaternions , 2021, IEEE Robotics and Automation Letters.

[10]  Gerd Hirzinger,et al.  Optimal Hand-Eye Calibration , 2006, 2006 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[11]  Cyrill Stachniss,et al.  Extrinsic multi-sensor calibration for mobile robots using the Gauss-Helmert model , 2017, 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[12]  Yiu Cheung Shiu,et al.  Calibration of wrist-mounted robotic sensors by solving homogeneous transform equations of the form AX=XB , 1989, IEEE Trans. Robotics Autom..

[13]  Nakju Lett Doh,et al.  Automatic targetless camera–LIDAR calibration by aligning edge with Gaussian mixture model , 2019, J. Field Robotics.

[14]  Fernando García,et al.  Automatic extrinsic calibration for lidar-stereo vehicle sensor setups , 2017, 2017 IEEE 20th International Conference on Intelligent Transportation Systems (ITSC).

[15]  Andreas Geiger,et al.  Automatic camera and range sensor calibration using a single shot , 2012, 2012 IEEE International Conference on Robotics and Automation.

[16]  Song Han,et al.  Searching Efficient 3D Architectures with Sparse Point-Voxel Convolution , 2020, ECCV.

[17]  Zoltan Kato,et al.  Targetless Calibration of a Lidar - Perspective Camera Pair , 2013, 2013 IEEE International Conference on Computer Vision Workshops.

[18]  Song Han,et al.  Point-Voxel CNN for Efficient 3D Deep Learning , 2019, NeurIPS.

[19]  Sebastian Ramos,et al.  The Cityscapes Dataset for Semantic Urban Scene Understanding , 2016, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[20]  Dong Ye,et al.  LCCNet: LiDAR and Camera Self-Calibration using Cost Volume Network , 2021, 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW).

[21]  Michael Bosse,et al.  Line-based extrinsic calibration of range and image sensors , 2013, 2013 IEEE International Conference on Robotics and Automation.

[22]  Andreas Geiger,et al.  Are we ready for autonomous driving? The KITTI vision benchmark suite , 2012, 2012 IEEE Conference on Computer Vision and Pattern Recognition.

[23]  Eung-Su Kim,et al.  Extrinsic Calibration between Camera and LiDAR Sensors by Matching Multiple 3D Planes † , 2019, Sensors.

[24]  Takeo Kanade,et al.  Extrinsic calibration of a single line scanning lidar and a camera , 2011, 2011 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[25]  Cyrill Stachniss,et al.  SemanticKITTI: A Dataset for Semantic Scene Understanding of LiDAR Sequences , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).

[26]  Kyungeun Cho,et al.  Calibration between Color Camera and 3D LIDAR Instruments with a Polygonal Planar Board , 2014, Sensors.

[27]  Nick Schneider,et al.  RegNet: Multimodal sensor registration using deep neural networks , 2017, 2017 IEEE Intelligent Vehicles Symposium (IV).

[28]  Sridha Sridharan,et al.  Spatiotemporal Camera-LiDAR Calibration: A Targetless and Structureless Approach , 2020, IEEE Robotics and Automation Letters.

[29]  Brett Browning,et al.  Automatic Calibration of a Range Sensor and Camera System , 2012, 2012 Second International Conference on 3D Imaging, Modeling, Processing, Visualization & Transmission.