High-speed object detection with a single-photon time-of-flight image sensor

3D time-of-flight (ToF) imaging is used in a variety of applications such as augmented reality (AR), computer interfaces, robotics and autonomous systems. Single-photon avalanche diodes (SPADs) are one of the enabling technologies providing accurate depth data even over long ranges. By developing SPADs in array format with integrated processing combined with pulsed, flood-type illumination, high-speed 3D capture is possible. However, array sizes tend to be relatively small, limiting the lateral resolution of the resulting depth maps and, consequently, the information that can be extracted from the image for applications such as object detection. In this paper, we demonstrate that these limitations can be overcome through the use of convolutional neural networks (CNNs) for high-performance object detection. We present outdoor results from a portable SPAD camera system that outputs 16-bin photon timing histograms with 64×32 spatial resolution, with each histogram containing thousands of photons. The results, obtained with exposure times down to 2 ms (equivalent to 500 FPS) and in signal-to-background (SBR) ratios as low as 0.05, point to the advantages of providing the CNN with full histogram data rather than point clouds alone. Alternatively, a combination of point cloud and active intensity data may be used as input, for a similar level of performance. In either case, the GPU-accelerated processing time is less than 1 ms per frame, leading to an overall latency (image acquisition plus processing) in the millisecond range, making the results relevant for safety-critical computer vision applications which would benefit from faster than human reaction times.

[1]  Tao Deng,et al.  Image Transmission Based on Spiking Dynamics of Electrically Controlled VCSEL-SA Neuron , 2021, Photonics.

[2]  Kevin Barraclough,et al.  I and i , 2001, BMJ : British Medical Journal.

[3]  Naimul Mefraz Khan,et al.  A Novel Focal Tversky Loss Function With Improved Attention U-Net for Lesion Segmentation , 2018, 2019 IEEE 16th International Symposium on Biomedical Imaging (ISBI 2019).

[4]  Gordon Wetzstein,et al.  SPADnet: deep RGB-SPAD sensor fusion assisted by monocular depth estimation. , 2020, Optics express.

[5]  Markus Ulrich,et al.  Machine Vision Algorithms and Applications , 2007 .

[6]  Xiaogang Wang,et al.  PV-RCNN: Point-Voxel Feature Set Abstraction for 3D Object Detection , 2019, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[7]  Brian G. Stewart,et al.  Deep Learning for LiDAR Waveforms with Multiple Returns , 2020, 2020 28th European Signal Processing Conference (EUSIPCO).

[8]  You Li,et al.  Lidar for Autonomous Driving: The Principles, Challenges, and Trends for Automotive Lidar and Perception Systems , 2020, IEEE Signal Processing Magazine.

[9]  David Stoppa,et al.  A CMOS Range Camera based on Single Photon Avalanche Diodes (CMOS-Range-Kamera mit Single-Photon-Avalanche-Photodioden) , 2007 .

[10]  Remo Sala,et al.  A Survey on 3D Cameras: Metrological Comparison of Time-of-Flight, Structured-Light and Active Stereoscopy Technologies , 2018, SpringerBriefs in Computer Science.

[11]  Yanan Sun,et al.  3DSSD: Point-Based 3D Single Stage Object Detector , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[12]  C. Niclass,et al.  A 100-m Range 10-Frame/s 340 $\,\times\, $96-Pixel Time-of-Flight Depth Sensor in 0.18-$\mu\hbox{m}$ CMOS , 2013, IEEE Journal of Solid-State Circuits.

[13]  Jian-Wei Pan,et al.  Single-photon imaging over 200  km , 2021, 2103.05860.

[14]  Vivek K Goyal,et al.  High-flux single-photon lidar , 2020 .

[15]  Ruili Wang,et al.  A comparison and analysis of RGB-D cameras' depth performance for robotics application , 2017, 2017 24th International Conference on Mechatronics and Machine Vision in Practice (M2VIP).

[16]  Saeed Afshar,et al.  Event-Based Processing of Single Photon Avalanche Diode Sensors , 2019, IEEE Sensors Journal.

[17]  Robert K. Henderson,et al.  A Reconfigurable 3-D-Stacked SPAD Imager With In-Pixel Histogramming for Flash LIDAR or High-Speed Time-of-Flight Imaging , 2019, IEEE Journal of Solid-State Circuits.

[18]  Gordon Wetzstein,et al.  Single-photon 3D imaging with deep sensor fusion , 2018, ACM Trans. Graph..

[19]  Vivek K. Goyal,et al.  A Few Photons Among Many: Unmixing Signal and Noise for Photon-Efficient Active Imaging , 2016, IEEE Transactions on Computational Imaging.

[20]  Zhaohui Wu,et al.  3D Face Recognition using Mapped Depth Images , 2005, 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'05) - Workshops.

[21]  Ji Wan,et al.  Multi-view 3D Object Detection Network for Autonomous Driving , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[22]  Thomas Brox,et al.  U-Net: Convolutional Networks for Biomedical Image Segmentation , 2015, MICCAI.

[23]  ISTVAN GYONGY,et al.  High-speed 3D sensing via hybrid-mode imaging and guided upsampling , 2020 .

[24]  Cyrill Stachniss,et al.  SuMa++: Efficient LiDAR-based Semantic SLAM , 2019, 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[25]  Mohammed Bennamoun,et al.  RGB-D image-based Object Detection: from Traditional Methods to Deep Learning Techniques , 2019, RGB-D Image Analysis and Processing.

[26]  Matěj Hejda,et al.  Ultrafast optical integration and pattern classification for neuromorphic photonics based on spiking VCSEL neurons , 2020, Scientific Reports.

[27]  Gordon Wetzstein,et al.  Disambiguating Monocular Depth Estimation with a Single Transient , 2020, ECCV.

[28]  Mohit Gupta,et al.  Asynchronous Single-Photon 3D Imaging , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).

[29]  HoraudRadu,et al.  An overview of depth cameras and range scanners based on time-of-flight technologies , 2016 .

[30]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[31]  Xavier Rottenberg,et al.  Benchmarking time-of-flight based depth measurement techniques , 2016, SPIE OPTO.

[32]  Ali Farhadi,et al.  YOLOv3: An Incremental Improvement , 2018, ArXiv.

[33]  Min H. Kim,et al.  DeepToF: off-the-shelf real-time correction of multipath interference in time-of-flight imaging , 2017, ACM Trans. Graph..

[34]  Jonathan Leach,et al.  High-speed object detection using SPAD sensors , 2021, OPTO.

[35]  Silvio Savarese,et al.  Generalized Intersection Over Union: A Metric and a Loss for Bounding Box Regression , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[36]  David J. S. Birch,et al.  A $192\times128$ Time Correlated SPAD Image Sensor in 40-nm CMOS Technology , 2019, IEEE Journal of Solid-State Circuits.

[37]  Martin Wolf,et al.  A 30-frames/s, $252\times144$ SPAD Flash LiDAR With 1728 Dual-Clock 48.8-ps TDCs, and Pixel-Wise Integrated Histogramming , 2019, IEEE Journal of Solid-State Circuits.

[38]  Robert K. Henderson,et al.  Robust super-resolution depth imaging via a multi-feature fusion deep network , 2020, Optics express.

[39]  Roderick Murray-Smith,et al.  Spatial images from temporal data , 2019, Optica.

[40]  Radu Horaud,et al.  An overview of depth cameras and range scanners based on time-of-flight technologies , 2016, Machine Vision and Applications.

[41]  Tareq Abed Mohammed,et al.  Understanding of a convolutional neural network , 2017, 2017 International Conference on Engineering and Technology (ICET).

[42]  Gerald S. Buller,et al.  Full Waveform LiDAR for Adverse Weather Conditions , 2020, IEEE Transactions on Vehicular Technology.