Focus Is All You Need: Loss Functions for Event-Based Vision

Event cameras are novel vision sensors that output pixel-level brightness changes ("events") instead of traditional video frames. These asynchronous sensors offer several advantages over traditional cameras, such as, high temporal resolution, very high dynamic range, and no motion blur. To unlock the potential of such sensors, motion compensation methods have been recently proposed. We present a collection and taxonomy of twenty two objective functions to analyze event alignment in motion compensation approaches. We call them focus loss functions since they have strong connections with functions used in traditional shape-from-focus applications. The proposed loss functions allow bringing mature computer vision tools to the realm of event cameras. We compare the accuracy and runtime performance of all loss functions on a publicly available dataset, and conclude that the variance, the gradient and the Laplacian magnitudes are among the best loss functions. The applicability of the loss functions is shown on multiple tasks: rotational motion, depth and optical flow estimation. The proposed focus loss functions allow to unlock the outstanding properties of event cameras.

[1]  Narciso García,et al.  Optimal Piecewise Linear Function Approximation for GPU-Based Applications , 2015, IEEE Transactions on Cybernetics.

[2]  Lindsay Kleeman,et al.  Simultaneous Optical Flow and Segmentation (SOFAS) using Dynamic Vision Sensor , 2018, ICRA 2018.

[3]  Tobi Delbrück,et al.  The event-camera dataset and simulator: Event-based data for pose estimation, visual odometry, and SLAM , 2016, Int. J. Robotics Res..

[4]  Tobi Delbrück,et al.  EV-IMO: Motion Segmentation Dataset and Learning Pipeline for Event Cameras , 2019, 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[5]  Tobi Delbrück,et al.  A 128$\times$ 128 120 dB 15 $\mu$s Latency Asynchronous Temporal Contrast Vision Sensor , 2008, IEEE Journal of Solid-State Circuits.

[6]  Kostas Daniilidis,et al.  Event-Based Visual Inertial Odometry , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[7]  Alexander Andreopoulos,et al.  A Low Power, High Throughput, Fully Event-Based Stereo System , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[8]  J. Neumann Distribution of the Ratio of the Mean Square Successive Difference to the Variance , 1941 .

[9]  Garrick Orchard,et al.  HOTS: A Hierarchy of Event-Based Time-Surfaces for Pattern Recognition , 2017, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[10]  Tobi Delbrück,et al.  Asynchronous Event-Based Binocular Stereo Matching , 2012, IEEE Transactions on Neural Networks and Learning Systems.

[11]  Davide Scaramuzza,et al.  Event-based, 6-DOF pose tracking for high-speed maneuvers , 2014, 2014 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[12]  Jörg Conradt,et al.  Simultaneous Localization and Mapping for Event-Based Vision Systems , 2013, ICVS.

[13]  Margarita Chli,et al.  Asynchronous Corner Detection and Tracking for Event Cameras in Real Time , 2018, IEEE Robotics and Automation Letters.

[14]  Tom Drummond,et al.  Event-Based Motion Segmentation by Motion Compensation , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).

[15]  Ryad Benosman,et al.  Visual Tracking Using Neuromorphic Asynchronous Event-Based Cameras , 2015, Neural Computation.

[16]  Davide Scaramuzza,et al.  Ultimate SLAM? Combining Events, Images, and IMU for Robust Visual SLAM in HDR and High-Speed Scenarios , 2017, IEEE Robotics and Automation Letters.

[17]  Yiannis Aloimonos,et al.  Event-Based Moving Object Detection and Tracking , 2018, 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[18]  Francesco Fedele,et al.  Variational Stereo Imaging of Oceanic Waves With Statistical Constraints , 2013, IEEE Transactions on Image Processing.

[19]  Davide Scaramuzza,et al.  Continuous-Time Visual-Inertial Odometry for Event Cameras , 2017, IEEE Transactions on Robotics.

[20]  K Cook,et al.  Comparison of autofocus methods for automated microscopy. , 1991, Cytometry.

[21]  Davide Scaramuzza,et al.  EMVS: Event-Based Multi-View Stereo—3D Reconstruction with an Event Camera in Real-Time , 2017, International Journal of Computer Vision.

[22]  Nitish V. Thakor,et al.  HFirst: A Temporal Approach to Object Recognition , 2015, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[23]  Davide Scaramuzza,et al.  Low-latency event-based visual odometry , 2014, 2014 IEEE International Conference on Robotics and Automation (ICRA).

[24]  Tobi Delbrück,et al.  Real-Time Gesture Interface Based on Event-Driven Processing From Stereo Silicon Retinas , 2014, IEEE Transactions on Neural Networks and Learning Systems.

[25]  Kostas Daniilidis,et al.  Realtime Time Synchronized Event-based Stereo , 2018, ECCV.

[26]  Tobi Delbruck,et al.  A 240 × 180 130 dB 3 µs Latency Global Shutter Spatiotemporal Vision Sensor , 2014, IEEE Journal of Solid-State Circuits.

[27]  Davide Scaramuzza,et al.  Accurate Angular Velocity Estimation With an Event Camera , 2017, IEEE Robotics and Automation Letters.

[28]  P. Moran Notes on continuous stochastic phenomena. , 1950, Biometrika.

[29]  Narciso García Santos,et al.  Optimal Piecewise Linear Function Approximation for GPU-based Applications , 2015 .

[30]  Lindsay Kleeman,et al.  Event Cameras, Contrast Maximization and Reward Functions: An Analysis , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[31]  Ryad Benosman,et al.  A spiking neural network model of 3D perception for event-based neuromorphic stereo vision systems , 2017, Scientific Reports.

[32]  Narciso García Santos,et al.  Optimal polygonal L1 linearization and fast interpolation of nonlinear systems , 2014 .

[33]  Tobi Delbrück,et al.  A Low Power, Fully Event-Based Gesture Recognition System , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[34]  Yiannis Aloimonos,et al.  Unsupervised Learning of Dense Optical Flow and Depth from Sparse Event Data , 2018, ArXiv.

[35]  Narciso García,et al.  Optimal Polygonal $L_{1}$ Linearization and Fast Interpolation of Nonlinear Systems , 2013, IEEE Transactions on Circuits and Systems I: Regular Papers.

[36]  Kostas Daniilidis,et al.  Unsupervised Event-Based Learning of Optical Flow, Depth, and Egomotion , 2018, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[37]  Ryad Benosman,et al.  HATS: Histograms of Averaged Time Surfaces for Robust Event-Based Object Classification , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[38]  Peter van Beek,et al.  An extensive empirical evaluation of focus measures for digital photography , 2014, Electronic Imaging.

[39]  Horst Bischof,et al.  Event-driven stereo matching for real-time 3D panoramic vision , 2015, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[40]  Chiara Bartolozzi,et al.  Robust visual tracking with a freely-moving event camera , 2017, 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[41]  Davide Scaramuzza,et al.  Feature detection and tracking with the dynamic and active-pixel vision sensor (DAVIS) , 2016, 2016 Second International Conference on Event-based Control, Communication, and Signal Processing (EBCCSP).

[42]  Domenec Puig,et al.  Analysis of focus measure operators for shape-from-focus , 2013, Pattern Recognit..

[43]  Kostas Daniilidis,et al.  Event-based feature tracking with probabilistic data association , 2017, 2017 IEEE International Conference on Robotics and Automation (ICRA).

[44]  Shengyong Chen,et al.  Event-Based Stereo Depth Estimation Using Belief Propagation , 2017, Front. Neurosci..

[45]  R. Geary,et al.  The Contiguity Ratio and Statistical Mapping , 1954 .

[46]  Stefan Leutenegger,et al.  Real-Time 3D Reconstruction and 6-DoF Tracking with an Event Camera , 2016, ECCV.

[47]  Feng Shi,et al.  Context-aware event-driven stereo matching , 2016, 2016 IEEE International Conference on Image Processing (ICIP).

[48]  Davide Scaramuzza,et al.  A Unifying Contrast Maximization Framework for Event Cameras, with Applications to Motion, Depth, and Optical Flow Estimation , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[49]  Mohit Gupta,et al.  MC3D: Motion Contrast 3D Scanning , 2015, 2015 IEEE International Conference on Computational Photography (ICCP).

[50]  Kostas Daniilidis,et al.  EV-FlowNet: Self-Supervised Optical Flow Estimation for Event-based Cameras , 2018, Robotics: Science and Systems.

[51]  Ryad Benosman,et al.  Simultaneous Mosaicing and Tracking with an Event Camera , 2014, BMVC.

[52]  Davide Scaramuzza,et al.  EVO: A Geometric Approach to Event-Based 6-DOF Parallel Tracking and Mapping in Real Time , 2017, IEEE Robotics and Automation Letters.

[53]  Davide Scaramuzza,et al.  Event-Based, 6-DOF Camera Tracking from Photometric Depth Maps , 2016, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[54]  Thomas Pock,et al.  Real-time panoramic tracking for event cameras , 2017, 2017 IEEE International Conference on Computational Photography (ICCP).

[55]  Davide Scaramuzza,et al.  Asynchronous, Photometric Feature Tracking using Events and Frames , 2018, ECCV.

[56]  T. Delbruck,et al.  > Replace This Line with Your Paper Identification Number (double-click Here to Edit) < 1 , 2022 .

[57]  Yiannis Aloimonos,et al.  Unsupervised Learning of Dense Optical Flow, Depth and Egomotion from Sparse Event Data , 2018 .

[58]  P. J. Narayanan,et al.  Composite Focus Measure for High Quality Depth Maps , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).

[59]  Yi Zhou,et al.  Semi-Dense 3D Reconstruction with a Stereo Event Camera , 2018, ECCV.