DENSE: Environment Perception in Bad Weather—First Results

This paper presents the first results of the publicly funded ECSEL project DENSE (Adverse weather environmental sensing system). DENSE seeks to eliminate one of the most pressing problems of automated driving: the inability of current systems to sense their surroundings under all weather conditions. The task in DENSE is to develop a sensor suite for automatic driving, by means of which the vehicle environment can be reliably detected 24/7 under these bad weather conditions. In this paper, the state of the art of environmental sensor technology is first examined and evaluated in the CEREMA weather chamber. Then, the architecture of the DENSE Sensor Suite is presented and the development results of the most important system components are described. The results show that the realization of a 24/7 all-weather sensor suite is absolutely feasible with these components .

[1]  John P. Oakley,et al.  Improving image quality in poor visibility conditions using a physical model for contrast degradation , 1998, IEEE Trans. Image Process..

[2]  Matti Kutila,et al.  Automotive LiDAR performance verification in fog and rain , 2018, 2018 21st International Conference on Intelligent Transportation Systems (ITSC).

[3]  Matti Kutila,et al.  Spectral attenuation in low visibility artificial fog: Experimental study and comparison to literature models , 2017, 2017 13th IEEE International Conference on Intelligent Computer Communication and Processing (ICCP).

[4]  Alexei A. Efros,et al.  The Unreasonable Effectiveness of Deep Features as a Perceptual Metric , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[5]  Wei Liu,et al.  SSD: Single Shot MultiBox Detector , 2015, ECCV.

[6]  Jan Kautz,et al.  High-Resolution Image Synthesis and Semantic Manipulation with Conditional GANs , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[7]  Werner Ritter,et al.  Benchmarking Image Sensors Under Adverse Weather Conditions for Autonomous Driving , 2018, 2018 IEEE Intelligent Vehicles Symposium (IV).

[8]  Shree K. Nayar,et al.  Removing weather effects from monochrome images , 2001, Proceedings of the 2001 IEEE Computer Society Conference on Computer Vision and Pattern Recognition. CVPR 2001.

[9]  Werner Ritter,et al.  A Benchmark for Lidar Sensors in Fog: Is Detection Breaking Down? , 2018, 2018 IEEE Intelligent Vehicles Symposium (IV).

[10]  Luc Van Gool,et al.  Semantic Foggy Scene Understanding with Synthetic Data , 2017, International Journal of Computer Vision.

[11]  Klaus C. J. Dietmayer,et al.  Seeing Through Fog Without Seeing Fog: Deep Sensor Fusion in the Absence of Labeled Training Data , 2019, ArXiv.

[12]  Matti Kutila,et al.  Automotive LIDAR sensor development scenarios for harsh weather conditions , 2016, 2016 IEEE 19th International Conference on Intelligent Transportation Systems (ITSC).

[13]  Dariu Gavrila,et al.  Multi-cue pedestrian classification with partial occlusion handling , 2010, 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition.

[14]  Jean-Jacques Boreux,et al.  An innovative artificial fog production device improved in the European project “FOG” , 2008 .

[15]  Shree K. Nayar,et al.  Instant dehazing of images using polarization , 2001, Proceedings of the 2001 IEEE Computer Society Conference on Computer Vision and Pattern Recognition. CVPR 2001.

[16]  Klaus C. J. Dietmayer,et al.  Learning Super-resolved Depth from Active Gated Imaging , 2018, 2018 21st International Conference on Intelligent Transportation Systems (ITSC).

[17]  Klaus C. J. Dietmayer,et al.  Robustness Against Unknown Noise for Raw Data Fusing Neural Networks , 2018, 2018 21st International Conference on Intelligent Transportation Systems (ITSC).

[18]  Eero P. Simoncelli,et al.  Image quality assessment: from error visibility to structural similarity , 2004, IEEE Transactions on Image Processing.