Improving in-car emotion classification by NIR database augmentation

On-board detection of driver’s emotions has become a task of high importance for car manufacturers, as negative emotions appear to be one of the major risks for car accidents. Deep neural networks have become over the last years the state of the art methods for computer vision and image classification. Yet, their success depends upon their being trained on a comprehensive database, which should cover all of the real-life situations that may arise in practice. Most of the in-car driver monitoring cameras capture images in the near infra-red (NIR) domain therefore one needs a large database with images featuring emotions in the NIR domain. As most databases featuring human emotions contain images acquired in the visible domain, we discuss in this paper two methods of transferring the NIR-like look into the "visible" images, by using a CycleGAN style-transferring neural network trained using "paired" and "unpaired" images. We show that the resulted database augmented with NIR-like images leads to a much improved performance in emotion classification for a deep neural network, when tested on real NIR images.

[1]  Paul A. Jennings,et al.  Towards hybrid driver state monitoring: Review, future perspectives and the role of consumer electronics , 2016, 2016 IEEE Intelligent Vehicles Symposium (IV).

[2]  Fei-Fei Li,et al.  ImageNet: A large-scale hierarchical image database , 2009, 2009 IEEE Conference on Computer Vision and Pattern Recognition.

[3]  Alexei A. Efros,et al.  Image-to-Image Translation with Conditional Adversarial Networks , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[4]  Yoshua Bengio,et al.  Generative Adversarial Nets , 2014, NIPS.

[5]  Li Fei-Fei,et al.  Perceptual Losses for Real-Time Style Transfer and Super-Resolution , 2016, ECCV.

[6]  Angel Domingo Sappa,et al.  Infrared Image Colorization Based on a Triplet DCGAN Architecture , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW).

[7]  Yuan-Hsiang Lin,et al.  A Driver's Physiological Monitoring System Based on a Wearable PPG Sensor and a Smartphone , 2011, SUComS.

[8]  Andrew Zisserman,et al.  Very Deep Convolutional Networks for Large-Scale Image Recognition , 2014, ICLR.

[9]  Michael Felsberg,et al.  Generating Visible Spectrum Images from Thermal Infrared , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW).

[10]  拓海 杉山,et al.  “Unpaired Image-to-Image Translation using Cycle-Consistent Adversarial Networks”の学習報告 , 2017 .

[11]  ImageNet Classification with Deep Convolutional Neural , 2013 .

[12]  Geoffrey E. Hinton,et al.  ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.

[13]  Yong Yu,et al.  Unsupervised Diverse Colorization via Generative Adversarial Networks , 2017, ECML/PKDD.

[14]  Yorgos Goletsis,et al.  A wearable system for the affective monitoring of car racing drivers during simulated conditions , 2011 .

[15]  Feng Guo,et al.  Driver crash risk factors and prevalence evaluation using naturalistic driving data , 2016, Proceedings of the National Academy of Sciences.