MIND-Net: A Deep Mutual Information Distillation Network for Realistic Low-Resolution Face Recognition

Realistic low-resolution (LR) face images refer to those captured by the real-world surveillance cameras at extreme standoff distances, thereby LR and poor in quality essentially. Owing to severe scarcity of labeled data, a high-capacity deep convolution neural networks (CNN) is hardly trained to confront the realistic LR face recognition (LRFR) challenge. We introduce in this letter a dual-stream mutual information distillation network (MIND-Net), whereby the non-identity specific mutual information (MI) characterized by generic face features coexistent on realistic and synthetic LR face images are distilled to render a resolution-invariant embedding space for LRFR. For a thorough analysis, we quantify the degree of MI distillation in terms normalized MI index. Our experimental results on the realistic LR face datasets substantiate that the MIND-Net instances assembled from the pre-learned CNNs stand out from the baselines and other state of the arts by a notable margin.

[1]  Mei Wang,et al.  Deep Face Recognition: A Survey , 2018, Neurocomputing.

[2]  Thomas S. Huang,et al.  Studying Very Low Resolution Recognition Using Deep Networks , 2016, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[3]  Wei Liu,et al.  Super-Identity Convolutional Neural Network for Face Hallucination , 2018, ECCV.

[4]  Yu Qiao,et al.  Joint Face Detection and Alignment Using Multitask Cascaded Convolutional Networks , 2016, IEEE Signal Processing Letters.

[5]  Xing Ji,et al.  CosFace: Large Margin Cosine Loss for Deep Face Recognition , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[6]  Omkar M. Parkhi,et al.  VGGFace2: A Dataset for Recognising Faces across Pose and Age , 2017, 2018 13th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2018).

[7]  Mengjiao Wang,et al.  Improved Knowledge Distillation for Training Fast Low Resolution Face Recognition Model , 2019, 2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW).

[8]  Mislav Grgic,et al.  SCface – surveillance cameras face database , 2011, Multimedia Tools and Applications.

[9]  Shiming Ge,et al.  Low-Resolution Face Recognition in the Wild via Selective Knowledge Distillation , 2018, IEEE Transactions on Image Processing.

[10]  Stefanos Zafeiriou,et al.  ArcFace: Additive Angular Margin Loss for Deep Face Recognition , 2018, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[11]  Jian Zhao,et al.  Cross-Resolution Face Recognition via Prior-Aided Face Hallucination and Residual Knowledge Distillation , 2019, ArXiv.

[12]  Fabrizio Falchi,et al.  Cross-Resolution Learning for Face Recognition , 2020, Image Vis. Comput..

[13]  James Philbin,et al.  FaceNet: A unified embedding for face recognition and clustering , 2015, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[14]  Bhiksha Raj,et al.  SphereFace: Deep Hypersphere Embedding for Face Recognition , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[15]  Alex Krizhevsky,et al.  Learning Multiple Layers of Features from Tiny Images , 2009 .

[16]  Patrick J. Flynn,et al.  On Low-Resolution Face Recognition in the Wild: Comparisons and New Techniques , 2018, IEEE Transactions on Information Forensics and Security.

[17]  Ya Le,et al.  Tiny ImageNet Visual Recognition Challenge , 2015 .

[18]  Xu Ji,et al.  Invariant Information Clustering for Unsupervised Image Classification and Segmentation , 2018, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).

[19]  Patrick J. Flynn,et al.  Face Recognition in Low Quality Images: A Survey , 2018, ArXiv.

[20]  Yuxiao Hu,et al.  MS-Celeb-1M: A Dataset and Benchmark for Large-Scale Face Recognition , 2016, ECCV.

[21]  Xiaoming Liu,et al.  Improving Face Recognition from Hard Samples via Distribution Distillation Loss , 2020, ECCV.

[22]  Xudong Jiang,et al.  Deep Coupled ResNet for Low-Resolution Face Recognition , 2018, IEEE Signal Processing Letters.

[23]  Marwan Mattar,et al.  Labeled Faces in the Wild: A Database forStudying Face Recognition in Unconstrained Environments , 2008 .

[24]  Yonglong Tian,et al.  Contrastive Representation Distillation , 2019, ICLR.

[25]  Sergey Ioffe,et al.  Inception-v4, Inception-ResNet and the Impact of Residual Connections on Learning , 2016, AAAI.

[26]  Shengcai Liao,et al.  Learning Face Representation from Scratch , 2014, ArXiv.

[27]  Anil K. Jain,et al.  Towards Universal Representation Learning for Deep Face Recognition , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[28]  Geoffrey E. Hinton,et al.  Distilling the Knowledge in a Neural Network , 2015, ArXiv.

[29]  Shaogang Gong,et al.  Surveillance Face Recognition Challenge , 2018, ArXiv.

[30]  Alexander J. Smola,et al.  Multitask Learning without Label Correspondences , 2010, NIPS.

[31]  Shiming Ge,et al.  Efficient Low-Resolution Face Recognition via Bridge Distillation , 2020, IEEE Transactions on Image Processing.

[32]  Qijun Zhao,et al.  Towards resolution invariant face recognition in uncontrolled scenarios , 2016, 2016 International Conference on Biometrics (ICB).

[33]  Shiming Ge,et al.  Look One and More: Distilling Hybrid Order Relational Knowledge for Cross-Resolution Image Recognition , 2020, AAAI.

[34]  Shaogang Gong,et al.  Low-Resolution Face Recognition , 2018, ACCV.

[35]  Xiaoming Liu,et al.  FAN: Feature Adaptation Network for Surveillance Face Recognition and Normalization , 2019, ACCV.

[36]  Klemen Grm,et al.  Face Hallucination Using Cascaded Super-Resolution and Identity Priors , 2018, IEEE Transactions on Image Processing.