Deep Face Model Compression Using Entropy-Based Filter Selection

The state-of-the-art face recognition systems are built on deep convolutional neural networks (CNNs). However, these CNNs contain millions of parameters, leading to the deployment difficulties on mobile and embedded devices. One solution is to reduce the size of the trained CNNs by model compression. In this work, we propose an entropy-based prune metric to reduce the size of intermediate activations so as to accelerate and compress CNN models both in training and inference stages. First the importance of each filter in each layer is evaluated by our entropy-based method. Then some unimportant filters are removed according to a predefined compressing rate. Finally, we fine-tune the pruned model to improve its discrimination ability. Experiments conducted on LFW face dataset shows the effectiveness of our entropy-based method. We achieve \(1.92\times \) compression and \(1.88\times \) speed-up on VGG-16 model, \(2\times \) compression and \(1.74\times \) speed-up on WebFace model, both with only about 1% accuracy decrease evaluated on LFW.

[1]  Misha Denil,et al.  Predicting Parameters in Deep Learning , 2014 .

[2]  Li Fei-Fei,et al.  ImageNet: A large-scale hierarchical image database , 2009, CVPR.

[3]  Shengcai Liao,et al.  Learning Face Representation from Scratch , 2014, ArXiv.

[4]  Martín Abadi,et al.  TensorFlow: Large-Scale Machine Learning on Heterogeneous Distributed Systems , 2016, ArXiv.

[5]  Jian Cheng,et al.  Quantized Convolutional Neural Networks for Mobile Devices , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[6]  Geoffrey Zweig,et al.  From captions to visual concepts and back , 2014, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[7]  Tara N. Sainath,et al.  Structured Transforms for Small-Footprint Deep Learning , 2015, NIPS.

[8]  Geoffrey E. Hinton,et al.  ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.

[9]  Forrest N. Iandola,et al.  SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and <1MB model size , 2016, ArXiv.

[10]  Pascal Vincent,et al.  Representation Learning: A Review and New Perspectives , 2012, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[11]  Andrew Zisserman,et al.  Very Deep Convolutional Networks for Large-Scale Image Recognition , 2014, ICLR.

[12]  Jianxin Wu,et al.  An Entropy-based Pruning Method for CNN Compression , 2017, ArXiv.

[13]  Xiaoou Tang,et al.  Learning a Deep Convolutional Network for Image Super-Resolution , 2014, ECCV.

[14]  Jian Sun,et al.  Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[15]  Song Han,et al.  Learning both Weights and Connections for Efficient Neural Network , 2015, NIPS.

[16]  Geoffrey E. Hinton,et al.  Learning distributed representations of concepts. , 1989 .

[17]  Dumitru Erhan,et al.  Going deeper with convolutions , 2014, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[18]  Qiang Chen,et al.  Network In Network , 2013, ICLR.