Few Sample Knowledge Distillation for Efficient Network Compression
暂无分享,去创建一个
Changshui Zhang | Jianguo Li | Zhuang Liu | Tianhong Li | Zhuang Liu | Jianguo Li | Changshui Zhang | Tianhong Li
[1] Geoffrey E. Hinton,et al. Distilling the Knowledge in a Neural Network , 2015, ArXiv.
[2] Hugo Larochelle,et al. Optimization as a Model for Few-Shot Learning , 2016, ICLR.
[3] Rich Caruana,et al. Do Deep Nets Really Need to be Deep? , 2013, NIPS.
[4] Pietro Perona,et al. One-shot learning of object categories , 2006, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[5] Kartikeya Bhardwaj,et al. Dream Distillation: A Data-Independent Model Compression Framework , 2019, ArXiv.
[6] Huchuan Lu,et al. Deep Mutual Learning , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[7] Yoshua Bengio,et al. FitNets: Hints for Thin Deep Nets , 2014, ICLR.
[8] Kilian Q. Weinberger,et al. Densely Connected Convolutional Networks , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[9] Junmo Kim,et al. A Gift from Knowledge Distillation: Fast Optimization, Network Minimization and Transfer Learning , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[10] Lukás Burget,et al. Recurrent neural network based language model , 2010, INTERSPEECH.
[11] Geoffrey E. Hinton,et al. Rectified Linear Units Improve Restricted Boltzmann Machines , 2010, ICML.
[12] François Fleuret,et al. Knowledge Transfer with Jacobian Matching , 2018, ICML.
[13] Sergey Levine,et al. Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks , 2017, ICML.
[14] Song Han,et al. Deep Compression: Compressing Deep Neural Network with Pruning, Trained Quantization and Huffman Coding , 2015, ICLR.
[15] Weiyao Lin,et al. Network Decoupling: From Regular to Depthwise Separable Convolutions , 2018, BMVC.
[16] Qi Tian,et al. Data-Free Learning of Student Networks , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).
[17] Rich Caruana,et al. Model compression , 2006, KDD '06.
[18] Oriol Vinyals,et al. Matching Networks for One Shot Learning , 2016, NIPS.
[19] Eunhyeok Park,et al. Compression of Deep Convolutional Neural Networks for Fast and Low Power Mobile Applications , 2015, ICLR.
[20] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[21] Zhi-Quan Luo,et al. Iteration complexity analysis of block coordinate descent methods , 2013, Mathematical Programming.
[22] Andrew Zisserman,et al. Very Deep Convolutional Networks for Large-Scale Image Recognition , 2014, ICLR.
[23] Joan Bruna,et al. Exploiting Linear Structure Within Convolutional Networks for Efficient Evaluation , 2014, NIPS.
[24] Andrew Zisserman,et al. Speeding up Convolutional Neural Networks with Low Rank Expansions , 2014, BMVC.
[25] Tara N. Sainath,et al. Deep Neural Networks for Acoustic Modeling in Speech Recognition: The Shared Views of Four Research Groups , 2012, IEEE Signal Processing Magazine.
[26] Jianxin Wu,et al. ThiNet: A Filter Level Pruning Method for Deep Neural Network Compression , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).
[27] Wotao Yin,et al. A Globally Convergent Algorithm for Nonconvex Optimization Based on Block Coordinate Update , 2014, J. Sci. Comput..
[28] Jian Sun,et al. Accelerating Very Deep Convolutional Networks for Classification and Detection , 2015, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[29] Joshua B. Tenenbaum,et al. One shot learning of simple visual concepts , 2011, CogSci.
[30] Naiyan Wang,et al. Like What You Like: Knowledge Distill via Neuron Selectivity Transfer , 2017, ArXiv.
[31] Zhijian Liu,et al. GAN Compression: Efficient Architectures for Interactive Conditional GANs , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[32] Shimon Ullman,et al. Cross-generalization: learning novel classes from a single example by feature replacement , 2005, 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'05).
[33] Thad Starner,et al. Data-Free Knowledge Distillation for Deep Neural Networks , 2017, ArXiv.
[34] Zhiqiang Shen,et al. Learning Efficient Convolutional Networks through Network Slimming , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).
[35] Koh Takeuchi,et al. Few-shot learning of neural networks from scratch by pseudo example optimization , 2018, BMVC.
[36] Geoffrey E. Hinton,et al. ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.
[37] Tianqi Chen,et al. Net2Net: Accelerating Learning via Knowledge Transfer , 2015, ICLR.
[38] Song Han,et al. Learning both Weights and Connections for Efficient Neural Network , 2015, NIPS.
[39] Hanan Samet,et al. Pruning Filters for Efficient ConvNets , 2016, ICLR.