Training Shallow and Thin Networks for Acceleration via Knowledge Distillation with Conditional Adversarial Networks
暂无分享,去创建一个
[1] Andrew Zisserman,et al. Very Deep Convolutional Networks for Large-Scale Image Recognition , 2014, ICLR.
[2] R. Srikant,et al. Why Deep Neural Networks for Function Approximation? , 2016, ICLR.
[3] Nikos Komodakis,et al. Paying More Attention to Attention: Improving the Performance of Convolutional Neural Networks via Attention Transfer , 2016, ICLR.
[4] Frank Hutter,et al. A Downsampled Variant of ImageNet as an Alternative to the CIFAR datasets , 2017, ArXiv.
[5] Guorui Zhou,et al. Rocket Launching: A Universal and Efficient Framework for Training Well-performing Light Net , 2017, AAAI.
[6] Yoshua Bengio,et al. A Closer Look at Memorization in Deep Networks , 2017, ICML.
[7] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[8] Hanan Samet,et al. Training Quantized Nets: A Deeper Understanding , 2017, NIPS.
[9] Hanan Samet,et al. Pruning Filters for Efficient ConvNets , 2016, ICLR.
[10] Vishnu Naresh Boddeti,et al. In Teacher We Trust: Learning Compressed Models for Pedestrian Detection , 2016, ArXiv.
[11] Geoffrey E. Hinton,et al. Distilling the Knowledge in a Neural Network , 2015, ArXiv.
[12] G. Lewicki,et al. Approximation by Superpositions of a Sigmoidal Function , 2003 .
[13] Naiyan Wang,et al. Like What You Like: Knowledge Distill via Neuron Selectivity Transfer , 2017, ArXiv.
[14] Sergey Ioffe,et al. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift , 2015, ICML.
[15] Nitish Srivastava,et al. Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..
[16] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[17] Léon Bottou,et al. Wasserstein GAN , 2017, ArXiv.
[18] Rich Caruana,et al. Do Deep Nets Really Need to be Deep? , 2013, NIPS.
[19] Matus Telgarsky,et al. Benefits of Depth in Neural Networks , 2016, COLT.
[20] Igor Carron,et al. XNOR-Net: ImageNet Classification Using Binary Convolutional Neural Networks , 2016 .
[21] Nikos Komodakis,et al. Wide Residual Networks , 2016, BMVC.
[22] Matthew Richardson,et al. Do Deep Convolutional Nets Really Need to be Deep and Convolutional? , 2016, ICLR.
[23] Yoshua Bengio,et al. FitNets: Hints for Thin Deep Nets , 2014, ICLR.
[24] Yoshua Bengio,et al. Generative Adversarial Nets , 2014, NIPS.
[25] Raymond Y. K. Lau,et al. Least Squares Generative Adversarial Networks , 2016, 2017 IEEE International Conference on Computer Vision (ICCV).
[26] Yee Whye Teh,et al. Distral: Robust multitask reinforcement learning , 2017, NIPS.
[27] Geoffrey E. Hinton,et al. ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.
[28] Rich Caruana,et al. Model compression , 2006, KDD '06.
[29] Michael S. Bernstein,et al. ImageNet Large Scale Visual Recognition Challenge , 2014, International Journal of Computer Vision.
[30] Ohad Shamir,et al. The Power of Depth for Feedforward Neural Networks , 2015, COLT.
[31] Zheng Xu,et al. Stabilizing Adversarial Nets With Prediction Methods , 2017, ICLR.
[32] Germán Ros,et al. Training Constrained Deconvolutional Networks for Road Scene Semantic Segmentation , 2016, ArXiv.
[33] Bernt Schiele,et al. Generative Adversarial Text to Image Synthesis , 2016, ICML.
[34] Zhaoxiang Zhang,et al. DarkRank: Accelerating Deep Metric Learning via Cross Sample Similarities Transfer , 2017, AAAI.
[35] Junmo Kim,et al. A Gift from Knowledge Distillation: Fast Optimization, Network Minimization and Transfer Learning , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[36] Pieter Abbeel,et al. InfoGAN: Interpretable Representation Learning by Information Maximizing Generative Adversarial Nets , 2016, NIPS.
[37] Simon Osindero,et al. Conditional Generative Adversarial Nets , 2014, ArXiv.
[38] Zhuowen Tu,et al. Aggregated Residual Transformations for Deep Neural Networks , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[39] Kurt Hornik,et al. Multilayer feedforward networks are universal approximators , 1989, Neural Networks.
[40] Hyo-Eun Kim,et al. Transferring Knowledge to Smaller Network with Class-Distance Loss , 2017, ICLR.
[41] Samy Bengio,et al. Understanding deep learning requires rethinking generalization , 2016, ICLR.
[42] Jonathon Shlens,et al. Conditional Image Synthesis with Auxiliary Classifier GANs , 2016, ICML.
[43] Vineeth N. Balasubramanian,et al. Deep Model Compression: Distilling Knowledge from Noisy Teachers , 2016, ArXiv.
[44] P. Cochat,et al. Et al , 2008, Archives de pediatrie : organe officiel de la Societe francaise de pediatrie.
[45] Alexei A. Efros,et al. Image-to-Image Translation with Conditional Adversarial Networks , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[46] Xiaogang Wang,et al. Face Model Compression by Distilling Knowledge from Neurons , 2016, AAAI.
[47] Bo Chen,et al. MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications , 2017, ArXiv.
[48] Wenjun Zeng,et al. Deeply-Fused Nets , 2016, ArXiv.
[49] Dacheng Tao,et al. Learning from Multiple Teacher Networks , 2017, KDD.