Variational Information Distillation for Knowledge Transfer
暂无分享,去创建一个
Neil D. Lawrence | Zhenwen Dai | Sungsoo Ahn | Andreas C. Damianou | Shell Xu Hu | Neil D. Lawrence | A. Damianou | Zhenwen Dai | Sungsoo Ahn | S. Hu
[1] Zachary Chase Lipton,et al. Born Again Neural Networks , 2018, ICML.
[2] Georg Langs,et al. Unsupervised Pre-training Across Image Domains Improves Lung Tissue Classification , 2014, MCV.
[3] Ming Dong,et al. Coupled End-to-End Transfer Learning with Generalized Fisher Information , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[4] S. Srihari. Mixture Density Networks , 1994 .
[5] Tapani Raiko,et al. Semi-supervised Learning with Ladder Networks , 2015, NIPS.
[6] Naiyan Wang,et al. Like What You Like: Knowledge Distill via Neuron Selectivity Transfer , 2017, ArXiv.
[7] Pascal Vincent,et al. Stacked Denoising Autoencoders: Learning Useful Representations in a Deep Network with a Local Denoising Criterion , 2010, J. Mach. Learn. Res..
[8] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[9] Andrew Zisserman,et al. Very Deep Convolutional Networks for Large-Scale Image Recognition , 2014, ICLR.
[10] Matthew Richardson,et al. Do Deep Convolutional Nets Really Need to be Deep and Convolutional? , 2016, ICLR.
[11] Aaron C. Courville,et al. MINE: Mutual Information Neural Estimation , 2018, ArXiv.
[12] Yoshua Bengio,et al. FitNets: Hints for Thin Deep Nets , 2014, ICLR.
[13] Christian Szegedy,et al. DeepPose: Human Pose Estimation via Deep Neural Networks , 2013, 2014 IEEE Conference on Computer Vision and Pattern Recognition.
[14] David Barber,et al. The IM algorithm: a variational approach to Information Maximization , 2003, NIPS 2003.
[15] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[16] Fabio Galasso,et al. Adversarial Network Compression , 2018, ECCV Workshops.
[17] Max Welling,et al. Improved Variational Inference with Inverse Autoregressive Flow , 2016, NIPS 2016.
[18] Qiang Yang,et al. A Survey on Transfer Learning , 2010, IEEE Transactions on Knowledge and Data Engineering.
[19] Rob Fergus,et al. Depth Map Prediction from a Single Image using a Multi-Scale Deep Network , 2014, NIPS.
[20] Pietro Perona,et al. Caltech-UCSD Birds 200 , 2010 .
[21] Thomas Brox,et al. FlowNet: Learning Optical Flow with Convolutional Networks , 2015, 2015 IEEE International Conference on Computer Vision (ICCV).
[22] Ralph Linsker,et al. An Application of the Principle of Maximum Information Preservation to Linear Systems , 1988, NIPS.
[23] Roland Memisevic,et al. How far can we go without convolution: Improving fully-connected networks , 2015, ArXiv.
[24] Michael S. Bernstein,et al. ImageNet Large Scale Visual Recognition Challenge , 2014, International Journal of Computer Vision.
[25] Ross B. Girshick,et al. Fast R-CNN , 2015, 1504.08083.
[26] Nikos Komodakis,et al. Wide Residual Networks , 2016, BMVC.
[27] Junmo Kim,et al. A Gift from Knowledge Distillation: Fast Optimization, Network Minimization and Transfer Learning , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[28] Derek Hoiem,et al. Learning without Forgetting , 2016, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[29] Nikos Komodakis,et al. Paying More Attention to Attention: Improving the Performance of Convolutional Neural Networks via Attention Transfer , 2016, ICLR.
[30] Antonio Torralba,et al. Recognizing indoor scenes , 2009, 2009 IEEE Conference on Computer Vision and Pattern Recognition.
[31] Geoffrey E. Hinton,et al. Distilling the Knowledge in a Neural Network , 2015, ArXiv.
[32] Rich Caruana,et al. Do Deep Nets Really Need to be Deep? , 2013, NIPS.