Knowledge Extraction with No Observable Data
暂无分享,去创建一个
U Kang | Jaemin Yoo | Minyong Cho | Taebum Kim | U. Kang | Jaemin Yoo | Minyong Cho | Taebum Kim
[1] Jangho Kim,et al. Paraphrasing Complex Network: Network Compression via Factor Transfer , 2018, NeurIPS.
[2] Tao Zhang,et al. A Survey of Model Compression and Acceleration for Deep Neural Networks , 2017, ArXiv.
[3] Anima Anandkumar,et al. Tensor Regression Networks , 2017, J. Mach. Learn. Res..
[4] Sepp Hochreiter,et al. Fast and Accurate Deep Network Learning by Exponential Linear Units (ELUs) , 2015, ICLR.
[5] Geoffrey E. Hinton,et al. Distilling the Knowledge in a Neural Network , 2015, ArXiv.
[6] Senén Barro,et al. Do we need hundreds of classifiers to solve real world classification problems? , 2014, J. Mach. Learn. Res..
[7] Shinichi Nakajima,et al. Global analytic solution of fully-observed variational Bayesian matrix factorization , 2013, J. Mach. Learn. Res..
[8] Alexander M. Rush,et al. Sequence-Level Knowledge Distillation , 2016, EMNLP.
[9] Abraham J. Wyner,et al. Modern Neural Networks Generalize on Small Data Sets , 2018, NeurIPS.
[10] Changshui Zhang,et al. Knowledge Distillation from Few Samples , 2018, ArXiv.
[11] Eunhyeok Park,et al. Compression of Deep Convolutional Neural Networks for Fast and Low Power Mobile Applications , 2015, ICLR.
[12] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[13] Yoshua Bengio,et al. Generative Adversarial Nets , 2014, NIPS.
[14] Koh Takeuchi,et al. Few-shot learning of neural networks from scratch by pseudo example optimization , 2018, BMVC.
[15] Martín Abadi,et al. Semi-supervised Knowledge Transfer for Deep Learning from Private Training Data , 2016, ICLR.
[16] Nitish Srivastava,et al. Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..
[17] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[18] Thad Starner,et al. Data-Free Knowledge Distillation for Deep Neural Networks , 2017, ArXiv.
[19] Harris Drucker,et al. Learning algorithms for classification: A comparison on handwritten digit recognition , 1995 .
[20] Yoshua Bengio,et al. Gradient-based learning applied to document recognition , 1998, Proc. IEEE.
[21] Jonathon Shlens,et al. Conditional Image Synthesis with Auxiliary Classifier GANs , 2016, ICML.
[22] Maja Pantic,et al. TensorLy: Tensor Learning in Python , 2016, J. Mach. Learn. Res..
[23] Tianqi Chen,et al. Net2Net: Accelerating Learning via Knowledge Transfer , 2015, ICLR.
[24] Soumith Chintala,et al. Unsupervised Representation Learning with Deep Convolutional Generative Adversarial Networks , 2015, ICLR.
[25] Maksym Kholiavchenko,et al. Iterative Low-Rank Approximation for CNN Compression , 2018, ArXiv.
[26] Sergey Ioffe,et al. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift , 2015, ICML.
[27] Tamara G. Kolda,et al. Tensor Decompositions and Applications , 2009, SIAM Rev..
[28] Roland Vollgraf,et al. Fashion-MNIST: a Novel Image Dataset for Benchmarking Machine Learning Algorithms , 2017, ArXiv.
[29] Hanqing Lu,et al. Recent advances in efficient computation of deep convolutional neural networks , 2018, Frontiers of Information Technology & Electronic Engineering.
[30] Dan Alistarh,et al. Model compression via distillation and quantization , 2018, ICLR.
[31] Stephan Günnemann,et al. Introduction to Tensor Decompositions and their Applications in Machine Learning , 2017, ArXiv.
[32] Masashi Sugiyama,et al. Bayesian Dark Knowledge , 2015 .
[33] Andrew Y. Ng,et al. Reading Digits in Natural Images with Unsupervised Feature Learning , 2011 .