暂无分享,去创建一个
[1] Nikos Komodakis,et al. Wide Residual Networks , 2016, BMVC.
[2] Prem Raj Adhikari,et al. Multiresolution Mixture Modeling using Merging of Mixture Components , 2012, ACML.
[3] Xiaogang Wang,et al. Convolutional neural networks with low-rank regularization , 2015, ICLR.
[4] Yoshua Bengio,et al. Gradient-based learning applied to document recognition , 1998, Proc. IEEE.
[5] C. S. Wallace,et al. Classification by Minimum-Message-Length Inference , 1991, ICCI.
[6] Misha Denil,et al. Predicting Parameters in Deep Learning , 2014 .
[7] H. Robbins. A Stochastic Approximation Method , 1951 .
[8] Song Han,et al. Deep Compression: Compressing Deep Neural Network with Pruning, Trained Quantization and Huffman Coding , 2015, ICLR.
[9] Omer Levy,et al. Published as a conference paper at ICLR 2018 S IMULATING A CTION D YNAMICS WITH N EURAL P ROCESS N ETWORKS , 2018 .
[10] Hassan Foroosh,et al. Sparse Convolutional Neural Networks , 2015, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[11] Jasper Snoek,et al. Practical Bayesian Optimization of Machine Learning Algorithms , 2012, NIPS.
[12] Sachin S. Talathi,et al. Fixed Point Quantization of Deep Convolutional Networks , 2015, ICML.
[13] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[14] Dharmendra S. Modha,et al. Deep neural networks are robust to weight binarization and other non-linear distortions , 2016, ArXiv.
[15] Yoshua Bengio,et al. BinaryConnect: Training Deep Neural Networks with binary weights during propagations , 2015, NIPS.
[16] Alex Graves,et al. Practical Variational Inference for Neural Networks , 2011, NIPS.
[17] Antti Honkela,et al. Variational learning and bits-back coding: an information-theoretic view to Bayesian learning , 2004, IEEE Transactions on Neural Networks.
[18] Ian H. Witten,et al. Arithmetic coding for data compression , 1987, CACM.
[19] Forrest N. Iandola,et al. SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and <1MB model size , 2016, ArXiv.
[20] J. Rissanen,et al. Modeling By Shortest Data Description* , 1978, Autom..
[21] Jian Cheng,et al. Quantized Convolutional Neural Networks for Mobile Devices , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[22] Roberto Cipolla,et al. Training CNNs with Low-Rank Filters for Efficient Image Classification , 2015, ICLR.
[23] Eriko Nurvitadhi,et al. Accelerating Deep Convolutional Networks using low-precision and sparsity , 2016, 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
[24] Ran El-Yaniv,et al. Binarized Neural Networks , 2016, NIPS.
[25] Yoshua Bengio,et al. Training deep neural networks with low precision multiplications , 2014 .
[26] Yiran Chen,et al. Learning Structured Sparsity in Deep Neural Networks , 2016, NIPS.
[27] Geoffrey E. Hinton,et al. Distilling the Knowledge in a Neural Network , 2015, ArXiv.
[28] Yoshua Bengio,et al. BinaryNet: Training Deep Neural Networks with Weights and Activations Constrained to +1 or -1 , 2016, ArXiv.
[29] Joan Bruna,et al. Exploiting Linear Structure Within Convolutional Networks for Efficient Evaluation , 2014, NIPS.
[30] J. Rissanen. Stochastic Complexity and Modeling , 1986 .
[31] Ming Yang,et al. Compressing Deep Convolutional Networks using Vector Quantization , 2014, ArXiv.
[32] Pritish Narayanan,et al. Deep Learning with Limited Numerical Precision , 2015, ICML.
[33] Geoffrey E. Hinton,et al. Keeping the neural networks simple by minimizing the description length of the weights , 1993, COLT '93.
[34] Yixin Chen,et al. Compressing Convolutional Neural Networks , 2015, ArXiv.
[35] Jian Sun,et al. Identity Mappings in Deep Residual Networks , 2016, ECCV.
[36] Andrew Zisserman,et al. Speeding up Convolutional Neural Networks with Low Rank Expansions , 2014, BMVC.
[37] Geoffrey E. Hinton,et al. Simplifying Neural Networks by Soft Weight-Sharing , 1992, Neural Computation.
[38] Song Han,et al. Learning both Weights and Connections for Efficient Neural Network , 2015, NIPS.
[39] Xinyun Chen. Under Review as a Conference Paper at Iclr 2017 Delving into Transferable Adversarial Ex- Amples and Black-box Attacks , 2016 .
[40] Yann LeCun,et al. Optimal Brain Damage , 1989, NIPS.
[41] B. Ripley,et al. Pattern Recognition , 1968, Nature.
[42] Dmitry P. Vetrov,et al. Variational Dropout Sparsifies Deep Neural Networks , 2017, ICML.
[43] Yurong Chen,et al. Dynamic Network Surgery for Efficient DNNs , 2016, NIPS.
[44] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.