暂无分享,去创建一个
[1] Hongbin Zha,et al. Alternating Multi-bit Quantization for Recurrent Neural Networks , 2018, ICLR.
[2] Erich Elsen,et al. Exploring Sparsity in Recurrent Neural Networks , 2017, ICLR.
[3] Yann LeCun,et al. The Loss Surfaces of Multilayer Networks , 2014, AISTATS.
[4] Geoffrey E. Hinton,et al. Distilling the Knowledge in a Neural Network , 2015, ArXiv.
[5] Song Han,et al. Learning both Weights and Connections for Efficient Neural Network , 2015, NIPS.
[6] Dan Alistarh,et al. Model compression via distillation and quantization , 2018, ICLR.
[7] Dongsoo Lee,et al. Viterbi-based Pruning for Sparse Matrix with Fixed and High Index Compression Ratio , 2018, ICLR.
[8] Yurong Chen,et al. Dynamic Network Surgery for Efficient DNNs , 2016, NIPS.
[9] Guigang Zhang,et al. Deep Learning , 2016, Int. J. Semantic Comput..
[10] Beatrice Santorini,et al. Building a Large Annotated Corpus of English: The Penn Treebank , 1993, CL.
[11] Shuchang Zhou,et al. DoReFa-Net: Training Low Bitwidth Convolutional Neural Networks with Low Bitwidth Gradients , 2016, ArXiv.
[12] Wojciech Zaremba,et al. Recurrent Neural Network Regularization , 2014, ArXiv.
[13] Ran El-Yaniv,et al. Quantized Neural Networks: Training Neural Networks with Low Precision Weights and Activations , 2016, J. Mach. Learn. Res..
[14] Yurong Chen,et al. Network Sketching: Exploiting Binary Structure in Deep CNNs , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[15] Dmitry P. Vetrov,et al. Variational Dropout Sparsifies Deep Neural Networks , 2017, ICML.
[16] Song Han,et al. Trained Ternary Quantization , 2016, ICLR.
[17] Yann LeCun,et al. Regularization of Neural Networks using DropConnect , 2013, ICML.
[18] Jürgen Schmidhuber,et al. Simplifying Neural Nets by Discovering Flat Minima , 1994, NIPS.
[19] Ali Farhadi,et al. XNOR-Net: ImageNet Classification Using Binary Convolutional Neural Networks , 2016, ECCV.
[20] Yoshua Bengio,et al. BinaryConnect: Training Deep Neural Networks with binary weights during propagations , 2015, NIPS.
[21] Shuchang Zhou,et al. Effective Quantization Methods for Recurrent Neural Networks , 2016, ArXiv.
[22] Geoffrey E. Hinton,et al. Keeping the neural networks simple by minimizing the description length of the weights , 1993, COLT '93.
[23] Yann LeCun,et al. Optimal Brain Damage , 1989, NIPS.
[24] Ian McGraw,et al. On the compression of recurrent neural networks with an application to LVCSR acoustic modeling for embedded speech recognition , 2016, 2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
[25] Suyog Gupta,et al. To prune, or not to prune: exploring the efficacy of pruning for model compression , 2017, ICLR.
[26] Ebru Arisoy,et al. Low-rank matrix factorization for Deep Neural Network training with high-dimensional output targets , 2013, 2013 IEEE International Conference on Acoustics, Speech and Signal Processing.
[27] Song Han,et al. Deep Compression: Compressing Deep Neural Network with Pruning, Trained Quantization and Huffman Coding , 2015, ICLR.
[28] Shuang Wu,et al. Training and Inference with Integers in Deep Neural Networks , 2018, ICLR.
[29] Geoffrey E. Hinton,et al. Deep Learning , 2015, Nature.
[30] Jürgen Schmidhuber,et al. Long Short-Term Memory , 1997, Neural Computation.