Flexpoint: An Adaptive Numerical Format for Efficient Training of Deep Neural Networks
暂无分享,去创建一个
Xin Wang | Stewart Hall | Arjun K. Bansal | William Constable | Marcel Nassar | Urs Köster | Luke Hornof | Amir Khosrowshahi | Oguz H. Elibol | Tristan Webb | Oguz Elibol | Carey Kloss | Ruby J. Pai | Naveen Rao | Urs Köster | T. Webb | Xin Wang | M. Nassar | W. Constable | Stewart Hall | Luke Hornof | A. Khosrowshahi | Carey Kloss | Ruby J. Pai | N. Rao | William Constable
[1] D. Williamson. Dynamically scaled fixed point arithmetic , 1991, [1991] IEEE Pacific Rim Conference on Communications, Computers and Signal Processing Conference Proceedings.
[2] Vincent Vanhoucke,et al. Improving the speed of neural networks on CPUs , 2011 .
[3] Geoffrey E. Hinton,et al. ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.
[4] Yoshua Bengio,et al. Training deep neural networks with low precision multiplications , 2014 .
[5] Dong Yu,et al. 1-bit stochastic gradient descent and its application to data-parallel distributed training of speech DNNs , 2014, INTERSPEECH.
[6] Kyuyeon Hwang,et al. Fixed-point feedforward deep neural network design using weights +1, 0, and −1 , 2014, 2014 IEEE Workshop on Signal Processing Systems (SiPS).
[7] Yinda Zhang,et al. LSUN: Construction of a Large-scale Image Dataset using Deep Learning with Humans in the Loop , 2015, ArXiv.
[8] Yoshua Bengio,et al. BinaryConnect: Training Deep Neural Networks with binary weights during propagations , 2015, NIPS.
[9] Pritish Narayanan,et al. Deep Learning with Limited Numerical Precision , 2015, ICML.
[10] Yoshua Bengio,et al. BinaryNet: Training Deep Neural Networks with Weights and Activations Constrained to +1 or -1 , 2016, ArXiv.
[11] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[12] Sachin S. Talathi,et al. Fixed Point Quantization of Deep Convolutional Networks , 2015, ICML.
[13] Yoshua Bengio,et al. Neural Networks with Few Multiplications , 2015, ICLR.
[14] Ran El-Yaniv,et al. Binarized Neural Networks , 2016, NIPS.
[15] Jian Sun,et al. Identity Mappings in Deep Residual Networks , 2016, ECCV.
[16] Shuchang Zhou,et al. DoReFa-Net: Training Low Bitwidth Convolutional Neural Networks with Low Bitwidth Gradients , 2016, ArXiv.
[17] Paris Smaragdis,et al. Bitwise Neural Networks , 2016, ArXiv.
[18] Martín Abadi,et al. TensorFlow: Large-Scale Machine Learning on Heterogeneous Distributed Systems , 2016, ArXiv.
[19] Ali Farhadi,et al. XNOR-Net: ImageNet Classification Using Binary Convolutional Neural Networks , 2016, ECCV.
[20] Daisuke Miyashita,et al. Convolutional Neural Networks using Logarithmic Data Representation , 2016, ArXiv.
[21] Sepp Hochreiter,et al. GANs Trained by a Two Time-Scale Update Rule Converge to a Local Nash Equilibrium , 2017, NIPS.
[22] Abhisek Kundu,et al. Mixed Low-precision Deep Learning Inference using Dynamic Fixed Point , 2017, ArXiv.
[23] David A. Patterson,et al. In-datacenter performance analysis of a tensor processing unit , 2017, 2017 ACM/IEEE 44th Annual International Symposium on Computer Architecture (ISCA).
[24] Eriko Nurvitadhi,et al. Accelerating Deep Convolutional Networks using low-precision and sparsity , 2016, 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
[25] Ran El-Yaniv,et al. Quantized Neural Networks: Training Neural Networks with Low Precision Weights and Activations , 2016, J. Mach. Learn. Res..