Accurate and Efficient 2-bit Quantized Neural Networks
暂无分享,去创建一个
Swagath Venkataramani | Vijayalakshmi Srinivasan | Zhuo Wang | Kailash Gopalakrishnan | Jungwook Choi | Pierce Chuang | K. Gopalakrishnan | Jungwook Choi | Swagath Venkataramani | V. Srinivasan | P. Chuang | Zhuo Wang
[1] Nikos Komodakis,et al. Wide Residual Networks , 2016, BMVC.
[2] Mark D. McDonnell,et al. Training wide residual networks for deployment using a single bit for each weight , 2018, ICLR.
[3] Kyuyeon Hwang,et al. Fixed-point feedforward deep neural network design using weights +1, 0, and −1 , 2014, 2014 IEEE Workshop on Signal Processing Systems (SiPS).
[4] Ran El-Yaniv,et al. Quantized Neural Networks: Training Neural Networks with Low Precision Weights and Activations , 2016, J. Mach. Learn. Res..
[5] Avi Mendelson,et al. UNIQ: Uniform Noise Injection for the Quantization of Neural Networks , 2018, ArXiv.
[6] Swagath Venkataramani,et al. POSTER: Design Space Exploration for Performance Optimization of Deep Neural Networks on Shared Memory Accelerators , 2017, 2017 26th International Conference on Parallel Architectures and Compilation Techniques (PACT).
[7] Asit K. Mishra,et al. Apprentice: Using Knowledge Distillation Techniques To Improve Low-Precision Network Accuracy , 2017, ICLR.
[8] G. Hua,et al. LQ-Nets: Learned Quantization for Highly Accurate and Compact Deep Neural Networks , 2018, ECCV.
[9] Ying Zhang,et al. Towards End-to-End Speech Recognition with Deep Convolutional Neural Networks , 2016, INTERSPEECH.
[10] Geoffrey E. Hinton,et al. Rectified Linear Units Improve Restricted Boltzmann Machines , 2010, ICML.
[11] Dan Alistarh,et al. Model compression via distillation and quantization , 2018, ICLR.
[12] A. Krizhevsky. Convolutional Deep Belief Networks on CIFAR-10 , 2010 .
[13] Geoffrey E. Hinton,et al. ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.
[14] Jian Sun,et al. Deep Learning with Low Precision by Half-Wave Gaussian Quantization , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[15] Jae-Joon Han,et al. Joint Training of Low-Precision Neural Network with Quantization Interval Parameters , 2018, ArXiv.
[16] Jian Sun,et al. Identity Mappings in Deep Residual Networks , 2016, ECCV.
[17] Jungwon Lee,et al. Learning Low Precision Deep Neural Networks through Regularization , 2018, ArXiv.
[18] Yoshua Bengio,et al. Estimating or Propagating Gradients Through Stochastic Neurons for Conditional Computation , 2013, ArXiv.
[19] Paris Smaragdis,et al. Bitwise Neural Networks , 2016, ArXiv.
[20] James T. Kwok,et al. Loss-aware Weight Quantization of Deep Networks , 2018, ICLR.
[21] Joel Silberman,et al. A Scalable Multi- TeraOPS Deep Learning Processor Core for AI Trainina and Inference , 2018, 2018 IEEE Symposium on VLSI Circuits.
[22] Eunhyeok Park,et al. Weighted-Entropy-Based Quantization for Deep Neural Networks , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[23] Shuchang Zhou,et al. DoReFa-Net: Training Low Bitwidth Convolutional Neural Networks with Low Bitwidth Gradients , 2016, ArXiv.
[24] Michael S. Bernstein,et al. ImageNet Large Scale Visual Recognition Challenge , 2014, International Journal of Computer Vision.
[25] Shuchang Zhou,et al. Balanced Quantization: An Effective and Efficient Approach to Quantized Neural Networks , 2017, Journal of Computer Science and Technology.
[26] Yoshua Bengio,et al. BinaryConnect: Training Deep Neural Networks with binary weights during propagations , 2015, NIPS.
[27] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).