Differentiable Soft Quantization: Bridging Full-Precision and Low-Bit Neural Networks
暂无分享,去创建一个
Xianglong Liu | Junjie Yan | Ruihao Gong | Fengwei Yu | Peng Hu | Tianxiang Li | Jiazhen Lin | Shenghu Jiang | Junjie Yan | Xianglong Liu | F. Yu | Tian-Hao Li | Ruihao Gong | Peng Hu | Shenghu Jiang | Jiazhen Lin
[1] Song Han,et al. Trained Ternary Quantization , 2016, ICLR.
[2] Wei Pan,et al. Towards Accurate Binary Convolutional Neural Network , 2017, NIPS.
[3] Zhijian Liu,et al. HAQ: Hardware-Aware Automated Quantization With Mixed Precision , 2018, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[4] Hanan Samet,et al. Training Quantized Nets: A Deeper Understanding , 2017, NIPS.
[5] Ran El-Yaniv,et al. Binarized Neural Networks , 2016, NIPS.
[6] Ran El-Yaniv,et al. Binarized Neural Networks , 2016, ArXiv.
[7] Song Han,et al. HAQ: Hardware-Aware Automated Quantization , 2018, ArXiv.
[8] Houqiang Li,et al. Adaptive Layerwise Quantization for Deep Neural Network Compression , 2018, 2018 IEEE International Conference on Multimedia and Expo (ICME).
[9] Daisuke Miyashita,et al. Convolutional Neural Networks using Logarithmic Data Representation , 2016, ArXiv.
[10] Yoshua Bengio,et al. Estimating or Propagating Gradients Through Stochastic Neurons for Conditional Computation , 2013, ArXiv.
[11] Fei-Fei Li,et al. ImageNet: A large-scale hierarchical image database , 2009, 2009 IEEE Conference on Computer Vision and Pattern Recognition.
[12] G. Hua,et al. LQ-Nets: Learned Quantization for Highly Accurate and Compact Deep Neural Networks , 2018, ECCV.
[13] Daniel Soudry,et al. Post training 4-bit quantization of convolutional networks for rapid-deployment , 2018, NeurIPS.
[14] Bo Chen,et al. Quantization and Training of Neural Networks for Efficient Integer-Arithmetic-Only Inference , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[15] Shuchang Zhou,et al. Balanced Quantization: An Effective and Efficient Approach to Quantized Neural Networks , 2017, Journal of Computer Science and Technology.
[16] Igor Carron,et al. XNOR-Net: ImageNet Classification Using Binary Convolutional Neural Networks , 2016 .
[17] Jack Xin,et al. Blended coarse gradient descent for full quantization of deep neural networks , 2018, Research in the Mathematical Sciences.
[18] Yang Liu,et al. Two-Step Quantization for Low-bit Neural Networks , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[19] Elad Hoffer,et al. ACIQ: Analytical Clipping for Integer Quantization of neural networks , 2018, ArXiv.
[20] Greg Mori,et al. CLIP-Q: Deep Network Compression Learning by In-parallel Pruning-Quantization , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[21] Eriko Nurvitadhi,et al. WRPN: Wide Reduced-Precision Networks , 2017, ICLR.
[22] Yuhui Xu,et al. Deep Neural Network Compression with Single and Multiple Level Quantization , 2018, AAAI.
[23] Swagath Venkataramani,et al. PACT: Parameterized Clipping Activation for Quantized Neural Networks , 2018, ArXiv.
[24] Yoshua Bengio,et al. BinaryConnect: Training Deep Neural Networks with binary weights during propagations , 2015, NIPS.
[25] Swagath Venkataramani,et al. Bridging the Accuracy Gap for 2-bit Quantized Neural Networks (QNN) , 2018, ArXiv.
[26] Avi Mendelson,et al. UNIQ , 2018, ACM Trans. Comput. Syst..
[27] Jian Sun,et al. Identity Mappings in Deep Residual Networks , 2016, ECCV.
[28] Mark Sandler,et al. MobileNetV2: Inverted Residuals and Linear Bottlenecks , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[29] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[30] Jae-Joon Han,et al. Learning to Quantize Deep Networks by Optimizing Quantization Intervals With Task Loss , 2018, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[31] Lin Xu,et al. Incremental Network Quantization: Towards Lossless CNNs with Low-Precision Weights , 2017, ICLR.
[32] Yuan Yu,et al. TensorFlow: A system for large-scale machine learning , 2016, OSDI.
[33] Xiaoli Liu,et al. Highly Efficient 8-bit Low Precision Inference of Convolutional Neural Networks with IntelCaffe , 2018, ReQuEST@ASPLOS.
[34] Dharmendra S. Modha,et al. Discovering Low-Precision Networks Close to Full-Precision Networks for Efficient Embedded Inference , 2018, ArXiv.
[35] Song Han,et al. Deep Compression: Compressing Deep Neural Network with Pruning, Trained Quantization and Huffman Coding , 2015, ICLR.
[36] Bin Liu,et al. Ternary Weight Networks , 2016, ICASSP 2023 - 2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
[37] Trevor Darrell,et al. Caffe: Convolutional Architecture for Fast Feature Embedding , 2014, ACM Multimedia.
[38] Luca Antiga,et al. Automatic differentiation in PyTorch , 2017 .
[39] Jian Sun,et al. Deep Learning with Low Precision by Half-Wave Gaussian Quantization , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[40] Hadi Esmaeilzadeh,et al. ReLeQ: A Reinforcement Learning Approach for Deep Quantization of Neural Networks , 2018, ArXiv.
[41] Raghuraman Krishnamoorthi,et al. Quantizing deep convolutional networks for efficient inference: A whitepaper , 2018, ArXiv.
[42] Hadi Esmaeilzadeh,et al. ReLeQ: An Automatic Reinforcement Learning Approach for Deep Quantization of Neural Networks , 2018 .
[43] Asit K. Mishra,et al. Apprentice: Using Knowledge Distillation Techniques To Improve Low-Precision Network Accuracy , 2017, ICLR.
[44] Shuchang Zhou,et al. DoReFa-Net: Training Low Bitwidth Convolutional Neural Networks with Low Bitwidth Gradients , 2016, ArXiv.