Extremely Low Bit Neural Network: Squeeze the Last Bit Out with ADMM
暂无分享,去创建一个
Shenghuo Zhu | Hao Li | Rong Jin | Cong Leng | Rong Jin | Shenghuo Zhu | Hao Li | Cong Leng
[1] Nate Derbinsky,et al. An Improved Three-Weight Message-Passing Algorithm , 2013, ArXiv.
[2] Lin Xu,et al. Incremental Network Quantization: Towards Lossless CNNs with Low-Precision Weights , 2017, ICLR.
[3] Wotao Yin,et al. Global Convergence of ADMM in Nonconvex Nonsmooth Optimization , 2015, Journal of Scientific Computing.
[4] Ali Farhadi,et al. XNOR-Net: ImageNet Classification Using Binary Convolutional Neural Networks , 2016, ECCV.
[5] Song Han,et al. Trained Ternary Quantization , 2016, ICLR.
[6] Ali Farhadi,et al. You Only Look Once: Unified, Real-Time Object Detection , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[7] Ming Yang,et al. Compressing Deep Convolutional Networks using Vector Quantization , 2014, ArXiv.
[8] Ivan V. Oseledets,et al. Speeding-up Convolutional Neural Networks Using Fine-tuned CP-Decomposition , 2014, ICLR.
[9] Stephen P. Boyd,et al. Distributed Optimization and Statistical Learning via the Alternating Direction Method of Multipliers , 2011, Found. Trends Mach. Learn..
[10] Wei Liu,et al. SSD: Single Shot MultiBox Detector , 2015, ECCV.
[11] Yoshua Bengio,et al. Neural Networks with Few Multiplications , 2015, ICLR.
[12] Jian Cheng,et al. Quantized Convolutional Neural Networks for Mobile Devices , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[13] Shuchang Zhou,et al. DoReFa-Net: Training Low Bitwidth Convolutional Neural Networks with Low Bitwidth Gradients , 2016, ArXiv.
[14] Forrest N. Iandola,et al. SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and <1MB model size , 2016, ArXiv.
[15] G. M. Korpelevich. The extragradient method for finding saddle points and other problems , 1976 .
[16] Sergey Ioffe,et al. Inception-v4, Inception-ResNet and the Impact of Residual Connections on Learning , 2016, AAAI.
[17] Arkadi Nemirovski,et al. Prox-Method with Rate of Convergence O(1/t) for Variational Inequalities with Lipschitz Continuous Monotone Operators and Smooth Convex-Concave Saddle Point Problems , 2004, SIAM J. Optim..
[18] Yoshua Bengio,et al. BinaryConnect: Training Deep Neural Networks with binary weights during propagations , 2015, NIPS.
[19] Stephen P. Boyd,et al. A simple effective heuristic for embedded mixed-integer quadratic programming , 2015, 2016 American Control Conference (ACC).
[20] Alborz Alavian,et al. Improving ADMM-based optimization of Mixed Integer objectives , 2017, 2017 51st Annual Conference on Information Sciences and Systems (CISS).
[21] Bin Liu,et al. Ternary Weight Networks , 2016, ICASSP 2023 - 2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
[22] H. John Caulfield,et al. Weight discretization paradigm for optical neural networks , 1990, Other Conferences.
[23] Geoffrey E. Hinton,et al. ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.
[24] Trevor Darrell,et al. Caffe: Convolutional Architecture for Fast Feature Embedding , 2014, ACM Multimedia.
[25] M. Hestenes. Multiplier and gradient methods , 1969 .
[26] Francesco Piazza,et al. Fast neural networks without multipliers , 1993, IEEE Trans. Neural Networks.
[27] Bruce W. Suter,et al. Extragradient Method in Optimization: Convergence and Complexity , 2016, J. Optim. Theory Appl..
[28] Jian Sun,et al. Accelerating Very Deep Convolutional Networks for Classification and Detection , 2015, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[29] Eriko Nurvitadhi,et al. Accelerating Deep Convolutional Networks using low-precision and sparsity , 2016, 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
[30] Jian Sun,et al. Deep Learning with Low Precision by Half-Wave Gaussian Quantization , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[31] Sachin S. Talathi,et al. Fixed Point Quantization of Deep Convolutional Networks , 2015, ICML.
[32] Upal Mahbub,et al. New methods for handling binary constraints , 2016, 2016 54th Annual Allerton Conference on Communication, Control, and Computing (Allerton).
[33] Tim Dettmers,et al. 8-Bit Approximations for Parallelism in Deep Learning , 2015, ICLR.
[34] Yoshua Bengio,et al. Training deep neural networks with low precision multiplications , 2014 .
[35] Song Han,et al. Deep Compression: Compressing Deep Neural Network with Pruning, Trained Quantization and Huffman Coding , 2015, ICLR.
[36] Eunhyeok Park,et al. Compression of Deep Convolutional Neural Networks for Fast and Low Power Mobile Applications , 2015, ICLR.
[37] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[38] Misha Denil,et al. Predicting Parameters in Deep Learning , 2014 .
[39] Yoshua Bengio,et al. Low precision arithmetic for deep learning , 2014, ICLR.
[40] Andrew Zisserman,et al. Very Deep Convolutional Networks for Large-Scale Image Recognition , 2014, ICLR.
[41] Joan Bruna,et al. Exploiting Linear Structure Within Convolutional Networks for Efficient Evaluation , 2014, NIPS.
[42] Dumitru Erhan,et al. Going deeper with convolutions , 2014, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[43] Stephen P. Boyd,et al. A General System for Heuristic Solution of Convex Problems over Nonconvex Sets , 2016, 1601.07277.
[44] Ran El-Yaniv,et al. Binarized Neural Networks , 2016, NIPS.
[45] Ran El-Yaniv,et al. Binarized Neural Networks , 2016, ArXiv.
[46] Sergey Ioffe,et al. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift , 2015, ICML.
[47] Daniel Soudry,et al. Training Binary Multilayer Neural Networks for Image Classification using Expectation Backpropagation , 2015, ArXiv.