暂无分享,去创建一个
Brian Chmiel | Liad Ben-Uri | Moran Shkolnik | Elad Hoffer | Ron Banner | Daniel Soudry | Daniel Soudry | Elad Hoffer | Ron Banner | Liad Ben-Uri | Brian Chmiel | Moran Shkolnik
[1] Jose Javier Gonzalez Ortiz,et al. What is the State of Neural Network Pruning? , 2020, MLSys.
[2] Janowsky,et al. Pruning versus clipping in neural networks. , 1989, Physical review. A, General physics.
[3] Michael Carbin,et al. The Lottery Ticket Hypothesis: Finding Sparse, Trainable Neural Networks , 2018, ICLR.
[4] C. A. R. Hoare,et al. Algorithm 65: find , 1961, Commun. ACM.
[5] Daniel Soudry,et al. Post training 4-bit quantization of convolutional networks for rapid-deployment , 2018, NeurIPS.
[6] David Thorsley,et al. Post-training Piecewise Linear Quantization for Deep Neural Networks , 2020, ECCV.
[7] Avi Mendelson,et al. Loss Aware Post-training Quantization , 2019, ArXiv.
[8] Avi Mendelson,et al. NICE: Noise Injection and Clamping Estimation for Neural Network Quantization , 2018, Mathematics.
[9] Tor M. Aamodt,et al. Sparse Weight Activation Training , 2020, NeurIPS.
[10] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[11] Andrew McCallum,et al. Energy and Policy Considerations for Deep Learning in NLP , 2019, ACL.
[12] Michael C. Mozer,et al. Using Relevance to Reduce Network Size Automatically , 1989 .
[13] N. Smirnov. Table for Estimating the Goodness of Fit of Empirical Distributions , 1948 .
[14] Kilian Q. Weinberger,et al. Densely Connected Convolutional Networks , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[15] Hao Wu,et al. Mixed Precision Training , 2017, ICLR.
[16] Erich Elsen,et al. The State of Sparsity in Deep Neural Networks , 2019, ArXiv.
[17] Elad Hoffer,et al. Scalable Methods for 8-bit Training of Neural Networks , 2018, NeurIPS.
[18] Kamyar Azizzadenesheli,et al. signSGD: compressed optimisation for non-convex problems , 2018, ICML.
[19] Albert Gural,et al. Trained Uniform Quantization for Accurate and Efficient Neural Network Inference on Fixed-Point Hardware , 2019, ArXiv.
[20] Swagath Venkataramani,et al. Bridging the Accuracy Gap for 2-bit Quantized Neural Networks (QNN) , 2018, ArXiv.
[21] Alexander Heinecke,et al. Leveraging the bfloat16 Artificial Intelligence Datatype For Higher-Precision Computations , 2019, 2019 IEEE 26th Symposium on Computer Arithmetic (ARITH).
[22] Andrew Zisserman,et al. Very Deep Convolutional Networks for Large-Scale Image Recognition , 2014, ICLR.
[23] Pradeep Dubey,et al. A Study of BFLOAT16 for Deep Learning Training , 2019, ArXiv.
[24] Philip H. S. Torr,et al. A Signal Propagation Perspective for Pruning Neural Networks at Initialization , 2019, ICLR.
[25] Shuang Wu,et al. Training and Inference with Integers in Deep Neural Networks , 2018, ICLR.
[26] Xu Sun,et al. meProp: Sparsified Back Propagation for Accelerated Deep Learning with Reduced Overfitting , 2017, ICML.
[27] Daniel Brand,et al. Training Deep Neural Networks with 8-bit Floating Point Numbers , 2018, NeurIPS.
[28] Li Fei-Fei,et al. ImageNet: A large-scale hierarchical image database , 2009, CVPR.
[29] Swagath Venkataramani,et al. Hybrid 8-bit Floating Point (HFP8) Training and Inference for Deep Neural Networks , 2019, NeurIPS.
[30] Wojciech Samek,et al. Dithered backprop: A sparse and quantized backpropagation algorithm for more efficient deep neural network training , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW).
[31] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[32] Raghuraman Krishnamoorthi,et al. Quantizing deep convolutional networks for efficient inference: A whitepaper , 2018, ArXiv.
[33] G. Biau,et al. High-Dimensional \(p\)-Norms , 2013, 1311.0587.
[34] Max Welling,et al. Learning Sparse Neural Networks through L0 Regularization , 2017, ICLR.
[35] Weisheng Zhao,et al. Accelerating CNN Training by Sparsifying Activation Gradients , 2019, ArXiv.
[36] Michael C. Mozer,et al. Skeletonization: A Technique for Trimming the Fat from a Network via Relevance Assessment , 1988, NIPS.
[37] Ehud D. Karnin,et al. A simple procedure for pruning back-propagation trained neural networks , 1990, IEEE Trans. Neural Networks.