A Mean Field Theory of Quantized Deep Networks: The Quantization-Depth Trade-Off
暂无分享,去创建一个
[1] Jascha Sohl-Dickstein,et al. Dynamical Isometry and a Mean Field Theory of CNNs: How to Train 10, 000-Layer Vanilla Convolutional Neural Networks , 2018, ICML.
[2] Ran El-Yaniv,et al. Binarized Neural Networks , 2016, ArXiv.
[3] Eduardo Sontag,et al. Turing computability with neural nets , 1991 .
[4] Hanan Samet,et al. Training Quantized Nets: A Deeper Understanding , 2017, NIPS.
[5] Jaehoon Lee,et al. Deep Neural Networks as Gaussian Processes , 2017, ICLR.
[6] Daisuke Miyashita,et al. Convolutional Neural Networks using Logarithmic Data Representation , 2016, ArXiv.
[7] Jack Xin,et al. Understanding Straight-Through Estimator in Training Activation Quantized Neural Nets , 2019, ICLR.
[8] Eriko Nurvitadhi,et al. WRPN: Wide Reduced-Precision Networks , 2017, ICLR.
[9] Daniel Brand,et al. Training Deep Neural Networks with 8-bit Floating Point Numbers , 2018, NeurIPS.
[10] Sebastian Nowozin,et al. Deterministic Variational Inference for Robust Bayesian Neural Networks , 2018, ICLR.
[11] Ran El-Yaniv,et al. Quantized Neural Networks: Training Neural Networks with Low Precision Weights and Activations , 2016, J. Mach. Learn. Res..
[12] Alexander G. Anderson,et al. The High-Dimensional Geometry of Binary Neural Networks , 2017, ICLR.
[13] Samuel S. Schoenholz,et al. Dynamical Isometry and a Mean Field Theory of RNNs: Gating Enables Signal Propagation in Recurrent Neural Networks , 2018, ICML.
[14] Tengyu Ma,et al. Fixup Initialization: Residual Learning Without Normalization , 2019, ICLR.
[15] Elad Hoffer,et al. Scalable Methods for 8-bit Training of Neural Networks , 2018, NeurIPS.
[16] Arthur Jacot,et al. Neural tangent kernel: convergence and generalization in neural networks (invited paper) , 2018, NeurIPS.
[17] Wei Pan,et al. Towards Accurate Binary Convolutional Neural Network , 2017, NIPS.
[18] Richard E. Turner,et al. Gaussian Process Behaviour in Wide Deep Neural Networks , 2018, ICLR.
[19] Richard F. Lyon,et al. Neural Networks for Machine Learning , 2017 .
[20] Surya Ganguli,et al. Exponential expressivity in deep neural networks through transient chaos , 2016, NIPS.
[21] Igor Carron,et al. XNOR-Net: ImageNet Classification Using Binary Convolutional Neural Networks , 2016 .
[22] Jascha Sohl-Dickstein,et al. A Mean Field Theory of Batch Normalization , 2019, ICLR.
[23] Samy Bengio,et al. Understanding deep learning requires rethinking generalization , 2016, ICLR.
[24] Surya Ganguli,et al. Deep Information Propagation , 2016, ICLR.
[25] Surya Ganguli,et al. Resurrecting the sigmoid in deep learning through dynamical isometry: theory and practice , 2017, NIPS.
[26] Yoshua Bengio,et al. Understanding the difficulty of training deep feedforward neural networks , 2010, AISTATS.
[27] Yixin Chen,et al. Compressing Neural Networks with the Hashing Trick , 2015, ICML.
[28] Ruosong Wang,et al. On Exact Computation with an Infinitely Wide Neural Net , 2019, NeurIPS.
[29] Pritish Narayanan,et al. Deep Learning with Limited Numerical Precision , 2015, ICML.
[30] Jaehoon Lee,et al. Wide neural networks of any depth evolve as linear models under gradient descent , 2019, NeurIPS.
[31] Bo Chen,et al. MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications , 2017, ArXiv.
[32] Pekka Orponen,et al. On the Effect of Analog Noise in Discrete-Time Analog Computations , 1996, Neural Computation.
[33] Pradeep Dubey,et al. Mixed Precision Training of Convolutional Neural Networks using Integer Operations , 2018, ICLR.
[34] Seyed-Mohsen Moosavi-Dezfooli,et al. Adaptive Quantization for Deep Neural Network , 2017, AAAI.
[35] Samuel S. Schoenholz,et al. Dynamical Isometry and a Mean Field Theory of LSTMs and GRUs , 2019, ArXiv.
[36] Parul Parashar,et al. Neural Networks in Machine Learning , 2014 .
[37] Samuel S. Schoenholz,et al. Mean Field Residual Networks: On the Edge of Chaos , 2017, NIPS.