暂无分享,去创建一个
[1] Gulshan Kumar,et al. A Survey of Deep Learning and Its Applications: A New Paradigm to Machine Learning , 2019, Archives of Computational Methods in Engineering.
[2] J. Sola,et al. Importance of input data normalization for the application of neural networks to complex industrial problems , 1997 .
[3] Klaus-Robert Müller,et al. Efficient BackProp , 2012, Neural Networks: Tricks of the Trade.
[4] Geoffrey E. Hinton,et al. On the importance of initialization and momentum in deep learning , 2013, ICML.
[5] Quoc V. Le,et al. Swish: a Self-Gated Activation Function , 2017, 1710.05941.
[6] Andrew L. Maas. Rectifier Nonlinearities Improve Neural Network Acoustic Models , 2013 .
[7] Scott C. Douglas,et al. Why RELU Units Sometimes Die: Analysis of Single-Unit Error Backpropagation in Neural Networks , 2018, 2018 52nd Asilomar Conference on Signals, Systems, and Computers.
[8] Geoffrey E. Hinton,et al. Rectified Linear Units Improve Restricted Boltzmann Machines , 2010, ICML.
[9] Geoffrey E. Hinton,et al. Learning representations by back-propagating errors , 1986, Nature.
[10] Abien Fred Agarap. Deep Learning using Rectified Linear Units (ReLU) , 2018, ArXiv.
[11] David Silver,et al. Learning values across many orders of magnitude , 2016, NIPS.
[12] Sergey Ioffe,et al. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift , 2015, ICML.
[13] Shou-De Lin,et al. ANS: Adaptive Network Scaling for Deep Rectifier Reinforcement Learning Models , 2018, ArXiv.
[14] Brahim Chaib-draa,et al. Parametric Exponential Linear Unit for Deep Convolutional Neural Networks , 2016, 2017 16th IEEE International Conference on Machine Learning and Applications (ICMLA).
[15] Geoffrey E. Hinton,et al. Deep Learning , 2015, Nature.
[16] Sepp Hochreiter,et al. Self-Normalizing Neural Networks , 2017, NIPS.
[17] Aleksander Madry,et al. How Does Batch Normalization Help Optimization? (No, It Is Not About Internal Covariate Shift) , 2018, NeurIPS.
[18] Geoffrey E. Hinton,et al. Layer Normalization , 2016, ArXiv.
[19] Lu Lu,et al. Dying ReLU and Initialization: Theory and Numerical Examples , 2019, Communications in Computational Physics.
[20] Tapani Raiko,et al. Deep Learning Made Easier by Linear Transformations in Perceptrons , 2012, AISTATS.
[21] Sepp Hochreiter,et al. Fast and Accurate Deep Network Learning by Exponential Linear Units (ELUs) , 2015, ICLR.
[22] Quoc V. Le,et al. Searching for Activation Functions , 2018, arXiv.
[23] Philip Bachman,et al. Deep Reinforcement Learning that Matters , 2017, AAAI.
[24] Xiaogang Wang,et al. Deeply learned face representations are sparse, selective, and robust , 2014, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[25] Jian Sun,et al. Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification , 2015, 2015 IEEE International Conference on Computer Vision (ICCV).
[26] Liwei Wang,et al. Gradient Descent Finds Global Minima of Deep Neural Networks , 2018, ICML.
[27] Yoshua Bengio,et al. Deep Sparse Rectifier Neural Networks , 2011, AISTATS.
[28] Yoshua Bengio,et al. Understanding the difficulty of training deep feedforward neural networks , 2010, AISTATS.
[29] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[30] Boris Hanin,et al. Which Neural Net Architectures Give Rise To Exploding and Vanishing Gradients? , 2018, NeurIPS.
[31] Gabriel Goh,et al. Why Momentum Really Works , 2017 .
[32] Jürgen Schmidhuber,et al. Training Very Deep Networks , 2015, NIPS.
[33] Surya Ganguli,et al. Exponential expressivity in deep neural networks through transient chaos , 2016, NIPS.
[34] Dina Q. Goldin,et al. On Similarity Queries for Time-Series Data: Constraint Specification and Implementation , 1995, CP.