暂无分享,去创建一个
[1] Surya Ganguli,et al. On the Expressive Power of Deep Neural Networks , 2016, ICML.
[2] Ohad Shamir,et al. The Power of Depth for Feedforward Neural Networks , 2015, COLT.
[3] M. Nica,et al. Products of Many Large Random Matrices and Gradients in Deep Neural Networks , 2018, Communications in Mathematical Physics.
[4] T. Poggio,et al. Deep vs. shallow networks : An approximation theory perspective , 2016, ArXiv.
[5] Rich Caruana,et al. Do Deep Nets Really Need to be Deep? , 2013, NIPS.
[6] Dmitry Yarotsky,et al. Error bounds for approximations with deep ReLU networks , 2016, Neural Networks.
[7] Ohad Shamir,et al. Failures of Gradient-Based Deep Learning , 2017, ICML.
[8] Philipp Petersen,et al. Optimal approximation of piecewise smooth functions using deep ReLU neural networks , 2017, Neural Networks.
[9] Boris Hanin,et al. Which Neural Net Architectures Give Rise To Exploding and Vanishing Gradients? , 2018, NeurIPS.
[10] Christian Tjandraatmadja,et al. Bounding and Counting Linear Regions of Deep Neural Networks , 2017, ICML.
[11] Yoshua Bengio,et al. A Closer Look at Memorization in Deep Networks , 2017, ICML.
[12] Matus Telgarsky,et al. Representation Benefits of Deep Feedforward Networks , 2015, ArXiv.
[13] Tomaso Poggio,et al. Learning Functions: When Is Deep Better Than Shallow , 2016, 1603.00988.
[14] Razvan Pascanu,et al. On the Number of Linear Regions of Deep Neural Networks , 2014, NIPS.
[15] Raman Arora,et al. Understanding Deep Neural Networks with Rectified Linear Units , 2016, Electron. Colloquium Comput. Complex..
[16] Max Tegmark,et al. Why Does Deep and Cheap Learning Work So Well? , 2016, Journal of Statistical Physics.
[17] Franco Scarselli,et al. On the Complexity of Neural Network Classifiers: A Comparison Between Shallow and Deep Architectures , 2014, IEEE Transactions on Neural Networks and Learning Systems.
[18] Matus Telgarsky,et al. Benefits of Depth in Neural Networks , 2016, COLT.
[19] Dmitry Yarotsky,et al. Optimal approximation of continuous functions by very deep ReLU networks , 2018, COLT.
[20] David Rolnick,et al. How to Start Training: The Effect of Initialization and Architecture , 2018, NeurIPS.
[21] Nathan Srebro,et al. Exploring Generalization in Deep Learning , 2017, NIPS.
[22] Lorenzo Rosasco,et al. Why and when can deep-but not shallow-networks avoid the curse of dimensionality: A review , 2016, International Journal of Automation and Computing.
[23] David Rolnick,et al. The power of deeper networks for expressing natural functions , 2017, ICLR.
[24] Surya Ganguli,et al. Exponential expressivity in deep neural networks through transient chaos , 2016, NIPS.
[25] Jascha Sohl-Dickstein,et al. Sensitivity and Generalization in Neural Networks: an Empirical Study , 2018, ICLR.
[26] Samy Bengio,et al. Understanding deep learning requires rethinking generalization , 2016, ICLR.