暂无分享,去创建一个
[1] Misha Denil,et al. Predicting Parameters in Deep Learning , 2014 .
[2] Eduardo Sontag,et al. For neural networks, function determines form , 1992, [1992] Proceedings of the 31st IEEE Conference on Decision and Control.
[3] Ryan P. Adams,et al. Gradient-based Hyperparameter Optimization through Reversible Learning , 2015, ICML.
[4] Peter L. Bartlett,et al. Efficient agnostic learning of neural networks with bounded fan-in , 1996, IEEE Trans. Inf. Theory.
[5] C. Fefferman. Reconstructing a neural net from its output , 1994 .
[6] Nicolas Le Roux,et al. Convex Neural Networks , 2005, NIPS.
[7] Raquel Urtasun,et al. The Reversible Residual Network: Backpropagation Without Storing Activations , 2017, NIPS.
[8] Francis R. Bach,et al. Breaking the Curse of Dimensionality with Convex Neural Networks , 2014, J. Mach. Learn. Res..
[9] Razvan Pascanu,et al. On the Number of Linear Regions of Deep Neural Networks , 2014, NIPS.
[10] Kurt Hornik,et al. Approximation capabilities of multilayer feedforward networks , 1991, Neural Networks.
[11] Tomaso Poggio,et al. Learning Functions: When Is Deep Better Than Shallow , 2016, 1603.00988.
[12] Tengyu Ma,et al. Identity Matters in Deep Learning , 2016, ICLR.
[13] Samy Bengio,et al. Density estimation using Real NVP , 2016, ICLR.
[14] Philip M. Long,et al. Gradient Descent with Identity Initialization Efficiently Learns Positive-Definite Linear Transformations by Deep Residual Networks , 2018, Neural Computation.
[15] Matus Telgarsky,et al. Representation Benefits of Deep Feedforward Networks , 2015, ArXiv.
[16] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[17] Surya Ganguli,et al. Exponential expressivity in deep neural networks through transient chaos , 2016, NIPS.