暂无分享,去创建一个
[1] Yoram Singer,et al. Toward Deeper Understanding of Neural Networks: The Power of Initialization and a Dual View on Expressivity , 2016, NIPS.
[2] Yann Dauphin,et al. Convolutional Sequence to Sequence Learning , 2017, ICML.
[3] Joel A. Tropp,et al. User-Friendly Tail Bounds for Sums of Random Matrices , 2010, Found. Comput. Math..
[4] Inderjit S. Dhillon,et al. Recovery Guarantees for One-hidden-layer Neural Networks , 2017, ICML.
[5] Inderjit S. Dhillon,et al. Mixed Linear Regression with Multiple Components , 2016, NIPS.
[6] Tengyu Ma,et al. Identity Matters in Deep Learning , 2016, ICLR.
[7] René Vidal,et al. Global Optimality in Tensor Factorization, Deep Learning, and Beyond , 2015, ArXiv.
[8] Roi Livni,et al. On the Computational Efficiency of Training Neural Networks , 2014, NIPS.
[9] Johan Håstad,et al. Tensor Rank is NP-Complete , 1989, ICALP.
[10] Yuandong Tian,et al. Symmetry-Breaking Convergence Analysis of Certain Two-layered Neural Networks with ReLU nonlinearity , 2017, ICLR.
[11] Yuandong Tian,et al. An Analytical Formula of Population Gradient for two-layered ReLU network and its Applications in Convergence and Critical Point Analysis , 2017, ICML.
[12] Surya Ganguli,et al. On the Expressive Power of Deep Neural Networks , 2016, ICML.
[13] Anima Anandkumar,et al. Provable Methods for Training Neural Networks with Sparse Connectivity , 2014, ICLR.
[14] David P. Woodruff,et al. Relative Error Tensor Low Rank Approximation , 2017, Electron. Colloquium Comput. Complex..
[15] Amnon Shashua,et al. Convolutional Rectifier Networks as Generalized Tensor Decompositions , 2016, ICML.
[16] Surya Ganguli,et al. Exponential expressivity in deep neural networks through transient chaos , 2016, NIPS.
[17] Percy Liang,et al. Tensor Factorization via Matrix Factorization , 2015, AISTATS.
[18] Dean Alderucci. A SPECTRAL ALGORITHM FOR LEARNING HIDDEN MARKOV MODELS THAT HAVE SILENT STATES , 2015 .
[19] Anima Anandkumar,et al. Tensor decompositions for learning latent variable models , 2012, J. Mach. Learn. Res..
[20] Demis Hassabis,et al. Mastering the game of Go with deep neural networks and tree search , 2016, Nature.
[21] Le Song,et al. Diversity Leads to Generalization in Neural Networks , 2016, ArXiv.
[22] Surya Ganguli,et al. Identifying and attacking the saddle point problem in high-dimensional non-convex optimization , 2014, NIPS.
[23] Alexander J. Smola,et al. Fast and Guaranteed Tensor Decomposition via Sketching , 2015, NIPS.
[24] Nadav Cohen,et al. On the Expressive Power of Deep Learning: A Tensor Analysis , 2015, COLT 2016.
[25] Geoffrey E. Hinton,et al. ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.
[26] Anima Anandkumar,et al. Online and Differentially-Private Tensor Decomposition , 2016, NIPS.
[27] Geoffrey E. Hinton,et al. Deep Learning , 2015, Nature.
[28] Christopher J. Hillar,et al. Most Tensor Problems Are NP-Hard , 2009, JACM.
[29] Prateek Jain,et al. Low-rank matrix completion using alternating minimization , 2012, STOC '13.
[30] Amir Globerson,et al. Globally Optimal Gradient Descent for a ConvNet with Gaussian Inputs , 2017, ICML.
[31] Ah Chung Tsoi,et al. Face recognition: a convolutional neural-network approach , 1997, IEEE Trans. Neural Networks.
[32] Yuandong Tian,et al. When is a Convolutional Filter Easy To Learn? , 2017, ICLR.
[33] David P. Woodruff,et al. Sublinear Time Orthogonal Tensor Decomposition , 2016, NIPS.
[34] Martin J. Wainwright,et al. Convexified Convolutional Neural Networks , 2016, ICML.
[35] Anima Anandkumar,et al. Beating the Perils of Non-Convexity: Guaranteed Training of Neural Networks using Tensor Methods , 2017 .
[36] Matus Telgarsky,et al. Benefits of Depth in Neural Networks , 2016, COLT.
[37] Ohad Shamir,et al. On the Quality of the Initial Basin in Overspecified Neural Networks , 2015, ICML.