A RANDOM MATRIX PERSPECTIVE
暂无分享,去创建一个
Jeffrey Pennington | Ben Adlam | Jake Levinson | Jeffrey Pennington | Ben Adlam | J. Levinson | HIGH DIMENSIONS | High Dimensions
[1] H. Yau,et al. A Dynamical Approach to Random Matrix Theory , 2017 .
[2] Jeffrey Pennington,et al. Geometry of Neural Network Loss Surfaces via Random Matrix Theory , 2017, ICML.
[3] Andrea Montanari,et al. A mean field view of the landscape of two-layer neural networks , 2018, Proceedings of the National Academy of Sciences.
[4] T. Tao. Topics in Random Matrix Theory , 2012 .
[5] M. Peligrad,et al. On the limiting spectral distribution for a large class of symmetric random matrices with correlated entries , 2015 .
[6] Zhenyu Liao,et al. On the Spectrum of Random Features Maps of High Dimensional Data , 2018, ICML.
[7] Zhidong Bai,et al. LARGE SAMPLE COVARIANCE MATRICES WITHOUT INDEPENDENCE STRUCTURES IN COLUMNS , 2008 .
[8] J. W. Silverstein,et al. On the empirical distribution of eigenvalues of a class of large dimensional random matrices , 1995 .
[9] Noureddine El Karoui,et al. The spectrum of kernel random matrices , 2010, 1001.0492.
[10] Andrea Montanari,et al. Mean-field theory of two-layers neural networks: dimension-free bounds and kernel limit , 2019, COLT.
[11] Surya Ganguli,et al. An analytic theory of generalization dynamics and transfer learning in deep linear networks , 2018, ICLR.
[12] Zhenyu Liao,et al. The Dynamics of Learning: A Random Matrix Approach , 2018, ICML.
[13] Andrew M. Saxe,et al. High-dimensional dynamics of generalization error in neural networks , 2017, Neural Networks.
[14] Francis Bach,et al. A Note on Lazy Training in Supervised Differentiable Programming , 2018, ArXiv.
[15] Debashis Paul,et al. No eigenvalues outside the support of the limiting empirical spectral distribution of a separable covariance matrix , 2009, J. Multivar. Anal..
[16] Jeffrey Pennington,et al. The Spectrum of the Fisher Information Matrix of a Single-Hidden-Layer Neural Network , 2018, NeurIPS.
[17] Tara N. Sainath,et al. Deep Neural Networks for Acoustic Modeling in Speech Recognition: The Shared Views of Four Research Groups , 2012, IEEE Signal Processing Magazine.
[18] Francis Bach,et al. On the Global Convergence of Gradient Descent for Over-parameterized Models using Optimal Transport , 2018, NeurIPS.
[19] Arthur Jacot,et al. Neural tangent kernel: convergence and generalization in neural networks (invited paper) , 2018, NeurIPS.
[20] Jeffrey Pennington,et al. Nonlinear random matrix theory for deep learning , 2019, NIPS.
[21] Surya Ganguli,et al. Statistical Mechanics of Optimal Convex Inference in High Dimensions , 2016 .
[22] Andrea Montanari,et al. Surprises in High-Dimensional Ridgeless Least Squares Interpolation , 2019, Annals of statistics.
[23] Zhenyu Liao,et al. A Random Matrix Approach to Neural Networks , 2017, ArXiv.
[24] Joan Bruna,et al. Global convergence of neuron birth-death dynamics , 2019, ICML 2019.
[25] George Kurian,et al. Google's Neural Machine Translation System: Bridging the Gap between Human and Machine Translation , 2016, ArXiv.
[26] Jaehoon Lee,et al. Wide neural networks of any depth evolve as linear models under gradient descent , 2019, NeurIPS.
[27] Grant M. Rotskoff,et al. Neural Networks as Interacting Particle Systems: Asymptotic Convexity of the Loss Landscape and Universal Scaling of the Approximation Error , 2018, ArXiv.