暂无分享,去创建一个
[1] Stephen I. Gallant,et al. Connectionist expert systems , 1988, CACM.
[2] Yan Liu,et al. Detecting Statistical Interactions from Neural Network Weights , 2017, ICLR.
[3] W. H. Zurek. Complexity, Entropy and the Physics of Information , 1990 .
[4] Franco Scarselli,et al. On the Complexity of Neural Network Classifiers: A Comparison Between Shallow and Deep Architectures , 2014, IEEE Transactions on Neural Networks and Learning Systems.
[5] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[6] Michael W. Mahoney,et al. Heavy-Tailed Universality Predicts Trends in Test Accuracies for Very Large Pre-Trained Deep Neural Networks , 2019, SDM.
[7] Surya Ganguli,et al. Resurrecting the sigmoid in deep learning through dynamical isometry: theory and practice , 2017, NIPS.
[8] M. Süzen,et al. Spectral Ergodicity in Deep Learning Architectures via Surrogate Random Matrices. , 2017 .
[9] Shane Legg,et al. Human-level control through deep reinforcement learning , 2015, Nature.
[10] Jürgen Schmidhuber,et al. Deep learning in neural networks: An overview , 2014, Neural Networks.
[11] Natalia Gimelshein,et al. PyTorch: An Imperative Style, High-Performance Deep Learning Library , 2019, NeurIPS.
[12] Karsten M. Borgwardt,et al. Neural Persistence: A Complexity Measure for Deep Neural Networks Using Algebraic Topology , 2018, ICLR.
[13] S M Pincus,et al. Approximate entropy as a measure of system complexity. , 1991, Proceedings of the National Academy of Sciences of the United States of America.
[14] P. Landsberg,et al. Simple measure for complexity , 1999 .
[15] Tin Kam Ho,et al. Complexity Measures of Supervised Classification Problems , 2002, IEEE Trans. Pattern Anal. Mach. Intell..
[16] Andrew Zisserman,et al. Very Deep Convolutional Networks for Large-Scale Image Recognition , 2014, ICLR.
[17] Luca Antiga,et al. Automatic differentiation in PyTorch , 2017 .
[18] Heikki Huttunen,et al. HARK Side of Deep Learning - From Grad Student Descent to Automated Machine Learning , 2019, ArXiv.
[19] Yann LeCun,et al. Towards Understanding the Role of Over-Parametrization in Generalization of Neural Networks , 2018, ArXiv.
[20] B. Pompe,et al. Permutation entropy: a natural complexity measure for time series. , 2002, Physical review letters.
[21] Surya Ganguli,et al. The Emergence of Spectral Universality in Deep Networks , 2018, AISTATS.
[22] Frank Hutter,et al. Neural Architecture Search: A Survey , 2018, J. Mach. Learn. Res..
[23] A. Jackson,et al. Spectral ergodicity and normal modes in ensembles of sparse matrices , 2001 .
[24] Michael W. Mahoney,et al. Implicit Self-Regularization in Deep Neural Networks: Evidence from Random Matrix Theory and Implications for Learning , 2018, J. Mach. Learn. Res..
[25] Michael W. Mahoney,et al. Traditional and Heavy-Tailed Self Regularization in Neural Network Models , 2019, ICML.
[26] Yann Dauphin,et al. Empirical Analysis of the Hessian of Over-Parametrized Neural Networks , 2017, ICLR.