Memorization in Overparameterized Autoencoders
暂无分享,去创建一个
Mikhail Belkin | Caroline Uhler | Adityanarayanan Radhakrishnan | Karren Yang | Karren D. Yang | Mikhail Belkin | Caroline Uhler | Adityanarayanan Radhakrishnan | M. Belkin
[1] Michael W. Mahoney,et al. Implicit Self-Regularization in Deep Neural Networks: Evidence from Random Matrix Theory and Implications for Learning , 2018, J. Mach. Learn. Res..
[2] Mikhail Belkin,et al. Eigenvectors of Orthogonally Decomposable Functions , 2014, SIAM J. Comput..
[3] Samy Bengio,et al. Understanding deep learning requires rethinking generalization , 2016, ICLR.
[4] Luca Antiga,et al. Automatic differentiation in PyTorch , 2017 .
[5] Tianqi Chen,et al. Empirical Evaluation of Rectified Activations in Convolutional Network , 2015, ArXiv.
[6] Gaël Varoquaux,et al. Scikit-learn: Machine Learning in Python , 2011, J. Mach. Learn. Res..
[7] Ryota Tomioka,et al. In Search of the Real Inductive Bias: On the Role of Implicit Regularization in Deep Learning , 2014, ICLR.
[8] Ping Li,et al. On Random Deep Weight-Tied Autoencoders: Exact Asymptotic Analysis, Phase Transitions, and Implications to Training , 2018, ICLR.
[9] Thomas Hofmann,et al. Greedy Layer-Wise Training of Deep Networks , 2007 .
[10] Yoshua Bengio,et al. Understanding the difficulty of training deep feedforward neural networks , 2010, AISTATS.
[11] Andrea Vedaldi,et al. Deep Image Prior , 2020, International Journal of Computer Vision.
[12] D. Chakrabarti,et al. A fast fixed - point algorithm for independent component analysis , 1997 .
[13] Nathan Srebro,et al. The Implicit Bias of Gradient Descent on Separable Data , 2017, J. Mach. Learn. Res..
[14] Michael Eickenberg,et al. Greedy Layerwise Learning Can Scale to ImageNet , 2018, ICML.
[15] H. Engl,et al. Regularization of Inverse Problems , 1996 .
[16] Yoshua Bengio,et al. Gradient-based learning applied to document recognition , 1998, Proc. IEEE.
[17] Aapo Hyvärinen,et al. A Fast Fixed-Point Algorithm for Independent Component Analysis , 1997, Neural Computation.
[18] Yuanzhi Li,et al. Learning Overparameterized Neural Networks via Stochastic Gradient Descent on Structured Data , 2018, NeurIPS.
[19] Jian Sun,et al. Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification , 2015, 2015 IEEE International Conference on Computer Vision (ICCV).
[20] Nathan Srebro,et al. How do infinite width bounded norm networks look in function space? , 2019, COLT.
[21] Mikhail Belkin,et al. Reconciling modern machine-learning practice and the classical bias–variance trade-off , 2018, Proceedings of the National Academy of Sciences.
[22] Thomas Brox,et al. U-Net: Convolutional Networks for Biomedical Image Segmentation , 2015, MICCAI.
[23] Peter Glöckner,et al. Why Does Unsupervised Pre-training Help Deep Learning? , 2013 .
[24] Pierre Baldi,et al. Autoencoders, Unsupervised Learning, and Deep Architectures , 2011, ICML Unsupervised and Transfer Learning.
[25] Yoshua Bengio,et al. A Closer Look at Memorization in Deep Networks , 2017, ICML.
[26] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[27] Cordelia Schmid,et al. Déjà Vu: an empirical evaluation of the memorization properties of ConvNets , 2018, ArXiv.
[28] Samy Bengio,et al. Identity Crisis: Memorization and Generalization under Extreme Overparameterization , 2019, ICLR.
[29] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[30] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[31] Mikhail Belkin,et al. Reconciling modern machine learning and the bias-variance trade-off , 2018, ArXiv.
[32] Shai Shalev-Shwartz,et al. SGD Learns Over-parameterized Networks that Provably Generalize on Linearly Separable Data , 2017, ICLR.