暂无分享,去创建一个
Joan Bruna | Stefano Soatto | René Vidal | Raja Giryes | R. Vidal | Joan Bruna | Stefano Soatto | R. Giryes
[1] P. Werbos,et al. Beyond Regression : "New Tools for Prediction and Analysis in the Behavioral Sciences , 1974 .
[2] Geoffrey E. Hinton,et al. Learning representations by back-propagating errors , 1986, Nature.
[3] J. Slawny,et al. Back propagation fails to separate where perceptrons succeed , 1989 .
[4] George Cybenko,et al. Approximation by superpositions of a sigmoidal function , 1989, Math. Control. Signals Syst..
[5] Kurt Hornik,et al. Neural networks and principal component analysis: Learning from examples without local minima , 1989, Neural Networks.
[6] Kurt Hornik,et al. Multilayer feedforward networks are universal approximators , 1989, Neural Networks.
[7] A. Tesi,et al. Backpropagation converges for multi-layered networks and linearly-separable patterns , 1991, IJCNN-91-Seattle International Joint Conference on Neural Networks.
[8] A. Barron. Approximation and Estimation Bounds for Artificial Neural Networks , 1991, COLT '91.
[9] Kurt Hornik,et al. Approximation capabilities of multilayer feedforward networks , 1991, Neural Networks.
[10] Alberto Tesi,et al. On the Problem of Local Minima in Backpropagation , 1992, IEEE Trans. Pattern Anal. Mach. Intell..
[11] Geoffrey E. Hinton,et al. Keeping the neural networks simple by minimizing the description length of the weights , 1993, COLT '93.
[12] Peter L. Bartlett,et al. Boosting Algorithms as Gradient Descent , 1999, NIPS.
[13] Vladimir Vapnik,et al. An overview of statistical learning theory , 1999, IEEE Trans. Neural Networks.
[14] Naftali Tishby,et al. The information bottleneck method , 2000, ArXiv.
[15] Peter L. Bartlett,et al. Rademacher and Gaussian Complexities: Risk Bounds and Structural Results , 2003, J. Mach. Learn. Res..
[16] J. Friedman. Greedy function approximation: A gradient boosting machine. , 2001 .
[17] Peter L. Bartlett,et al. Vapnik-Chervonenkis dimension of neural nets , 2003 .
[18] Lior Wolf,et al. Learning over Sets using Kernel Principal Angles , 2003, J. Mach. Learn. Res..
[19] Renato D. C. Monteiro,et al. Local Minima and Convergence in Low-Rank Semidefinite Programming , 2005, Math. Program..
[20] Nicolas Le Roux,et al. Convex Neural Networks , 2005, NIPS.
[21] Emmanuel J. Candès,et al. Near-Optimal Signal Recovery From Random Projections: Universal Encoding Strategies? , 2004, IEEE Transactions on Information Theory.
[22] Stefano Soatto,et al. On the set of images modulo viewpoint and contrast changes , 2009, CVPR.
[23] Fei-Fei Li,et al. ImageNet: A large-scale hierarchical image database , 2009, 2009 IEEE Conference on Computer Vision and Pattern Recognition.
[24] David A. McAllester,et al. Object Detection with Discriminatively Trained Part Based Models , 2010, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[25] Geoffrey E. Hinton,et al. Rectified Linear Units Improve Restricted Boltzmann Machines , 2010, ICML.
[26] Guillermo Sapiro,et al. Online Learning for Matrix Factorization and Sparse Coding , 2009, J. Mach. Learn. Res..
[27] Shie Mannor,et al. Robustness and generalization , 2010, Machine Learning.
[28] Stéphane Mallat,et al. Group Invariant Scattering , 2011, ArXiv.
[29] Pablo A. Parrilo,et al. The Convex Geometry of Linear Inverse Problems , 2010, Foundations of Computational Mathematics.
[30] S. Mallat,et al. Invariant Scattering Convolution Networks , 2013, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[31] Geoffrey E. Hinton,et al. ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.
[32] Tara N. Sainath,et al. Improving deep neural networks for LVCSR using rectified linear units and dropout , 2013, 2013 IEEE International Conference on Acoustics, Speech and Signal Processing.
[33] Andrew L. Maas. Rectifier Nonlinearities Improve Neural Network Acoustic Models , 2013 .
[34] Geoffrey E. Hinton,et al. On rectified linear units for speech processing , 2013, 2013 IEEE International Conference on Acoustics, Speech and Signal Processing.
[35] Wotao Yin,et al. A Block Coordinate Descent Method for Regularized Multiconvex Optimization with Applications to Nonnegative Tensor Factorization and Completion , 2013, SIAM J. Imaging Sci..
[36] R. Vidal,et al. Sparse Subspace Clustering: Algorithm, Theory, and Applications. , 2013, IEEE transactions on pattern analysis and machine intelligence.
[37] René Vidal,et al. Structured Low-Rank Matrix Factorization: Optimality, Algorithm, and Applications to Image Processing , 2014, ICML.
[38] Nitish Srivastava,et al. Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..
[39] Joan Bruna,et al. Signal recovery from Pooling Representations , 2013, ICML.
[40] Max Welling,et al. Auto-Encoding Variational Bayes , 2013, ICLR.
[41] Surya Ganguli,et al. Identifying and attacking the saddle point problem in high-dimensional non-convex optimization , 2014, NIPS.
[42] Yaniv Plan,et al. Dimension Reduction by Random Hyperplane Tessellations , 2014, Discret. Comput. Geom..
[43] Stefano Soatto,et al. Visual Representations: Defining Properties and Deep Approximations , 2014, ICLR 2016.
[44] Shai Ben-David,et al. Understanding Machine Learning - From Theory to Algorithms , 2014 .
[45] Ryota Tomioka,et al. Norm-Based Capacity Control in Neural Networks , 2015, COLT.
[46] Lorenzo Rosasco,et al. On Invariance and Selectivity in Representation Learning , 2015, ArXiv.
[47] René Vidal,et al. Global Optimality in Tensor Factorization, Deep Learning, and Beyond , 2015, ArXiv.
[48] Sergey Ioffe,et al. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift , 2015, ICML.
[49] Andrea Vedaldi,et al. Understanding deep image representations by inverting them , 2014, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[50] Ruslan Salakhutdinov,et al. Path-SGD: Path-Normalized Optimization in Deep Neural Networks , 2015, NIPS.
[51] Carlo Baldassi,et al. Local entropy as a measure for sampling solutions in Constraint Satisfaction Problems , 2015 .
[52] Geoffrey E. Hinton,et al. Deep Learning , 2015, Nature.
[53] Furong Huang,et al. Escaping From Saddle Points - Online Stochastic Gradient for Tensor Decomposition , 2015, COLT.
[54] Yann LeCun,et al. The Loss Surfaces of Multilayer Networks , 2014, AISTATS.
[55] A. Robert Calderbank,et al. Discriminative Robust Transformation Learning , 2015, NIPS.
[56] Ariel D. Procaccia,et al. Variational Dropout and the Local Reparameterization Trick , 2015, NIPS.
[57] Yoram Singer,et al. Train faster, generalize better: Stability of stochastic gradient descent , 2015, ICML.
[58] Stéphane Mallat,et al. Understanding deep convolutional networks , 2016, Philosophical Transactions of the Royal Society A: Mathematical, Physical and Engineering Sciences.
[59] Nikos Komodakis,et al. Wide Residual Networks , 2016, BMVC.
[60] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[61] Ohad Shamir,et al. On the Quality of the Initial Basin in Overspecified Neural Networks , 2015, ICML.
[62] Jian Sun,et al. Identity Mappings in Deep Residual Networks , 2016, ECCV.
[63] Guillermo Sapiro,et al. Deep Neural Networks with Random Gaussian Weights: A Universal Classification Strategy? , 2015, IEEE Transactions on Signal Processing.
[64] René Vidal,et al. Global Optimality in Neural Network Training , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[65] Pierre Vandergheynst,et al. Geometric Deep Learning: Going beyond Euclidean data , 2016, IEEE Signal Process. Mag..
[66] Joan Bruna,et al. Topology and Geometry of Half-Rectified Network Optimization , 2016, ICLR.
[67] Naftali Tishby,et al. Opening the Black Box of Deep Neural Networks via Information , 2017, ArXiv.
[68] Michael Elad,et al. Convolutional Neural Networks Analyzed via Convolutional Sparse Coding , 2016, J. Mach. Learn. Res..
[69] Samy Bengio,et al. Understanding deep learning requires rethinking generalization , 2016, ICLR.
[70] Razvan Pascanu,et al. Sharp Minima Can Generalize For Deep Nets , 2017, ICML.
[71] Anima Anandkumar,et al. Beating the Perils of Non-Convexity: Guaranteed Training of Neural Networks using Tensor Methods , 2017 .
[72] Mark W. Schmidt,et al. Minimizing finite sums with the stochastic average gradient , 2013, Mathematical Programming.
[73] Alexander A. Alemi,et al. Deep Variational Information Bottleneck , 2017, ICLR.
[74] Guillermo Sapiro,et al. Generalization Error of Invariant Classifiers , 2016, AISTATS.
[75] Stefano Soatto,et al. Entropy-SGD: biasing gradient descent into wide valleys , 2016, ICLR.
[76] Guillermo Sapiro,et al. Robust Large Margin Deep Neural Networks , 2017, IEEE Transactions on Signal Processing.
[77] Yonina C. Eldar,et al. Tradeoffs Between Convergence Speed and Reconstruction Accuracy in Inverse Problems , 2016, IEEE Transactions on Signal Processing.
[78] Stefano Soatto,et al. Deep relaxation: partial differential equations for optimizing deep neural networks , 2017, Research in the Mathematical Sciences.
[79] Stefano Soatto,et al. Information Dropout: Learning Optimal Representations Through Noisy Computation , 2016, IEEE Transactions on Pattern Analysis and Machine Intelligence.