暂无分享,去创建一个
[1] Nicolas Pinto,et al. Beyond simple features: A large-scale feature search approach to unconstrained face recognition , 2011, Face and Gesture 2011.
[2] P. Lancaster,et al. The theory of matrices : with applications , 1985 .
[3] Guigang Zhang,et al. Deep Learning , 2016, Int. J. Semantic Comput..
[4] Rob Fergus,et al. Stochastic Pooling for Regularization of Deep Convolutional Neural Networks , 2013, ICLR.
[5] Yan Wang,et al. A Powerful Generative Model Using Random Weights for the Deep Image Representation , 2016, NIPS.
[6] Andrew Zisserman,et al. Very Deep Convolutional Networks for Large-Scale Image Recognition , 2014, ICLR.
[7] Richard S. Sutton,et al. Online Learning with Random Representations , 1993, ICML.
[8] Geoffrey E. Hinton,et al. Reducing the Dimensionality of Data with Neural Networks , 2006, Science.
[9] B.M. Wilamowski,et al. Neural network architectures and learning algorithms , 2009, IEEE Industrial Electronics Magazine.
[10] Andrew R. Barron,et al. Universal approximation bounds for superpositions of a sigmoidal function , 1993, IEEE Trans. Inf. Theory.
[11] Guillermo Sapiro,et al. Deep Neural Networks with Random Gaussian Weights: A Universal Classification Strategy? , 2015, IEEE Transactions on Signal Processing.
[12] David D. Cox,et al. A High-Throughput Screening Approach to Discovering Good Forms of Biologically Inspired Visual Representation , 2009, PLoS Comput. Biol..
[13] Heikki Mannila,et al. Random projection in dimensionality reduction: applications to image and text data , 2001, KDD '01.
[14] Yann LeCun,et al. The Loss Surfaces of Multilayer Networks , 2014, AISTATS.
[15] David S. Broomhead,et al. Multivariable Functional Interpolation and Adaptive Networks , 1988, Complex Syst..
[16] James T. Kwok,et al. Objective functions for training new hidden units in constructive neural networks , 1997, IEEE Trans. Neural Networks.
[17] Yann LeCun,et al. What is the best multi-stage architecture for object recognition? , 2009, 2009 IEEE 12th International Conference on Computer Vision.
[18] Ivan Tyukin,et al. Approximation with random bases: Pro et Contra , 2015, Inf. Sci..
[19] Surya Ganguli,et al. Identifying and attacking the saddle point problem in high-dimensional non-convex optimization , 2014, NIPS.
[20] Bernard Widrow,et al. The No-Prop algorithm: A new learning algorithm for multilayer neural networks , 2013, Neural Networks.
[21] Kurt Hornik,et al. Multilayer feedforward networks are universal approximators , 1989, Neural Networks.
[22] Ohad Shamir,et al. The Power of Depth for Feedforward Neural Networks , 2015, COLT.
[23] Trevor Hastie,et al. Statistical Learning with Sparsity: The Lasso and Generalizations , 2015 .
[24] Yoshua Bengio,et al. Random Search for Hyper-Parameter Optimization , 2012, J. Mach. Learn. Res..
[25] Zhenghao Chen,et al. On Random Weights and Unsupervised Feature Learning , 2011, ICML.
[26] Mathieu Salzmann,et al. Learning the Number of Neurons in Deep Networks , 2016, NIPS.
[27] Leon A. Gatys,et al. Texture Synthesis Using Convolutional Neural Networks , 2015, NIPS.
[28] George Cybenko,et al. Approximation by superpositions of a sigmoidal function , 1992, Math. Control. Signals Syst..
[29] T. Blumensath,et al. Theory and Applications , 2011 .
[30] Pascal Vincent,et al. Representation Learning: A Review and New Perspectives , 2012, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[31] Nicolas Le Roux,et al. Representational Power of Restricted Boltzmann Machines and Deep Belief Networks , 2008, Neural Computation.
[32] Yee Whye Teh,et al. A Fast Learning Algorithm for Deep Belief Nets , 2006, Neural Computation.
[33] Andrew Y. Ng,et al. The Importance of Encoding Versus Training with Sparse Coding and Vector Quantization , 2011, ICML.
[34] Aditya Bhaskara,et al. Provable Bounds for Learning Some Deep Representations , 2013, ICML.
[35] Dianhui Wang,et al. Randomness in neural networks: an overview , 2017, WIREs Data Mining Knowl. Discov..
[36] Surya Ganguli,et al. Exact solutions to the nonlinear dynamics of learning in deep linear neural networks , 2013, ICLR.
[37] Dianhui Wang,et al. Stochastic Configuration Networks: Fundamentals and Algorithms , 2017, IEEE Transactions on Cybernetics.
[38] D. E. Rumelhart,et al. Learning internal representations by back-propagating errors , 1986 .
[39] R. Tibshirani. Regression Shrinkage and Selection via the Lasso , 1996 .
[40] Daniel Cownden,et al. Random feedback weights support learning in deep neural networks , 2014, ArXiv.
[41] Ming Li,et al. Insights into randomized algorithms for neural networks: Practical issues and common pitfalls , 2017, Inf. Sci..
[42] Y. Takefuji,et al. Functional-link net computing: theory, system architecture, and functionalities , 1992, Computer.
[43] Andrew Y. Ng,et al. Selecting Receptive Fields in Deep Networks , 2011, NIPS.
[44] Xin Yao,et al. A New Constructive Algorithm for Architectural and Functional Adaptation of Artificial Neural Networks , 2009, IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics).
[45] Jürgen Schmidhuber,et al. Deep learning in neural networks: An overview , 2014, Neural Networks.