暂无分享,去创建一个
[1] Ida G. Sprinkhuizen-Kuyper,et al. A local minimum for the 2-3-1 XOR network , 1999, IEEE Trans. Neural Networks.
[2] Ohad Shamir,et al. On the Quality of the Initial Basin in Overspecified Neural Networks , 2015, ICML.
[3] Joan Bruna,et al. Topology and Geometry of Half-Rectified Network Optimization , 2016, ICLR.
[4] Cristian Sminchisescu,et al. Building Roadmaps of Minima and Transitions in Visual Models , 2004, International Journal of Computer Vision.
[5] Kenji Kawaguchi,et al. Deep Learning without Poor Local Minima , 2016, NIPS.
[6] Matthias Hein,et al. The Loss Surface of Deep and Wide Neural Networks , 2017, ICML.
[7] Le Song,et al. Diverse Neural Network Learns True Target Functions , 2016, AISTATS.
[8] J. Slawny,et al. Back propagation fails to separate where perceptrons succeed , 1989 .
[9] Lorenzo Rosasco,et al. Why and when can deep-but not shallow-networks avoid the curse of dimensionality: A review , 2016, International Journal of Automation and Computing.
[10] Peter Auer,et al. Exponentially many local minima for single neurons , 1995, NIPS.
[11] Yann LeCun,et al. The Loss Surfaces of Multilayer Networks , 2014, AISTATS.
[12] Tomaso A. Poggio,et al. Theory II: Landscape of the Empirical Risk in Deep Learning , 2017, ArXiv.
[13] YoungJu Choie,et al. Local minima and back propagation , 1991, IJCNN-91-Seattle International Joint Conference on Neural Networks.
[14] Elad Hoffer,et al. Exponentially vanishing sub-optimal local minima in multilayer neural networks , 2017, ICLR.
[15] Eiji Mizutani,et al. An analysis on negative curvature induced by singularity in multi-layer neural-network learning , 2010, NIPS.
[16] Geoffrey E. Hinton,et al. Learning internal representations by error propagation , 1986 .
[17] Shun-ichi Amari,et al. Natural Gradient Works Efficiently in Learning , 1998, Neural Computation.
[18] Leonard G. C. Hamey,et al. XOR has no local minima: A case study in neural network error surface analysis , 1998, Neural Networks.
[19] Suvrit Sra,et al. Global optimality conditions for deep neural networks , 2017, ICLR.
[20] Tohru Nitta,et al. Resolution of Singularities Introduced by Hierarchical Structure in Deep Neural Networks , 2017, IEEE Transactions on Neural Networks and Learning Systems.
[21] Kurt Hornik,et al. Neural networks and principal component analysis: Learning from examples without local minima , 1989, Neural Networks.
[22] L. F. A. Wessels,et al. The Physical Correlates of Local Minima , 1990 .
[23] A. Bray,et al. Statistics of critical points of Gaussian fields on large-dimensional spaces. , 2006, Physical review letters.
[24] Etienne Barnard,et al. Avoiding false local minima by proper initialization of connections , 1992, IEEE Trans. Neural Networks.
[25] Le Song,et al. Diversity Leads to Generalization in Neural Networks , 2016, ArXiv.
[26] Surya Ganguli,et al. Identifying and attacking the saddle point problem in high-dimensional non-convex optimization , 2014, NIPS.
[27] H. Tamura,et al. An improved backpropagation algorithm to avoid the local minima problem , 2004, Neurocomputing.
[28] Joan Bruna,et al. Mathematics of Deep Learning , 2017, ArXiv.
[29] Shun-ichi Amari,et al. Dynamics of Learning Near Singularities in Layered Networks , 2008, Neural Computation.
[30] Haihao Lu,et al. Depth Creates No Bad Local Minima , 2017, ArXiv.
[31] Samy Bengio,et al. Understanding deep learning requires rethinking generalization , 2016, ICLR.
[32] Ronald L. Rivest,et al. Training a 3-node neural network is NP-complete , 1988, COLT '88.
[33] Kenji Fukumizu,et al. Local minima and plateaus in hierarchical structures of multilayer perceptrons , 2000, Neural Networks.
[34] Daniel Soudry,et al. No bad local minima: Data independent training error guarantees for multilayer neural networks , 2016, ArXiv.
[35] Razvan Pascanu,et al. Local minima in training of neural networks , 2016, 1611.06310.
[36] Yann LeCun,et al. Explorations on high dimensional landscapes , 2014, ICLR.
[37] Alberto Tesi,et al. On the Problem of Local Minima in Backpropagation , 1992, IEEE Trans. Pattern Anal. Mach. Intell..
[38] René Vidal,et al. Global Optimality in Neural Network Training , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).