Let's Agree to Agree: Neural Networks Share Classification Order on Real Datasets
暂无分享,去创建一个
[1] Andrew Zisserman,et al. Very Deep Convolutional Networks for Large-Scale Image Recognition , 2014, ICLR.
[2] Samy Bengio,et al. Understanding deep learning requires rethinking generalization , 2016, ICLR.
[3] Kilian Q. Weinberger,et al. Densely Connected Convolutional Networks , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[4] Timo Aila,et al. A Style-Based Generator Architecture for Generative Adversarial Networks , 2018, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[5] K. Pearson. Contributions to the Mathematical Theory of Evolution , 1894 .
[6] Geoffrey E. Hinton,et al. Learning representations by back-propagating errors , 1986, Nature.
[7] Ya Le,et al. Tiny ImageNet Visual Recognition Challenge , 2015 .
[8] Daphne Koller,et al. Self-Paced Learning for Latent Variable Models , 2010, NIPS.
[9] Roland Vollgraf,et al. Fashion-MNIST: a Novel Image Dataset for Benchmarking Machine Learning Algorithms , 2017, ArXiv.
[10] Lyle H. Ungar,et al. Machine Learning manuscript No. (will be inserted by the editor) Active Learning for Logistic Regression: , 2007 .
[11] Jeffrey Pennington,et al. GloVe: Global Vectors for Word Representation , 2014, EMNLP.
[12] Noam Slonim,et al. Learning to combine Grammatical Error Corrections , 2019, BEA@ACL.
[13] Andrea Vedaldi,et al. Understanding Image Representations by Measuring Their Equivariance and Equivalence , 2014, International Journal of Computer Vision.
[14] Jascha Sohl-Dickstein,et al. SVCCA: Singular Vector Canonical Correlation Analysis for Deep Learning Dynamics and Interpretability , 2017, NIPS.
[15] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[16] Samy Bengio,et al. Insights on representational similarity in neural networks with canonical correlation , 2018, NeurIPS.
[17] Yoshua Bengio,et al. Understanding intermediate layers using linear classifier probes , 2016, ICLR.
[18] Dumitru Erhan,et al. Going deeper with convolutions , 2014, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[19] Geoffrey E. Hinton,et al. ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.
[20] Yoshua Bengio,et al. Gradient-based learning applied to document recognition , 1998, Proc. IEEE.
[21] Tegan Maharaj,et al. Deep Nets Don't Learn via Memorization , 2017, ICLR.
[22] George Cybenko,et al. Approximation by superpositions of a sigmoidal function , 1989, Math. Control. Signals Syst..
[23] Hod Lipson,et al. Convergent Learning: Do different neural networks learn the same representations? , 2015, FE@NIPS.
[24] Daphna Weinshall,et al. On The Power of Curriculum Learning in Training Deep Networks , 2019, ICML.
[25] Li Fei-Fei,et al. ImageNet: A large-scale hierarchical image database , 2009, CVPR.
[26] Yoshua Bengio,et al. A Closer Look at Memorization in Deep Networks , 2017, ICML.
[27] Saman Ghili,et al. Tiny ImageNet Visual Recognition Challenge , 2014 .
[28] Haim Sompolinsky,et al. Separability and geometry of object manifolds in deep neural networks , 2019, Nature Communications.
[29] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[30] Abhinav Gupta,et al. Training Region-Based Object Detectors with Online Hard Example Mining , 2016, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[31] Leslie Pack Kaelbling,et al. Generalization in Deep Learning , 2017, ArXiv.
[32] Thomas B. Moeslund,et al. Long-Term Occupancy Analysis Using Graph-Based Optimisation in Thermal Imagery , 2013, 2013 IEEE Conference on Computer Vision and Pattern Recognition.
[33] Li Fei-Fei,et al. MentorNet: Learning Data-Driven Curriculum for Very Deep Neural Networks on Corrupted Labels , 2017, ICML.
[34] Trevor Hastie,et al. Multi-class AdaBoost ∗ , 2009 .
[35] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[36] Jürgen Schmidhuber,et al. Long Short-Term Memory , 1997, Neural Computation.
[37] Surya Ganguli,et al. A mathematical theory of semantic development in deep neural networks , 2018, Proceedings of the National Academy of Sciences.
[38] André F. T. Martins,et al. Marian: Fast Neural Machine Translation in C++ , 2018, ACL.
[39] Jason Weston,et al. Curriculum learning , 2009, ICML '09.
[40] Kurt Hornik,et al. Multilayer feedforward networks are universal approximators , 1989, Neural Networks.
[41] Antonio Torralba,et al. Recognizing indoor scenes , 2009, 2009 IEEE Conference on Computer Vision and Pattern Recognition.
[42] Alexei A. Efros,et al. The Unreasonable Effectiveness of Deep Features as a Perceptual Metric , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[43] Yoshua Bengio,et al. Understanding the difficulty of training deep feedforward neural networks , 2010, AISTATS.
[44] Yoshua Bengio,et al. Why Does Unsupervised Pre-training Help Deep Learning? , 2010, AISTATS.
[45] Omkar M. Parkhi,et al. VGGFace2: A Dataset for Recognising Faces across Pose and Age , 2017, 2018 13th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2018).
[46] Liwei Wang,et al. Towards Understanding Learning Representations: To What Extent Do Different Neural Networks Learn the Same Representation , 2018, NeurIPS.
[47] Hod Lipson,et al. Understanding Neural Networks Through Deep Visualization , 2015, ArXiv.
[48] Erkki Oja,et al. Principal components, minor components, and linear neural networks , 1992, Neural Networks.