暂无分享,去创建一个
[1] Amos Storkey,et al. Meta-Learning in Neural Networks: A Survey , 2020, IEEE transactions on pattern analysis and machine intelligence.
[2] Artem Molchanov,et al. Generalized Inner Loop Meta-Learning , 2019, ArXiv.
[3] Joan Bruna,et al. Intriguing properties of neural networks , 2013, ICLR.
[4] Amos Storkey,et al. Learning to Learn By Self-Critique , 2019, NeurIPS.
[5] C A Nelson,et al. Learning to Learn , 2017, Encyclopedia of Machine Learning and Data Mining.
[6] Jonathon Shlens,et al. A Learned Representation For Artistic Style , 2016, ICLR.
[7] Jeremy Kepner,et al. Interactive Supercomputing on 40,000 Cores for Machine Learning and Data Analysis , 2018, 2018 IEEE High Performance extreme Computing Conference (HPEC).
[8] Vladimir N. Vapnik,et al. The Nature of Statistical Learning Theory , 2000, Statistics for Engineering and Information Science.
[9] Sebastian Nowozin,et al. Fast and Flexible Multi-Task Classification Using Conditional Neural Adaptive Processes , 2019, NeurIPS.
[10] Leslie Pack Kaelbling,et al. Neural Relational Inference with Fast Modular Meta-learning , 2019, NeurIPS.
[11] Fu Jie Huang,et al. A Tutorial on Energy-Based Learning , 2006 .
[12] Sergey Levine,et al. Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks , 2017, ICML.
[13] Trevor Darrell,et al. Dynamic Scale Inference by Entropy Minimization , 2019, ArXiv.
[14] Shin Ishii,et al. Virtual Adversarial Training: A Regularization Method for Supervised and Semi-Supervised Learning , 2017, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[15] Kaiming He,et al. Momentum Contrast for Unsupervised Visual Representation Learning , 2019, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[16] Fabio Roli,et al. Evasion Attacks against Machine Learning at Test Time , 2013, ECML/PKDD.
[17] Geoffrey E. Hinton. Training Products of Experts by Minimizing Contrastive Divergence , 2002, Neural Computation.
[18] Jason Weston,et al. Transductive Inference for Estimating Values of Functions , 1999, NIPS.
[19] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[20] Geoffrey E. Hinton,et al. A Learning Algorithm for Boltzmann Machines , 1985, Cogn. Sci..
[21] Katja Hofmann,et al. CAML: Fast Context Adaptation via Meta-Learning , 2018, ArXiv.
[22] Natalia Gimelshein,et al. PyTorch: An Imperative Style, High-Performance Deep Learning Library , 2019, NeurIPS.
[23] Oriol Vinyals,et al. Representation Learning with Contrastive Predictive Coding , 2018, ArXiv.
[24] Aleksander Madry,et al. Adversarial Examples Are Not Bugs, They Are Features , 2019, NeurIPS.
[25] Eunho Yang,et al. Learning to Propagate Labels: Transductive Propagation Network for Few-Shot Learning , 2018, ICLR.
[26] Jason Yosinski,et al. Hamiltonian Neural Networks , 2019, NeurIPS.
[27] Sergey Levine,et al. Efficient Off-Policy Meta-Reinforcement Learning via Probabilistic Context Variables , 2019, ICML.
[28] Andreas Krause,et al. Differentiable Submodular Maximization , 2018, IJCAI.
[29] Sergey Ioffe,et al. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift , 2015, ICML.
[30] J. Zico Kolter,et al. OptNet: Differentiable Optimization as a Layer in Neural Networks , 2017, ICML.
[31] Katja Hofmann,et al. Fast Context Adaptation via Meta-Learning , 2018, ICML.
[32] Russ Tedrake,et al. The Surprising Effectiveness of Linear Models for Visual Foresight in Object Pile Manipulation , 2020, WAFR.
[33] Vladimir Vapnik,et al. The Nature of Statistical Learning , 1995 .
[34] Samy Bengio,et al. On the search for new learning rules for ANNs , 1995, Neural Processing Letters.
[35] Razvan Pascanu,et al. Meta-Learning with Warped Gradient Descent , 2020, ICLR.
[36] Peter L. Bartlett,et al. Rademacher and Gaussian Complexities: Risk Bounds and Structural Results , 2003, J. Mach. Learn. Res..
[37] Aleksander Madry,et al. Towards Deep Learning Models Resistant to Adversarial Attacks , 2017, ICLR.
[38] J. Zico Kolter,et al. Certified Adversarial Robustness via Randomized Smoothing , 2019, ICML.
[39] Li Fei-Fei,et al. ImageNet: A large-scale hierarchical image database , 2009, CVPR.
[40] Igor Mordatch,et al. Implicit Generation and Generalization with Energy Based Models , 2018 .
[41] Léon Bottou,et al. Local Learning Algorithms , 1992, Neural Computation.
[42] Aaron C. Courville,et al. FiLM: Visual Reasoning with a General Conditioning Layer , 2017, AAAI.
[43] B. Mityagin. The Zero Set of a Real Analytic Function , 2015, Mathematical Notes.
[44] Pradeep Ravikumar,et al. MACER: Attack-free and Scalable Robust Training via Maximizing Certified Radius , 2020, ICLR.
[45] Geoffrey E. Hinton,et al. A Simple Framework for Contrastive Learning of Visual Representations , 2020, ICML.
[46] Stefano Soatto,et al. A Baseline for Few-Shot Image Classification , 2019, ICLR.
[47] Alexei A. Efros,et al. Test-Time Training for Out-of-Distribution Generalization , 2019, ArXiv.
[48] Razvan Pascanu,et al. Learning to Navigate in Complex Environments , 2016, ICLR.
[49] Joan Bruna,et al. Few-Shot Learning with Graph Neural Networks , 2017, ICLR.
[50] Chelsea Finn,et al. Learning to Learn with Gradients , 2018 .
[51] Alexei A. Efros,et al. Test-Time Training with Self-Supervision for Generalization under Distribution Shifts , 2019, ICML.
[52] Yann LeCun,et al. Dimensionality Reduction by Learning an Invariant Mapping , 2006, 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06).
[53] Yoshua Bengio,et al. Equilibrium Propagation: Bridging the Gap between Energy-Based Models and Backpropagation , 2016, Front. Comput. Neurosci..
[54] Greg Yang,et al. Provably Robust Deep Learning via Adversarially Trained Smoothed Classifiers , 2019, NeurIPS.