Meta Dropout: Learning to Perturb Latent Features for Generalization
暂无分享,去创建一个
[1] Stefano Soatto,et al. Entropy-SGD: biasing gradient descent into wide valleys , 2016, ICLR.
[2] Gregory R. Koch,et al. Siamese Neural Networks for One-Shot Image Recognition , 2015 .
[3] Alex Beatson,et al. Amortized Bayesian Meta-Learning , 2018, ICLR.
[4] Max Welling,et al. Variational Dropout and the Local Reparameterization Trick , 2015, NIPS 2015.
[5] Bernt Schiele,et al. Disentangling Adversarial Robustness and Generalization , 2018, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[6] Rich Caruana,et al. Multitask Learning , 1998, Encyclopedia of Machine Learning and Data Mining.
[7] Aleksander Madry,et al. Towards Deep Learning Models Resistant to Adversarial Attacks , 2017, ICLR.
[8] Aleksander Madry,et al. How Does Batch Normalization Help Optimization? (No, It Is Not About Internal Covariate Shift) , 2018, NeurIPS.
[9] Swami Sankaranarayanan,et al. MetaReg: Towards Domain Generalization using Meta-Regularization , 2018, NeurIPS.
[10] Sebastian Thrun,et al. Learning to Learn , 1998, Springer US.
[11] Jonathon Shlens,et al. Explaining and Harnessing Adversarial Examples , 2014, ICLR.
[12] Sergey Levine,et al. Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks , 2017, ICML.
[13] Alex Kendall,et al. Concrete Dropout , 2017, NIPS.
[14] Ioannis Mitliagkas,et al. Manifold Mixup: Learning Better Representations by Interpolating Hidden States , 2018, 1806.05236.
[15] Brendan J. Frey,et al. Adaptive dropout for training deep neural networks , 2013, NIPS.
[16] Zoubin Ghahramani,et al. Dropout as a Bayesian Approximation: Representing Model Uncertainty in Deep Learning , 2015, ICML.
[17] Alexander A. Alemi,et al. Deep Variational Information Bottleneck , 2017, ICLR.
[18] Dan Boneh,et al. Adversarial Training and Robustness for Multiple Perturbations , 2019, NeurIPS.
[19] Honglak Lee,et al. Learning Structured Output Representation using Deep Conditional Generative Models , 2015, NIPS.
[20] Max Welling,et al. Auto-Encoding Variational Bayes , 2013, ICLR.
[21] Quoc V. Le,et al. DropBlock: A regularization method for convolutional networks , 2018, NeurIPS.
[22] Hang Li,et al. Meta-SGD: Learning to Learn Quickly for Few Shot Learning , 2017, ArXiv.
[23] Alexandre Lacoste,et al. TADAM: Task dependent adaptive metric for improved few-shot learning , 2018, NeurIPS.
[24] Nathan Srebro,et al. Exploring Generalization in Deep Learning , 2017, NIPS.
[25] Joshua Achiam,et al. On First-Order Meta-Learning Algorithms , 2018, ArXiv.
[26] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[27] Nitish Srivastava,et al. Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..
[28] Hongyi Zhang,et al. mixup: Beyond Empirical Risk Minimization , 2017, ICLR.
[29] Hugo Larochelle,et al. Optimization as a Model for Few-Shot Learning , 2016, ICLR.
[30] Naftali Tishby,et al. Deep learning and the information bottleneck principle , 2015, 2015 IEEE Information Theory Workshop (ITW).
[31] Subhransu Maji,et al. Meta-Learning With Differentiable Convex Optimization , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[32] Geoffrey E. Hinton,et al. Regularizing Neural Networks by Penalizing Confident Output Distributions , 2017, ICLR.
[33] Sergey Levine,et al. Probabilistic Model-Agnostic Meta-Learning , 2018, NeurIPS.
[34] Li Fei-Fei,et al. ImageNet: A large-scale hierarchical image database , 2009, CVPR.
[35] Richard S. Zemel,et al. Prototypical Networks for Few-shot Learning , 2017, NIPS.
[36] Oriol Vinyals,et al. Matching Networks for One Shot Learning , 2016, NIPS.
[37] Yoshua Bengio,et al. Show, Attend and Tell: Neural Image Caption Generation with Visual Attention , 2015, ICML.
[38] Naftali Tishby,et al. The information bottleneck method , 2000, ArXiv.
[39] Jorge Nocedal,et al. On Large-Batch Training for Deep Learning: Generalization Gap and Sharp Minima , 2016, ICLR.
[40] Seungjin Choi,et al. Gradient-Based Meta-Learning with Learned Layerwise Metric and Subspace , 2018, ICML.
[41] Sergey Ioffe,et al. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift , 2015, ICML.
[42] Katja Hofmann,et al. Fast Context Adaptation via Meta-Learning , 2018, ICML.
[43] Christopher D. Manning,et al. Fast dropout training , 2013, ICML.
[44] Razvan Pascanu,et al. Meta-Learning with Latent Embedding Optimization , 2018, ICLR.
[45] Stefano Soatto,et al. Information Dropout: Learning Optimal Representations Through Noisy Computation , 2016, IEEE Transactions on Pattern Analysis and Machine Intelligence.