Manifold Mixup: Better Representations by Interpolating Hidden States
暂无分享,去创建一个
Ioannis Mitliagkas | Yoshua Bengio | David Lopez-Paz | Christopher Beckham | Alex Lamb | Vikas Verma | Amir Najafi | Yoshua Bengio | Alex Lamb | Ioannis Mitliagkas | Vikas Verma | Christopher Beckham | David Lopez-Paz | Amir Najafi | Amir Najafi
[1] Jeffrey Dean,et al. Efficient Estimation of Word Representations in Vector Space , 2013, ICLR.
[2] Koby Crammer,et al. A theory of learning from different domains , 2010, Machine Learning.
[3] P. Bartlett,et al. Generalization Performance of Support Vector Machines and Other Pattern Classifiers , 1999 .
[4] Hongyu Guo,et al. Aggregated Learning: A Vector Quantization Approach to Learning with Neural Networks , 2018, ArXiv.
[5] Harri Valpola,et al. Weight-averaged consistency targets improve semi-supervised deep learning results , 2017, ArXiv.
[6] Hongyi Zhang,et al. mixup: Beyond Empirical Risk Minimization , 2017, ICLR.
[7] Tatsuya Harada,et al. Between-Class Learning for Image Classification , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[8] Colin Raffel,et al. Realistic Evaluation of Deep Semi-Supervised Learning Algorithms , 2018, NeurIPS.
[9] Aleksander Madry,et al. Towards Deep Learning Models Resistant to Adversarial Attacks , 2017, ICLR.
[10] Jonathon Shlens,et al. Explaining and Harnessing Adversarial Examples , 2014, ICLR.
[11] Sergey Ioffe,et al. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift , 2015, ICML.
[12] Naftali Tishby,et al. Opening the Black Box of Deep Neural Networks via Information , 2017, ArXiv.
[13] Rob Fergus,et al. Visualizing and Understanding Convolutional Networks , 2013, ECCV.
[14] Shin Ishii,et al. Virtual Adversarial Training: A Regularization Method for Supervised and Semi-Supervised Learning , 2017, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[15] Graham W. Taylor,et al. Improved Regularization of Convolutional Neural Networks with Cutout , 2017, ArXiv.
[16] Guigang Zhang,et al. Deep Learning , 2016, Int. J. Semantic Comput..
[17] Wojciech Zaremba,et al. Improved Techniques for Training GANs , 2016, NIPS.
[18] Sergey Levine,et al. InfoBot: Transfer and Exploration via the Information Bottleneck , 2019, ICLR.
[19] George Cybenko,et al. Approximation by superpositions of a sigmoidal function , 1989, Math. Control. Signals Syst..
[20] Nello Cristianini,et al. Machine Learning and Knowledge Discovery in Databases (ECML PKDD) , 2010 .
[21] David A. Wagner,et al. Obfuscated Gradients Give a False Sense of Security: Circumventing Defenses to Adversarial Examples , 2018, ICML.
[22] Aaron C. Courville,et al. MINE: Mutual Information Neural Estimation , 2018, ArXiv.
[23] Hongyu Guo,et al. MixUp as Locally Linear Out-Of-Manifold Regularization , 2018, AAAI.
[24] Léon Bottou,et al. Wasserstein Generative Adversarial Networks , 2017, ICML.
[25] Yuichi Yoshida,et al. Spectral Normalization for Generative Adversarial Networks , 2018, ICLR.
[26] Zhen Li,et al. Deep Learning with Data Dependent Implicit Activation Function , 2018, 1802.00168.
[27] Alexander A. Alemi,et al. Deep Variational Information Bottleneck , 2017, ICLR.
[28] Nitish Srivastava,et al. Improving neural networks by preventing co-adaptation of feature detectors , 2012, ArXiv.
[29] Stefano Soatto,et al. Information Dropout: Learning Optimal Representations Through Noisy Computation , 2016, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[30] Yoshua Bengio,et al. Difference Target Propagation , 2014, ECML/PKDD.
[31] Aaron C. Courville,et al. Improved Training of Wasserstein GANs , 2017, NIPS.
[32] Kyunghyun Cho,et al. Retrieval-Augmented Convolutional Neural Networks for Improved Robustness against Adversarial Examples , 2018, ArXiv.
[33] Yoshua Bengio,et al. Equilibrium Propagation: Bridging the Gap between Energy-Based Models and Backpropagation , 2016, Front. Comput. Neurosci..
[34] Geoffrey E. Hinton,et al. Assessing the Scalability of Biologically-Motivated Deep Learning Algorithms and Architectures , 2018, NeurIPS.
[35] Jian Sun,et al. Identity Mappings in Deep Residual Networks , 2016, ECCV.
[36] Naftali Tishby,et al. Deep learning and the information bottleneck principle , 2015, 2015 IEEE Information Theory Workshop (ITW).
[37] Joan Bruna,et al. Intriguing properties of neural networks , 2013, ICLR.
[38] Rafal Bogacz,et al. An Approximation of the Error Backpropagation Algorithm in a Predictive Coding Network with Local Hebbian Synaptic Plasticity , 2017, Neural Computation.
[39] John Shawe-Taylor,et al. Generalization Performance of Support Vector Machines and Other Pattern Classifiers , 1999 .