暂无分享,去创建一个
[1] Hongyi Zhang,et al. mixup: Beyond Empirical Risk Minimization , 2017, ICLR.
[2] Takashi Matsubara,et al. Data Augmentation Using Random Image Cropping and Patching for Deep CNNs , 2018, IEEE Transactions on Circuits and Systems for Video Technology.
[3] Kai A. Krueger,et al. Flexible shaping: How learning in small steps helps , 2009, Cognition.
[4] Léon Bottou,et al. Large-Scale Machine Learning with Stochastic Gradient Descent , 2010, COMPSTAT.
[5] Nikos Komodakis,et al. Wide Residual Networks , 2016, BMVC.
[6] Sergey Ioffe,et al. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift , 2015, ICML.
[7] Luis Perez,et al. The Effectiveness of Data Augmentation in Image Classification using Deep Learning , 2017, ArXiv.
[8] Satoshi Oyama,et al. Why Does Large Batch Training Result in Poor Generalization? A Comprehensive Explanation and a Better Strategy from the Viewpoint of Stochastic Optimization , 2018, Neural Computation.
[9] Daphne Koller,et al. Self-Paced Learning for Latent Variable Models , 2010, NIPS.
[10] Jürgen Schmidhuber,et al. Long Short-Term Memory , 1997, Neural Computation.
[11] Danna Zhou,et al. d. , 1840, Microbial pathogenesis.
[12] J. Elman. Learning and development in neural networks: the importance of starting small , 1993, Cognition.
[13] Jonghyun Choi,et al. ScreenerNet: Learning Self-Paced Curriculum for Deep Neural Networks , 2018 .
[14] Jason Weston,et al. Curriculum learning , 2009, ICML '09.
[15] Sepp Hochreiter,et al. Learning to Learn Using Gradient Descent , 2001, ICANN.
[16] Douglas L. T. Rohde,et al. Language acquisition in the absence of explicit negative evidence: how important is starting small? , 1999, Cognition.
[17] Sanjeev Khudanpur,et al. Audio augmentation for speech recognition , 2015, INTERSPEECH.
[18] Samy Bengio,et al. Understanding deep learning requires rethinking generalization , 2016, ICLR.
[19] Li Fei-Fei,et al. ImageNet: A large-scale hierarchical image database , 2009, CVPR.
[20] Taesup Kim,et al. Fast AutoAugment , 2019, NeurIPS.
[21] V. Cerný. Thermodynamical approach to the traveling salesman problem: An efficient simulation algorithm , 1985 .
[22] Andrew Y. Ng,et al. Reading Digits in Natural Images with Unsupervised Feature Learning , 2011 .
[23] Maoguo Gong,et al. Self-paced Convolutional Neural Networks , 2017, IJCAI.
[24] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[25] Mohamad Ivan Fanany,et al. Simulated Annealing Algorithm for Deep Learning , 2015 .
[26] C. D. Gelatt,et al. Optimization by Simulated Annealing , 1983, Science.
[27] Louis-Philippe Morency,et al. Visualizing and Understanding Curriculum Learning for Long Short-Term Memory Networks , 2016, ArXiv.
[28] Quoc V. Le,et al. Randaugment: Practical automated data augmentation with a reduced search space , 2019, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW).
[29] Roland Vollgraf,et al. Fashion-MNIST: a Novel Image Dataset for Benchmarking Machine Learning Algorithms , 2017, ArXiv.
[30] Yoshua Bengio,et al. Deep Sparse Rectifier Neural Networks , 2011, AISTATS.
[31] Satoshi Oyama,et al. Effective neural network training with adaptive learning rate based on training loss , 2018, Neural Networks.
[32] Loris Nanni,et al. General Purpose (GenP) Bioimage Ensemble of Handcrafted and Learned Features with Data Augmentation , 2019, ArXiv.
[33] Yoshua Bengio,et al. Gradient-based learning applied to document recognition , 1998, Proc. IEEE.
[34] Qi Tian,et al. Data Augmentation Revisited: Rethinking the Distribution Gap between Clean and Augmented Data , 2019, ArXiv.
[35] Kurt Keutzer,et al. Hessian-based Analysis of Large Batch Training and Robustness to Adversaries , 2018, NeurIPS.
[36] Graham W. Taylor,et al. Improved Regularization of Convolutional Neural Networks with Cutout , 2017, ArXiv.
[37] Honglak Lee,et al. An Analysis of Single-Layer Networks in Unsupervised Feature Learning , 2011, AISTATS.
[38] Elad Hoffer,et al. Train longer, generalize better: closing the generalization gap in large batch training of neural networks , 2017, NIPS.
[39] Valentin I. Spitkovsky,et al. Baby Steps: How “Less is More” in Unsupervised Dependency Parsing , 2009 .
[40] Marcin Andrychowicz,et al. Learning to learn by gradient descent by gradient descent , 2016, NIPS.
[41] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[42] Yi Yang,et al. Random Erasing Data Augmentation , 2017, AAAI.
[43] Quoc V. Le,et al. AutoAugment: Learning Augmentation Strategies From Data , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[44] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.