暂无分享,去创建一个
Roger B. Grosse | David Duvenaud | Paul Vicol | Jonathan Lorraine | Matthew MacKay | D. Duvenaud | M. Mackay | Paul Vicol | Jonathan Lorraine | R. Grosse
[1] Eugene L. Allgower,et al. Numerical continuation methods - an introduction , 1990, Springer series in computational mathematics.
[2] J. Urgen Schmidhuber. Learning to Control Fast-weight Memories: an Alternative to Dynamic Recurrent Networks , 1991 .
[3] Pierre Hansen,et al. New Branch-and-Bound Rules for Linear Bilevel Programming , 1989, SIAM J. Sci. Comput..
[4] Robert Gibbons,et al. A primer in game theory , 1992 .
[5] R. J. Williams,et al. Simple Statistical Gradient-Following Algorithms for Connectionist Reinforcement Learning , 2004, Machine Learning.
[6] Beatrice Santorini,et al. Building a Large Annotated Corpus of English: The Penn Treebank , 1993, CL.
[7] Jürgen Schmidhuber,et al. A ‘Self-Referential’ Weight Matrix , 1993 .
[8] L. N. Vicente,et al. Descent approaches for quadratic bilevel programming , 1994 .
[9] J. Larsen,et al. Design and regularization of neural networks: the optimal use of a validation set , 1996, Neural Networks for Signal Processing VI. Proceedings of the 1996 IEEE Signal Processing Society Workshop.
[10] Jürgen Schmidhuber,et al. Long Short-Term Memory , 1997, Neural Computation.
[11] Yoshua Bengio,et al. Gradient-based learning applied to document recognition , 1998, Proc. IEEE.
[12] Jonathan F. Bard,et al. Practical Bilevel Optimization: Algorithms and Applications , 1998 .
[13] Trevor Hastie,et al. The Elements of Statistical Learning , 2001 .
[14] Sayan Mukherjee,et al. Choosing Multiple Parameters for Support Vector Machines , 2002, Machine Learning.
[15] Yurii Nesterov,et al. Introductory Lectures on Convex Optimization - A Basic Course , 2014, Applied Optimization.
[16] Patrice Marcotte,et al. A Trust-Region Method for Nonlinear Bilevel Programming: Algorithm and Computational Experience , 2005, Comput. Optim. Appl..
[17] Matthias W. Seeger. Cross-Validation Optimization for Large Scale Hierarchical Classification Kernel Methods , 2006, NIPS.
[18] Patrice Marcotte,et al. An overview of bilevel optimization , 2007, Ann. Oper. Res..
[19] Chuan-Sheng Foo,et al. Efficient multiple hyperparameter learning for log-linear models , 2007, NIPS.
[20] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[21] Jason Weston,et al. Curriculum learning , 2009, ICML '09.
[22] Heinrich von Stackelberg. Market Structure and Equilibrium , 2010 .
[23] Yoshua Bengio,et al. Algorithms for Hyper-Parameter Optimization , 2011, NIPS.
[24] Kevin Leyton-Brown,et al. Sequential Model-Based Optimization for General Algorithm Configuration , 2011, LION.
[25] Yoshua Bengio,et al. Random Search for Hyper-Parameter Optimization , 2012, J. Mach. Learn. Res..
[26] Jasper Snoek,et al. Practical Bayesian Optimization of Machine Learning Algorithms , 2012, NIPS.
[27] Justin Domke,et al. Generic Methods for Optimization-Based Modeling , 2012, AISTATS.
[28] Geoffrey E. Hinton,et al. ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.
[29] David Barber,et al. Variational Optimization , 2012, ArXiv.
[30] Yann LeCun,et al. Regularization of Neural Networks using DropConnect , 2013, ICML.
[31] Karl Kunisch,et al. A Bilevel Optimization Approach for Parameter Learning in Variational Models , 2013, SIAM J. Imaging Sci..
[32] Misha Denil,et al. Predicting Parameters in Deep Learning , 2014 .
[33] Kalyanmoy Deb,et al. Efficient Evolutionary Algorithm for Single-Objective Bilevel Optimization , 2013, ArXiv.
[34] Nitish Srivastava,et al. Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..
[35] Jasper Snoek,et al. Freeze-Thaw Bayesian Optimization , 2014, ArXiv.
[36] Yoshua Bengio,et al. Generative Adversarial Nets , 2014, NIPS.
[37] Wojciech Zaremba,et al. Recurrent Neural Network Regularization , 2014, ArXiv.
[38] Carola-Bibiane Schönlieb,et al. Bilevel approaches for learning of variational imaging models , 2015, ArXiv.
[39] Le Song,et al. Deep Fried Convnets , 2014, 2015 IEEE International Conference on Computer Vision (ICCV).
[40] Prabhat,et al. Scalable Bayesian Optimization Using Deep Neural Networks , 2015, ICML.
[41] Julien Cornebise,et al. Weight Uncertainty in Neural Networks , 2015, ArXiv.
[42] Ryan P. Adams,et al. Gradient-based Hyperparameter Optimization through Reversible Learning , 2015, ICML.
[43] Zoubin Ghahramani,et al. A Theoretically Grounded Application of Dropout in Recurrent Neural Networks , 2015, NIPS.
[44] Fabian Pedregosa,et al. Hyperparameter optimization with approximate gradient , 2016, ICML.
[45] Ameet Talwalkar,et al. Non-stochastic Best Arm Identification and Hyperparameter Optimization , 2015, AISTATS.
[46] Tapani Raiko,et al. Scalable Gradient-Based Tuning of Continuous Regularization Hyperparameters , 2015, ICML.
[47] Sebastian Nowozin,et al. The Numerics of GANs , 2017, NIPS.
[48] David Pfau,et al. Unrolled Generative Adversarial Networks , 2016, ICLR.
[49] Quoc V. Le,et al. Neural Architecture Search with Reinforcement Learning , 2016, ICLR.
[50] Aaron Klein,et al. Learning Curve Prediction with Bayesian Neural Networks , 2016, ICLR.
[51] Christopher Burgess,et al. beta-VAE: Learning Basic Visual Concepts with a Constrained Variational Framework , 2016, ICLR 2016.
[52] Max Jaderberg,et al. Population Based Training of Neural Networks , 2017, ArXiv.
[53] Luca Antiga,et al. Automatic differentiation in PyTorch , 2017 .
[54] Ameet Talwalkar,et al. Hyperband: Bandit-Based Configuration Evaluation for Hyperparameter Optimization , 2016, ICLR.
[55] Graham W. Taylor,et al. Improved Regularization of Convolutional Neural Networks with Cutout , 2017, ArXiv.
[56] Paolo Frasconi,et al. Bilevel Programming for Hyperparameter Optimization and Meta-Learning , 2018, ICML.
[57] Didrik Nielsen,et al. Fast and Scalable Bayesian Deep Learning by Weight-Perturbation in Adam , 2018, ICML.
[58] Richard Socher,et al. Regularizing and Optimizing LSTM Language Models , 2017, ICLR.
[59] Dustin Tran,et al. Flipout: Efficient Pseudo-Independent Weight Perturbations on Mini-Batches , 2018, ICLR.
[60] David Duvenaud,et al. Stochastic Hyperparameter Optimization through Hypernetworks , 2018, ArXiv.
[61] Theodore Lim,et al. SMASH: One-Shot Model Architecture Search through HyperNetworks , 2017, ICLR.