暂无分享,去创建一个
[1] Jian Sun,et al. Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification , 2015, 2015 IEEE International Conference on Computer Vision (ICCV).
[2] Han Liu,et al. Nonparametrically Learning Activation Functions in Deep Neural Nets , 2016 .
[3] Geoffrey E. Hinton,et al. Reducing the Dimensionality of Data with Neural Networks , 2006, Science.
[4] Ron Kohavi,et al. Scaling Up the Accuracy of Naive-Bayes Classifiers: A Decision-Tree Hybrid , 1996, KDD.
[5] Andrew L. Maas. Rectifier Nonlinearities Improve Neural Network Acoustic Models , 2013 .
[6] S. P. Smith. Differentiation of the Cholesky Algorithm , 1995 .
[7] Patrick van der Smagt,et al. Automatic Differentiation for Tensor Algebras , 2017, ArXiv.
[8] Geoffrey E. Hinton,et al. Rectified Linear Units Improve Restricted Boltzmann Machines , 2010, ICML.
[9] Carl E. Rasmussen,et al. In Advances in Neural Information Processing Systems , 2011 .
[10] Max Welling,et al. Auto-Encoding Variational Bayes , 2013, ICLR.
[11] Simone Scardapane,et al. Kafnets: kernel-based non-parametric activation functions for neural networks , 2017, Neural Networks.
[12] Neil D. Lawrence,et al. Deep Gaussian Processes , 2012, AISTATS.
[13] Zhaoran Wang,et al. Nonparametrically Learning Activation Functions in Deep Neural Nets , 2017 .
[14] Li Fei-Fei,et al. ImageNet: A large-scale hierarchical image database , 2009, CVPR.
[15] Patrick van der Smagt,et al. A Neural Transfer Function for a Smooth and Differentiable Transition Between Additive and Multiplicative Interactions , 2015, ArXiv.
[16] Carl E. Rasmussen,et al. A Unifying View of Sparse Approximate Gaussian Process Regression , 2005, J. Mach. Learn. Res..
[17] Sepp Hochreiter,et al. Fast and Accurate Deep Network Learning by Exponential Linear Units (ELUs) , 2015, ICLR.
[18] Nitish Srivastava,et al. Improving neural networks by preventing co-adaptation of feature detectors , 2012, ArXiv.
[19] Geoffrey E. Hinton,et al. Deep Boltzmann Machines , 2009, AISTATS.
[20] Pierre Baldi,et al. Learning Activation Functions to Improve Deep Neural Networks , 2014, ICLR.
[21] Agathe Girard,et al. Prediction at an Uncertain Input for Gaussian Processes and Relevance Vector Machines Application to Multiple-Step Ahead Time-Series Forecasting , 2002 .
[22] Geoffrey E. Hinton,et al. ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.
[23] John Salvatier,et al. Theano: A Python framework for fast computation of mathematical expressions , 2016, ArXiv.
[24] Sepp Hochreiter,et al. Self-Normalizing Neural Networks , 2017, NIPS.
[25] Yoshua Bengio,et al. Understanding the difficulty of training deep feedforward neural networks , 2010, AISTATS.
[26] John B. Shoven,et al. I , Edinburgh Medical and Surgical Journal.
[27] Tianqi Chen,et al. Empirical Evaluation of Rectified Activations in Convolutional Network , 2015, ArXiv.
[28] Martín Abadi,et al. TensorFlow: Large-Scale Machine Learning on Heterogeneous Distributed Systems , 2016, ArXiv.
[29] S. Julier,et al. A General Method for Approximating Nonlinear Transformations of Probability Distributions , 1996 .
[30] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[31] Nitish Srivastava,et al. Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..
[32] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[33] Yann LeCun,et al. The mnist database of handwritten digits , 2005 .
[34] Christopher D. Manning,et al. Fast dropout training , 2013, ICML.
[35] Marc Peter Deisenroth,et al. Doubly Stochastic Variational Inference for Deep Gaussian Processes , 2017, NIPS.
[36] Michalis K. Titsias,et al. Variational Learning of Inducing Variables in Sparse Gaussian Processes , 2009, AISTATS.