暂无分享,去创建一个
[1] Robert A. Jacobs,et al. Hierarchical Mixtures of Experts and the EM Algorithm , 1993, Neural Computation.
[2] Yoshua Bengio,et al. An Input Output HMM Architecture , 1994, NIPS.
[3] Alexei A. Efros,et al. Texture synthesis by non-parametric sampling , 1999, Proceedings of the Seventh IEEE International Conference on Computer Vision.
[4] Geoffrey E. Hinton,et al. Variational Learning for Switching State-Space Models , 2000, Neural Computation.
[5] Zoubin Ghahramani,et al. Propagation Algorithms for Variational Bayesian Learning , 2000, NIPS.
[6] William T. Freeman,et al. Example-Based Super-Resolution , 2002, IEEE Computer Graphics and Applications.
[7] Eero P. Simoncelli,et al. A Parametric Texture Model Based on Joint Statistics of Complex Wavelet Coefficients , 2000, International Journal of Computer Vision.
[8] Michael I. Jordan,et al. Factorial Hidden Markov Models , 1995, Machine Learning.
[9] Bhaskar D. Rao,et al. Variational EM Algorithms for Non-Gaussian Latent Variable Models , 2005, NIPS.
[10] L. Bottou,et al. Training Invariant Support Vector Machines using Selective Sampling , 2005 .
[11] Michael I. Jordan,et al. A Probabilistic Interpretation of Canonical Correlation Analysis , 2005 .
[12] Alexei A. Efros,et al. Scene completion using millions of photographs , 2007, SIGGRAPH 2007.
[13] Francis R. Bach,et al. Sparse probabilistic projections , 2008, NIPS.
[14] Li Fei-Fei,et al. ImageNet: A large-scale hierarchical image database , 2009, CVPR.
[15] Honglak Lee,et al. Convolutional deep belief networks for scalable unsupervised learning of hierarchical representations , 2009, ICML '09.
[16] Geoffrey E. Hinton,et al. Rectified Linear Units Improve Restricted Boltzmann Machines , 2010, ICML.
[17] Pascal Vincent,et al. Stacked Denoising Autoencoders: Learning Useful Representations in a Deep Network with a Local Denoising Criterion , 2010, J. Mach. Learn. Res..
[18] Andrew Y. Ng,et al. Reading Digits in Natural Images with Unsupervised Feature Learning , 2011 .
[19] Andrew Y. Ng,et al. Selecting Receptive Fields in Deep Networks , 2011, NIPS.
[20] Andrew Y. Ng,et al. Learning Feature Representations with K-Means , 2012, Neural Networks: Tricks of the Trade.
[21] Yoshua Bengio,et al. Random Search for Hyper-Parameter Optimization , 2012, J. Mach. Learn. Res..
[22] Andrew L. Maas. Rectifier Nonlinearities Improve Neural Network Acoustic Models , 2013 .
[23] Jeffrey Dean,et al. Distributed Representations of Words and Phrases and their Compositionality , 2013, NIPS.
[24] Rob Fergus,et al. Visualizing and Understanding Convolutional Networks , 2013, ECCV.
[25] Yoshua Bengio,et al. Generative Adversarial Nets , 2014, NIPS.
[26] Max Welling,et al. Auto-Encoding Variational Bayes , 2013, ICLR.
[27] Ivan Laptev,et al. Learning and Transferring Mid-level Image Representations Using Convolutional Neural Networks , 2014, 2014 IEEE Conference on Computer Vision and Pattern Recognition.
[28] Thomas Brox,et al. Striving for Simplicity: The All Convolutional Net , 2014, ICLR.
[29] Surya Ganguli,et al. Deep Unsupervised Learning using Nonequilibrium Thermodynamics , 2015, ICML.
[30] Rob Fergus,et al. Deep Generative Image Models using a Laplacian Pyramid of Adversarial Networks , 2015, NIPS.
[31] Sergey Ioffe,et al. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift , 2015, ICML.
[32] Yinda Zhang,et al. LSUN: Construction of a Large-scale Image Dataset using Deep Learning with Humans in the Loop , 2015, ArXiv.
[33] Tapani Raiko,et al. Semi-supervised Learning with Ladder Networks , 2015, NIPS.
[34] Alex Graves,et al. DRAW: A Recurrent Neural Network For Image Generation , 2015, ICML.
[35] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[36] Jürgen Schmidhuber,et al. Understanding Locally Competitive Networks , 2014, ICLR.
[37] Alexander Mordvintsev,et al. Inceptionism: Going Deeper into Neural Networks , 2015 .
[38] Tianqi Chen,et al. Empirical Evaluation of Rectified Activations in Convolutional Network , 2015, ArXiv.
[39] Thomas Brox,et al. Learning to generate chairs with convolutional neural networks , 2014, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[40] Yann LeCun,et al. Stacked What-Where Auto-encoders , 2015, ArXiv.
[41] Tapani Raiko,et al. Semi-supervised Learning with Ladder Networks , 2015, NIPS.
[42] Yoram Singer,et al. Train faster, generalize better: Stability of stochastic gradient descent , 2015, ICML.
[43] Matthias Bethge,et al. A note on the evaluation of generative models , 2015, ICLR.
[44] John W. Fisher,et al. Dreaming More Data: Class-dependent Distributions over Diffeomorphisms for Learned Data Augmentation , 2015, AISTATS.
[45] Wojciech Zaremba,et al. Improved Techniques for Training GANs , 2016, NIPS.
[46] Thomas Brox,et al. Discriminative Unsupervised Feature Learning with Exemplar Convolutional Neural Networks , 2014, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[47] Xinyun Chen. Under Review as a Conference Paper at Iclr 2017 Delving into Transferable Adversarial Ex- Amples and Black-box Attacks , 2016 .
[48] Thomas Brox,et al. Synthesizing the preferred inputs for neurons in neural networks via deep generator networks , 2016, NIPS.
[49] Thomas Brox,et al. Learning to Generate Chairs, Tables and Cars with Convolutional Networks , 2014, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[50] David Pfau,et al. Unrolled Generative Adversarial Networks , 2016, ICLR.