暂无分享,去创建一个
[1] R. R. Bahadur. Sufficiency and Statistical Decision Functions , 1954 .
[2] Thomas M. Cover,et al. Elements of Information Theory , 2005 .
[3] Geoffrey E. Hinton,et al. Keeping the neural networks simple by minimizing the description length of the weights , 1993, COLT '93.
[4] G. Golub,et al. Some large-scale matrix computation problems , 1996 .
[5] Yoshua Bengio,et al. Gradient-based learning applied to document recognition , 1998, Proc. IEEE.
[6] Naftali Tishby,et al. The information bottleneck method , 2000, ArXiv.
[7] R. Dennis Cook,et al. Optimal sufficient dimension reduction in regressions with categorical predictors , 2002 .
[8] Stuart Geman,et al. Invariance and selectivity in the ventral visual pathway , 2006, Journal of Physiology-Paris.
[9] Yoshua. Bengio,et al. Learning Deep Architectures for AI , 2007, Found. Trends Mach. Learn..
[10] Stefano Soatto,et al. Actionable information in vision , 2009, 2009 IEEE 12th International Conference on Computer Vision.
[11] Stefano Soatto,et al. On the set of images modulo viewpoint and contrast changes , 2009, 2009 IEEE Conference on Computer Vision and Pattern Recognition.
[12] Michael I. Jordan,et al. Kernel dimension reduction in regression , 2009, 0908.1854.
[13] L. Rosasco. THE COMPUTATIONAL MAGIC OF THE VENTRAL STREAM , 2011 .
[14] Stéphane Mallat,et al. Classification with scattering operators , 2010, CVPR 2011.
[15] Yann LeCun,et al. Learning Invariant Feature Hierarchies , 2012, ECCV Workshops.
[16] David A. McAllester. A PAC-Bayesian Tutorial with A Dropout Bound , 2013, ArXiv.
[17] Nitish Srivastava,et al. Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..
[18] Yoshua Bengio,et al. Generative Adversarial Nets , 2014, NIPS.
[19] Max Welling,et al. Auto-Encoding Variational Bayes , 2013, ICLR.
[20] Stefano Soatto,et al. Visual Representations: Defining Properties and Deep Approximations , 2014, ICLR 2016.
[21] Donald Geman,et al. Visual Turing test for computer vision systems , 2015, Proceedings of the National Academy of Sciences.
[22] Aram Galstyan,et al. Maximally Informative Hierarchical Representations of High-Dimensional Data , 2014, AISTATS.
[23] Thomas Brox,et al. Striving for Simplicity: The All Convolutional Net , 2014, ICLR.
[24] Lorenzo Rosasco,et al. On Invariance and Selectivity in Representation Learning , 2015, ArXiv.
[25] Naftali Tishby,et al. Deep learning and the information bottleneck principle , 2015, 2015 IEEE Information Theory Workshop (ITW).
[26] Ariel D. Procaccia,et al. Variational Dropout and the Local Reparameterization Trick , 2015, NIPS.
[27] Shuo Yang,et al. From Facial Parts Responses to Face Detection: A Deep Learning Approach , 2015, 2015 IEEE International Conference on Computer Vision (ICCV).
[28] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[29] Stefano Soatto,et al. Information Dropout: learning optimal representations through noise , 2017, ArXiv.
[30] Soumith Chintala,et al. Unsupervised Representation Learning with Deep Convolutional Generative Adversarial Networks , 2015, ICLR.
[31] Sepp Hochreiter,et al. Fast and Accurate Deep Network Learning by Exponential Linear Units (ELUs) , 2015, ICLR.
[32] Naftali Tishby,et al. Opening the Black Box of Deep Neural Networks via Information , 2017, ArXiv.
[33] Tengyu Ma,et al. On the Ability of Neural Nets to Express Distributions , 2017, COLT.
[34] Dmitry P. Vetrov,et al. Variational Dropout Sparsifies Deep Neural Networks , 2017, ICML.
[35] Lucas Theis,et al. Amortised MAP Inference for Image Super-resolution , 2016, ICLR.
[36] Gintare Karolina Dziugaite,et al. Computing Nonvacuous Generalization Bounds for Deep (Stochastic) Neural Networks with Many More Parameters than Training Data , 2017, UAI.
[37] Samy Bengio,et al. Understanding deep learning requires rethinking generalization , 2016, ICLR.
[38] Razvan Pascanu,et al. Sharp Minima Can Generalize For Deep Nets , 2017, ICML.
[39] Alexander A. Alemi,et al. Deep Variational Information Bottleneck , 2017, ICLR.
[40] Stefano Soatto,et al. Entropy-SGD: biasing gradient descent into wide valleys , 2016, ICLR.
[41] Dmitry P. Vetrov,et al. Structured Bayesian Pruning via Log-Normal Multiplicative Noise , 2017, NIPS.
[42] Stefano Soatto,et al. Information Dropout: Learning Optimal Representations Through Noisy Computation , 2016, IEEE Transactions on Pattern Analysis and Machine Intelligence.