Sorting out Lipschitz function approximation
暂无分享,去创建一个
[1] Alston S. Householder,et al. Unitary Triangularization of a Nonsymmetric Matrix , 1958, JACM.
[2] Å. Björck,et al. An Iterative Algorithm for Computing the Best Estimate of an Orthogonal Matrix , 1971 .
[3] Kurt Hornik,et al. Approximation capabilities of multilayer feedforward networks , 1991, Neural Networks.
[4] Harris Drucker,et al. Improving generalization performance using double backpropagation , 1992, IEEE Trans. Neural Networks.
[5] Peter L. Bartlett,et al. The Sample Complexity of Pattern Classification with Neural Networks: The Size of the Weights is More Important than the Size of the Network , 1998, IEEE Trans. Inf. Theory.
[6] G. Lewicki,et al. Approximation by Superpositions of a Sigmoidal Function , 2003 .
[7] Robert E. Mahony,et al. Optimization Algorithms on Matrix Manifolds , 2007 .
[8] C. Villani. Optimal Transport: Old and New , 2008 .
[9] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[10] I. Yaacov. Lipschitz functions on topometric spaces , 2010, 1010.1600.
[11] S. Mallat,et al. Invariant Scattering Convolution Networks , 2013, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[12] Geoffrey E. Hinton,et al. ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.
[13] Yoshua Bengio,et al. Maxout Networks , 2013, ICML.
[14] Nitish Srivastava,et al. Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..
[15] Razvan Pascanu,et al. On the Number of Linear Regions of Deep Neural Networks , 2014, NIPS.
[16] Joan Bruna,et al. Intriguing properties of neural networks , 2013, ICLR.
[17] Shakir Mohamed,et al. Variational Inference with Normalizing Flows , 2015, ICML.
[18] Xiaogang Wang,et al. Deep Learning Face Attributes in the Wild , 2014, 2015 IEEE International Conference on Computer Vision (ICCV).
[19] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[20] Jonathon Shlens,et al. Explaining and Harnessing Adversarial Examples , 2014, ICLR.
[21] Honglak Lee,et al. Understanding and Improving Convolutional Neural Networks via Concatenated Rectified Linear Units , 2016, ICML.
[22] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[23] Pieter Abbeel,et al. InfoGAN: Interpretable Representation Learning by Information Maximizing Generative Adversarial Nets , 2016, NIPS.
[24] Laurent Condat,et al. A Fast Projection onto the Simplex and the l 1 Ball , 2015 .
[25] Les E. Atlas,et al. Full-Capacity Unitary Recurrent Neural Networks , 2016, NIPS.
[26] Artem N. Chernodub,et al. Norm-preserving Orthogonal Permutation Linear Unit Activation Functions (OPLU) , 2016, ArXiv.
[27] Yoshua Bengio,et al. Unitary Evolution Recurrent Neural Networks , 2015, ICML.
[28] Moustapha Cissé,et al. Parseval Networks: Improving Robustness to Adversarial Examples , 2017, ICML.
[29] Lawrence Carin,et al. Learning Structured Weight Uncertainty in Bayesian Neural Networks , 2017, AISTATS.
[30] Jakub M. Tomczak,et al. Variational Inference with Orthogonal Normalizing Flows , 2017 .
[31] Matus Telgarsky,et al. Spectrally-normalized margin bounds for neural networks , 2017, NIPS.
[32] Surya Ganguli,et al. Resurrecting the sigmoid in deep learning through dynamical isometry: theory and practice , 2017, NIPS.
[33] Luca Antiga,et al. Automatic differentiation in PyTorch , 2017 .
[34] Nathan Srebro,et al. Exploring Generalization in Deep Learning , 2017, NIPS.
[35] David A. Wagner,et al. Towards Evaluating the Robustness of Neural Networks , 2016, 2017 IEEE Symposium on Security and Privacy (SP).
[36] Gabriel Peyré,et al. Sinkhorn-AutoDiff: Tractable Wasserstein Learning of Generative Models , 2017 .
[37] Aaron C. Courville,et al. Improved Training of Wasserstein GANs , 2017, NIPS.
[38] Yuichi Yoshida,et al. Spectral Norm Regularization for Improving the Generalizability of Deep Learning , 2017, ArXiv.
[39] Guillermo Sapiro,et al. Robust Large Margin Deep Neural Networks , 2016, IEEE Transactions on Signal Processing.
[40] Masashi Sugiyama,et al. Lipschitz-Margin Training: Scalable Certification of Perturbation Invariance for Deep Neural Networks , 2018, NeurIPS.
[41] Nathan Srebro,et al. SPECTRALLY-NORMALIZED MARGIN BOUNDS FOR NEURAL NETWORKS , 2018 .
[42] Jacob Abernethy,et al. On Convergence and Stability of GANs , 2018 .
[43] David A. McAllester,et al. A PAC-Bayesian Approach to Spectrally-Normalized Margin Bounds for Neural Networks , 2017, ICLR.
[44] Max Welling,et al. Sylvester Normalizing Flows for Variational Inference , 2018, UAI.
[45] Ritu Chadha,et al. Limitations of the Lipschitz constant as a defense against adversarial examples , 2018, Nemesis/UrbReas/SoGood/IWAISe/GDM@PKDD/ECML.
[46] Bernhard Schölkopf,et al. Adversarial Vulnerability of Neural Networks Increases With Input Dimension , 2018, ArXiv.
[47] Aleksander Madry,et al. Towards Deep Learning Models Resistant to Adversarial Attacks , 2017, ICLR.
[48] Yuichi Yoshida,et al. Spectral Normalization for Generative Adversarial Networks , 2018, ICLR.
[49] Zeynep Akata,et al. Primal-Dual Wasserstein GAN , 2018, ArXiv.
[50] Gabriel Peyré,et al. Learning Generative Models with Sinkhorn Divergences , 2017, AISTATS.
[51] Jascha Sohl-Dickstein,et al. Dynamical Isometry and a Mean Field Theory of CNNs: How to Train 10, 000-Layer Vanilla Convolutional Neural Networks , 2018, ICML.
[52] Samuel S. Schoenholz,et al. Dynamical Isometry and a Mean Field Theory of RNNs: Gating Enables Signal Propagation in Recurrent Neural Networks , 2018, ICML.
[53] Aleksander Madry,et al. There Is No Free Lunch In Adversarial Robustness (But There Are Unexpected Benefits) , 2018, ArXiv.
[54] Aleksander Madry,et al. Robustness May Be at Odds with Accuracy , 2018, ICLR.
[55] Richard S. Zemel,et al. Aggregated Momentum: Stability Through Passive Damping , 2018, ICLR.
[56] Shin Ishii,et al. Virtual Adversarial Training: A Regularization Method for Supervised and Semi-Supervised Learning , 2017, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[57] Gabriel Peyré,et al. Computational Optimal Transport , 2018, Found. Trends Mach. Learn..
[58] Philip M. Long,et al. The Singular Values of Convolutional Layers , 2018, ICLR.
[59] Il Park,et al. Information Geometry of Orthogonal Initializations and Training , 2018, ICLR.
[60] Bernhard Pfahringer,et al. Regularisation of neural networks by enforcing Lipschitz continuity , 2018, Machine Learning.