暂无分享,去创建一个
[1] Robert H. Wilkinson,et al. A Method of Generating Functions of Several Variables Using Analog Diode Logic , 1963, IEEE Trans. Electron. Comput..
[2] Jason Weston,et al. Natural Language Processing (Almost) from Scratch , 2011, J. Mach. Learn. Res..
[3] Jeff Calder,et al. Lipschitz regularized Deep Neural Networks converge and generalize , 2018, ArXiv.
[4] David A. Wagner,et al. Obfuscated Gradients Give a False Sense of Security: Circumventing Defenses to Adversarial Examples , 2018, ICML.
[5] Ritu Chadha,et al. Limitations of the Lipschitz constant as a defense against adversarial examples , 2018, Nemesis/UrbReas/SoGood/IWAISe/GDM@PKDD/ECML.
[6] Samy Bengio,et al. Adversarial Machine Learning at Scale , 2016, ICLR.
[7] Dumitru Erhan,et al. Going deeper with convolutions , 2014, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[8] George Cybenko,et al. Approximation by superpositions of a sigmoidal function , 1989, Math. Control. Signals Syst..
[9] Aditi Raghunathan,et al. Certified Defenses against Adversarial Examples , 2018, ICLR.
[10] Moustapha Cissé,et al. Parseval Networks: Improving Robustness to Adversarial Examples , 2017, ICML.
[11] Shuning Wang,et al. General constructive representations for continuous piecewise-linear functions , 2004, IEEE Trans. Circuits Syst. I Regul. Pap..
[12] J. Zico Kolter,et al. Provable defenses against adversarial examples via the convex outer adversarial polytope , 2017, ICML.
[13] J. M. Tarela,et al. A representation method for PWL functions oriented to parallel processing , 1990 .
[14] Artem N. Chernodub,et al. Norm-preserving Orthogonal Permutation Linear Unit Activation Functions (OPLU) , 2016, ArXiv.
[15] J. Zico Kolter,et al. Scaling provable adversarial defenses , 2018, NeurIPS.
[16] David A. Wagner,et al. Towards Evaluating the Robustness of Neural Networks , 2016, 2017 IEEE Symposium on Security and Privacy (SP).
[17] Geoffrey E. Hinton,et al. ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.
[18] Masashi Sugiyama,et al. Lipschitz-Margin Training: Scalable Certification of Perturbation Invariance for Deep Neural Networks , 2018, NeurIPS.
[19] Cem Anil,et al. Sorting out Lipschitz function approximation , 2018, ICML.
[20] Chih-Hong Cheng,et al. Maximum Resilience of Artificial Neural Networks , 2017, ATVA.
[21] Bernhard Pfahringer,et al. Regularisation of neural networks by enforcing Lipschitz continuity , 2018, Machine Learning.
[22] J. M. Tarela,et al. Region configurations for realizability of lattice Piecewise-Linear models , 1999 .
[23] Aleksander Madry,et al. Towards Deep Learning Models Resistant to Adversarial Attacks , 2017, ICLR.
[24] Jonathon Shlens,et al. Explaining and Harnessing Adversarial Examples , 2014, ICLR.
[25] J. Zico Kolter,et al. Certified Adversarial Robustness via Randomized Smoothing , 2019, ICML.
[26] Mykel J. Kochenderfer,et al. Reluplex: An Efficient SMT Solver for Verifying Deep Neural Networks , 2017, CAV.
[27] Adam M. Oberman,et al. Lipschitz regularized Deep Neural Networks generalize and are adversarially robust , 2018 .