暂无分享,去创建一个
[1] D. Ruppert. Robust Statistics: The Approach Based on Influence Functions , 1987 .
[2] Charles R. Johnson,et al. Matrix analysis , 1985, Statistical Inference for Engineers and Data Scientists.
[3] Adam M. Oberman. A convergent difference scheme for the infinity Laplacian: construction of absolutely minimizing Lipschitz extensions , 2004, Math. Comput..
[4] Matthias Bethge,et al. Decision-Based Adversarial Attacks: Reliable Attacks Against Black-Box Machine Learning Models , 2017, ICLR.
[5] Dan Boneh,et al. Ensemble Adversarial Training: Attacks and Defenses , 2017, ICLR.
[6] Bernhard Pfahringer,et al. Regularisation of neural networks by enforcing Lipschitz continuity , 2018, Machine Learning.
[7] Matthias Bethge,et al. Foolbox v0.8.0: A Python toolbox to benchmark the robustness of machine learning models , 2017, ArXiv.
[8] L. Evans,et al. Optimal Lipschitz extensions and the infinity laplacian , 2001 .
[9] Thomas Brox,et al. Striving for Simplicity: The All Convolutional Net , 2014, ICLR.
[10] Matthias Hein,et al. Formal Guarantees on the Robustness of a Classifier against Adversarial Manipulation , 2017, NIPS.
[11] W. B. Johnson,et al. Extensions of Lipschitz mappings into Hilbert space , 1984 .
[12] Seyed-Mohsen Moosavi-Dezfooli,et al. DeepFool: A Simple and Accurate Method to Fool Deep Neural Networks , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[13] L. Evans. Measure theory and fine properties of functions , 1992 .
[14] Yuichi Yoshida,et al. Spectral Normalization for Generative Adversarial Networks , 2018, ICLR.
[15] Patrick D. McDaniel,et al. Making machine learning robust against adversarial inputs , 2018, Commun. ACM.
[16] Peter L. Bartlett,et al. For Valid Generalization the Size of the Weights is More Important than the Size of the Network , 1996, NIPS.
[17] Graham W. Taylor,et al. Improved Regularization of Convolutional Neural Networks with Cutout , 2017, ArXiv.
[18] Moustapha Cissé,et al. Parseval Networks: Improving Robustness to Adversarial Examples , 2017, ICML.
[19] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[20] Michael P. Wellman,et al. SoK: Security and Privacy in Machine Learning , 2018, 2018 IEEE European Symposium on Security and Privacy (EuroS&P).
[21] Arun Ross,et al. Score normalization in multimodal biometric systems , 2005, Pattern Recognit..
[22] Jeff Calder,et al. Lipschitz regularized Deep Neural Networks converge and generalize , 2018, ArXiv.
[23] Joan Bruna,et al. Intriguing properties of neural networks , 2013, ICLR.
[24] David A. Wagner,et al. Obfuscated Gradients Give a False Sense of Security: Circumventing Defenses to Adversarial Examples , 2018, ICML.
[25] F. A. Valentine. A Lipschitz Condition Preserving Extension for a Vector Function , 1945 .
[26] Aleksander Madry,et al. Towards Deep Learning Models Resistant to Adversarial Attacks , 2017, ICLR.
[27] Jonathon Shlens,et al. Explaining and Harnessing Adversarial Examples , 2014, ICLR.
[28] Geoffrey E. Hinton,et al. Distilling the Knowledge in a Neural Network , 2015, ArXiv.
[29] David A. Forsyth,et al. SafetyNet: Detecting and Rejecting Adversarial Examples Robustly , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).
[30] Samy Bengio,et al. Adversarial examples in the physical world , 2016, ICLR.
[31] Werner A. Stahel,et al. Robust Statistics: The Approach Based on Influence Functions , 1987 .
[32] Jian Sun,et al. Identity Mappings in Deep Residual Networks , 2016, ECCV.
[33] L. Rudin,et al. Nonlinear total variation based noise removal algorithms , 1992 .
[34] Guillermo Sapiro,et al. Image inpainting , 2000, SIGGRAPH.
[35] David A. Wagner,et al. Towards Evaluating the Robustness of Neural Networks , 2016, 2017 IEEE Symposium on Security and Privacy (SP).
[36] Stephen P. Boyd,et al. Convex Optimization , 2004, Algorithms and Theory of Computation Handbook.