暂无分享,去创建一个
Inderjit S. Dhillon | Cho-Jui Hsieh | Duane S. Boning | Huan Zhang | Zhao Song | Hongge Chen | Cho-Jui Hsieh | I. Dhillon | D. Boning | Huan Zhang | Zhao Song | Hongge Chen
[1] D. W. Scott,et al. Multivariate Density Estimation, Theory, Practice and Visualization , 1992 .
[2] Geoffrey E. Hinton,et al. Visualizing Data using t-SNE , 2008 .
[3] Adam Tauman Kalai,et al. Efficiently learning mixtures of two Gaussians , 2010, STOC '10.
[4] Ankur Moitra,et al. Settling the Polynomial Learnability of Mixtures of Gaussians , 2010, 2010 IEEE 51st Annual Symposium on Foundations of Computer Science.
[5] Johannes Stallkamp,et al. Detection of traffic signs in real-world images: The German traffic sign detection benchmark , 2013, The 2013 International Joint Conference on Neural Networks (IJCNN).
[6] Jiwen Lu,et al. Discriminative Deep Metric Learning for Face Verification in the Wild , 2014, 2014 IEEE Conference on Computer Vision and Pattern Recognition.
[7] Joan Bruna,et al. Intriguing properties of neural networks , 2013, ICLR.
[8] Jiwen Lu,et al. Deep transfer metric learning , 2015, CVPR.
[9] Jonathon Shlens,et al. Explaining and Harnessing Adversarial Examples , 2014, ICLR.
[10] Moritz Hardt,et al. Tight Bounds for Learning a Mixture of Two Gaussians , 2014, STOC.
[11] Ananthram Swami,et al. Distillation as a Defense to Adversarial Perturbations Against Deep Neural Networks , 2015, 2016 IEEE Symposium on Security and Privacy (SP).
[12] Matthias Hein,et al. Formal Guarantees on the Robustness of a Classifier against Adversarial Manipulation , 2017, NIPS.
[13] Mykel J. Kochenderfer,et al. Reluplex: An Efficient SMT Solver for Verifying Deep Neural Networks , 2017, CAV.
[14] David A. Wagner,et al. Towards Evaluating the Robustness of Neural Networks , 2016, 2017 IEEE Symposium on Security and Privacy (SP).
[15] Samy Bengio,et al. Adversarial Machine Learning at Scale , 2016, ICLR.
[16] Jerry Li,et al. Robustly Learning a Gaussian: Getting Optimal Error, Efficiently , 2017, SODA.
[17] Dan Boneh,et al. Ensemble Adversarial Training: Attacks and Defenses , 2017, ICLR.
[18] Cho-Jui Hsieh,et al. Towards Robust Neural Networks via Random Self-ensemble , 2017, ECCV.
[19] J. Zico Kolter,et al. Scaling provable adversarial defenses , 2018, NeurIPS.
[20] Cho-Jui Hsieh,et al. Efficient Neural Network Robustness Certification with General Activation Functions , 2018, NeurIPS.
[21] Kamyar Azizzadenesheli,et al. Stochastic Activation Pruning for Robust Adversarial Defense , 2018, ICLR.
[22] Yin Tat Lee,et al. Adversarial Examples from Cryptographic Pseudo-Random Generators , 2018, ArXiv.
[23] Leland McInnes,et al. UMAP: Uniform Manifold Approximation and Projection for Dimension Reduction , 2018, ArXiv.
[24] J. Zico Kolter,et al. Provable defenses against adversarial examples via the convex outer adversarial polytope , 2017, ICML.
[25] David A. Wagner,et al. Obfuscated Gradients Give a False Sense of Security: Circumventing Defenses to Adversarial Examples , 2018, ICML.
[26] Aleksander Madry,et al. Towards Deep Learning Models Resistant to Adversarial Attacks , 2017, ICLR.
[27] Yang Song,et al. Generative Adversarial Examples , 2018, NIPS 2018.
[28] Aleksander Madry,et al. Adversarially Robust Generalization Requires More Data , 2018, NeurIPS.
[29] John C. Duchi,et al. Certifying Some Distributional Robustness with Principled Adversarial Training , 2017, ICLR.
[30] Jinfeng Yi,et al. Is Robustness the Cost of Accuracy? - A Comprehensive Study on the Robustness of 18 Deep Image Classification Models , 2018, ECCV.
[31] Colin Raffel,et al. Thermometer Encoding: One Hot Way To Resist Adversarial Examples , 2018, ICLR.
[32] Pushmeet Kohli,et al. A Dual Approach to Scalable Verification of Deep Networks , 2018, UAI.
[33] Logan Engstrom,et al. Synthesizing Robust Adversarial Examples , 2017, ICML.
[34] Aditi Raghunathan,et al. Certified Defenses against Adversarial Examples , 2018, ICLR.
[35] Prateek Mittal,et al. PAC-learning in the presence of evasion adversaries , 2018, NIPS 2018.
[36] Alan L. Yuille,et al. Mitigating adversarial effects through randomization , 2017, ICLR.
[37] Inderjit S. Dhillon,et al. Towards Fast Computation of Certified Robustness for ReLU Networks , 2018, ICML.
[38] Yanjun Qi,et al. Feature Squeezing: Detecting Adversarial Examples in Deep Neural Networks , 2017, NDSS.
[39] James Bailey,et al. Characterizing Adversarial Subspaces Using Local Intrinsic Dimensionality , 2018, ICLR.
[40] Jerry Li,et al. Mixture models, robustness, and sum of squares proofs , 2017, STOC.
[41] Matthew Mirman,et al. Fast and Effective Robustness Certification , 2018, NeurIPS.
[42] Mingyan Liu,et al. Generating Adversarial Examples with Adversarial Networks , 2018, IJCAI.
[43] Moustapha Cissé,et al. Countering Adversarial Images using Input Transformations , 2018, ICLR.
[44] Yang Song,et al. PixelDefend: Leveraging Generative Models to Understand and Defend against Adversarial Examples , 2017, ICLR.
[45] Rama Chellappa,et al. Defense-GAN: Protecting Classifiers Against Adversarial Attacks Using Generative Models , 2018, ICLR.
[46] Aleksander Madry,et al. Robustness May Be at Odds with Accuracy , 2018, ICLR.
[47] Saeed Mahloujifar,et al. The Curse of Concentration in Robust Learning: Evasion and Poisoning Attacks from Concentration of Measure , 2018, AAAI.
[48] Ilya P. Razenshteyn,et al. Adversarial examples from computational constraints , 2018, ICML.
[49] Kui Ren,et al. Distributionally Adversarial Attack , 2018, AAAI.