Lower Bounds on Cross-Entropy Loss in the Presence of Test-time Adversaries
暂无分享,去创建一个
[1] Elvis Dohmatob,et al. Generalized No Free Lunch Theorem for Adversarial Robustness , 2018, ICML.
[2] Morgane Goibert,et al. Adversarial Robustness via Label-Smoothing , 2020 .
[3] Yann LeCun,et al. The mnist database of handwritten digits , 2005 .
[4] Ilias Diakonikolas,et al. Efficiently Learning Adversarially Robust Halfspaces with Noise , 2020, ICML.
[5] Aleksander Madry,et al. Adversarially Robust Generalization Requires More Data , 2018, NeurIPS.
[6] Richard M. Karp,et al. Theoretical Improvements in Algorithmic Efficiency for Network Flow Problems , 1972, Combinatorial Optimization.
[7] Nathan Srebro,et al. VC Classes are Adversarially Robustly Learnable, but Only Improperly , 2019, COLT.
[8] Jonathon Shlens,et al. Explaining and Harnessing Adversarial Examples , 2014, ICLR.
[9] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[10] Kannan Ramchandran,et al. Rademacher Complexity for Adversarially Robust Generalization , 2018, ICML.
[11] Joel Nothman,et al. SciPy 1.0-Fundamental Algorithms for Scientific Computing in Python , 2019, ArXiv.
[12] Stephen P. Boyd,et al. Convex Optimization , 2004, Algorithms and Theory of Computation Handbook.
[13] Prateek Mittal,et al. PAC-learning in the presence of adversaries , 2018, NeurIPS.
[14] Michael I. Jordan,et al. Theoretically Principled Trade-off between Robustness and Accuracy , 2019, ICML.
[15] Wei Cai,et al. A Survey on Security Threats and Defensive Techniques of Machine Learning: A Data Driven View , 2018, IEEE Access.
[16] Shai Ben-David,et al. Understanding Machine Learning: From Theory to Algorithms , 2014 .
[17] Somesh Jha,et al. Adversarially Robust Learning Could Leverage Computational Hardness , 2020, ALT.
[18] Suman Jana,et al. HYDRA: Pruning Adversarially Robust Neural Networks , 2020, NeurIPS.
[19] Saeed Mahloujifar,et al. The Curse of Concentration in Robust Learning: Evasion and Poisoning Attacks from Concentration of Measure , 2018, AAAI.
[20] Martin Wattenberg,et al. Adversarial Spheres , 2018, ICLR.
[21] Michael P. Wellman,et al. Towards the Science of Security and Privacy in Machine Learning , 2016, ArXiv.
[22] Aravindan Vijayaraghavan,et al. On Robustness to Adversarial Examples and Polynomial Optimization , 2019, NeurIPS.
[23] Ryan P. Adams,et al. Motivating the Rules of the Game for Adversarial Example Research , 2018, ArXiv.
[24] Yisen Wang,et al. Adversarial Weight Perturbation Helps Robust Generalization , 2020, NeurIPS.
[25] J. Zico Kolter,et al. Overfitting in adversarially robust deep learning , 2020, ICML.
[26] James Bailey,et al. Improving Adversarial Robustness Requires Revisiting Misclassified Examples , 2020, ICLR.
[27] Bo Li,et al. SoK: Certified Robustness for Deep Neural Networks , 2020, ArXiv.
[28] Saeed Mahloujifar,et al. Adversarial Risk and Robustness: General Definitions and Implications for the Uniform Distribution , 2018, NeurIPS.
[29] Quoc V. Le,et al. Smooth Adversarial Training , 2020, ArXiv.
[30] Pradeep Ravikumar,et al. Sharp Statistical Guarantees for Adversarially Robust Gaussian Classification , 2020, ArXiv.
[31] Fabio Roli,et al. Wild Patterns: Ten Years After the Rise of Adversarial Machine Learning , 2018, CCS.
[32] Ludwig Schmidt,et al. Unlabeled Data Improves Adversarial Robustness , 2019, NeurIPS.
[33] Joan Bruna,et al. Intriguing properties of neural networks , 2013, ICLR.
[34] Timothy A. Mann,et al. Uncovering the Limits of Adversarial Training against Norm-Bounded Adversarial Examples , 2020, ArXiv.
[35] Ilya P. Razenshteyn,et al. Adversarial examples from computational constraints , 2018, ICML.
[36] Hang Su,et al. Bag of Tricks for Adversarial Training , 2020, ICLR.
[37] Daniel Cullina,et al. Lower Bounds on Adversarial Robustness from Optimal Transport , 2019, NeurIPS.
[38] Andrew V. Goldberg,et al. A new approach to the maximum flow problem , 1986, STOC '86.
[39] Muni Sreenivas Pydi,et al. Adversarial Risk via Optimal Transport and Optimal Couplings , 2019, IEEE Transactions on Information Theory.
[40] David A. Wagner,et al. Towards Evaluating the Robustness of Neural Networks , 2016, 2017 IEEE Symposium on Security and Privacy (SP).
[41] Dawn Xiaodong Song,et al. Practical Black-Box Attacks on Deep Neural Networks Using Efficient Query Mechanisms , 2018, ECCV.
[42] Adel Javanmard,et al. Precise Tradeoffs in Adversarial Training for Linear Regression , 2020, COLT.
[43] Yang Song,et al. Improving the Robustness of Deep Neural Networks via Stability Training , 2016, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[44] Ivan Evtimov,et al. Security and Machine Learning in the Real World , 2020, ArXiv.
[45] Roland Vollgraf,et al. Fashion-MNIST: a Novel Image Dataset for Benchmarking Machine Learning Algorithms , 2017, ArXiv.
[46] Matthias Hein,et al. Reliable evaluation of adversarial robustness with an ensemble of diverse parameter-free attacks , 2020, ICML.
[47] Aleksander Madry,et al. Towards Deep Learning Models Resistant to Adversarial Attacks , 2017, ICLR.
[48] Prateek Mittal,et al. RobustBench: a standardized adversarial robustness benchmark , 2020, ArXiv.