暂无分享,去创建一个
[1] Blaine Nelson,et al. Can machine learning be secure? , 2006, ASIACCS '06.
[2] Brent Lagesse,et al. Analysis of Causative Attacks against SVMs Learning from Data Streams , 2017, IWSPA@CODASPY.
[3] Nathan Linial,et al. Collective Coin Flipping , 1989, Adv. Comput. Res..
[4] Ilya P. Razenshteyn,et al. Adversarial examples from computational constraints , 2018, ICML.
[5] Michael P. Wellman,et al. Towards the Science of Security and Privacy in Machine Learning , 2016, ArXiv.
[6] Santosh S. Vempala,et al. Agnostic Estimation of Mean and Covariance , 2016, 2016 IEEE 57th Annual Symposium on Foundations of Computer Science (FOCS).
[7] Ananthram Swami,et al. Practical Black-Box Attacks against Machine Learning , 2016, AsiaCCS.
[8] Yanjun Qi,et al. Feature Squeezing: Detecting Adversarial Examples in Deep Neural Networks , 2017, NDSS.
[9] Uriel Feige,et al. Robust Inference for Multiclass Classification , 2018, ALT.
[10] Michael E. Saks,et al. Some extremal problems arising from discrete control processes , 1989, Comb..
[11] Percy Liang,et al. Understanding Black-box Predictions via Influence Functions , 2017, ICML.
[12] Alexandr Andoni,et al. Near-Optimal Hashing Algorithms for Approximate Nearest Neighbor in High Dimensions , 2006, 2006 47th Annual IEEE Symposium on Foundations of Computer Science (FOCS'06).
[13] Michael Kearns,et al. Efficient noise-tolerant learning from statistical queries , 1993, STOC.
[14] Uriel Feige,et al. Learning and inference in the presence of corrupted inputs , 2015, COLT.
[15] Eyal Kushilevitz,et al. PAC learning with nasty noise , 1999, Theor. Comput. Sci..
[16] M. Talagrand. Concentration of measure and isoperimetric inequalities in product spaces , 1994, math/9406212.
[17] Ananthram Swami,et al. Distillation as a Defense to Adversarial Perturbations Against Deep Neural Networks , 2015, 2016 IEEE Symposium on Security and Privacy (SP).
[18] Alexandr Andoni,et al. Optimal Data-Dependent Hashing for Approximate Near Neighbors , 2015, STOC.
[19] Fabio Roli,et al. Security Evaluation of Pattern Classifiers under Attack , 2014, IEEE Transactions on Knowledge and Data Engineering.
[20] Yizhen Wang,et al. Data Poisoning Attacks against Online Learning , 2018, ArXiv.
[21] Fabio Roli,et al. Evasion Attacks against Machine Learning at Test Time , 2013, ECML/PKDD.
[22] Saeed Mahloujifar,et al. Adversarial Risk and Robustness: General Definitions and Implications for the Uniform Distribution , 2018, NeurIPS.
[23] Saeed Mahloujifar,et al. Multi-party Poisoning through Generalized p-Tampering , 2018, IACR Cryptol. ePrint Arch..
[24] Sivaraman Balakrishnan,et al. Robust estimation via robust gradient estimation , 2018, Journal of the Royal Statistical Society: Series B (Statistical Methodology).
[25] Moshe Tennenholtz,et al. Robust Probabilistic Inference , 2014, SODA.
[26] Alexandr Andoni,et al. Approximate Nearest Neighbor Search in High Dimensions , 2018, Proceedings of the International Congress of Mathematicians (ICM 2018).
[27] Blaine Nelson,et al. Poisoning Attacks against Support Vector Machines , 2012, ICML.
[28] Tom Goldstein,et al. Are adversarial examples inevitable? , 2018, ICLR.
[29] Daniel M. Kane,et al. Statistical Query Lower Bounds for Robust Estimation of High-Dimensional Gaussians and Gaussian Mixtures , 2016, 2017 IEEE 58th Annual Symposium on Foundations of Computer Science (FOCS).
[30] Piotr Indyk,et al. Approximate nearest neighbors: towards removing the curse of dimensionality , 1998, STOC '98.
[31] Yishay Mansour,et al. Improved generalization bounds for robust learning , 2018, ALT.
[32] Logan Engstrom,et al. Black-box Adversarial Attacks with Limited Queries and Information , 2018, ICML.
[33] Hamza Fawzi,et al. Adversarial vulnerability for any classifier , 2018, NeurIPS.
[34] David A. Wagner,et al. Towards Evaluating the Robustness of Neural Networks , 2016, 2017 IEEE Symposium on Security and Privacy (SP).
[35] Maria-Florina Balcan,et al. The Power of Localization for Efficiently Learning Linear Separators with Noise , 2013, J. ACM.
[36] Michael P. Wellman,et al. SoK: Security and Privacy in Machine Learning , 2018, 2018 IEEE European Symposium on Security and Privacy (EuroS&P).
[37] Claudia Eckert,et al. Is Feature Selection Secure against Training Data Poisoning? , 2015, ICML.
[38] Aleksander Madry,et al. Adversarially Robust Generalization Requires More Data , 2018, NeurIPS.
[39] Jerry Li,et al. Being Robust (in High Dimensions) Can Be Practical , 2017, ICML.
[40] Yael Tauman Kalai,et al. A Lower Bound for Adaptively-Secure Collective Coin-Flipping Protocols , 2018, Electron. Colloquium Comput. Complex..
[41] Tudor Dumitras,et al. Poison Frogs! Targeted Clean-Label Poisoning Attacks on Neural Networks , 2018, NeurIPS.
[42] Ming Li,et al. Learning in the presence of malicious errors , 1993, STOC '88.
[43] Saeed Mahloujifar,et al. Blockwise p-Tampering Attacks on Cryptographic Primitives, Extractors, and Learners , 2017, TCC.
[44] Saeed Mahloujifar,et al. The Curse of Concentration in Robust Learning: Evasion and Poisoning Attacks from Concentration of Measure , 2018, AAAI.
[45] Gregory Valiant,et al. Learning from untrusted data , 2016, STOC.
[46] Aleksander Madry,et al. Towards Deep Learning Models Resistant to Adversarial Attacks , 2017, ICLR.
[47] Jonathon Shlens,et al. Explaining and Harnessing Adversarial Examples , 2014, ICLR.
[48] Yael Tauman Kalai,et al. Adaptively Secure Coin-Flipping, Revisited , 2015, ICALP.
[49] Ilias Diakonikolas,et al. Efficient Algorithms and Lower Bounds for Robust Linear Regression , 2018, SODA.
[50] Colin McDiarmid,et al. Surveys in Combinatorics, 1989: On the method of bounded differences , 1989 .
[51] Martin Wattenberg,et al. Adversarial Spheres , 2018, ICLR.
[52] Pradeep Ravikumar,et al. On Adversarial Risk and Training , 2018, ArXiv.
[53] Jerry Li,et al. Sever: A Robust Meta-Algorithm for Stochastic Optimization , 2018, ICML.
[54] Yael Tauman Kalai,et al. A Lower Bound for Adaptively-Secure Collective Coin Flipping Protocols , 2020, Combinatorica.
[55] V. Milman,et al. Asymptotic Theory Of Finite Dimensional Normed Spaces , 1986 .
[56] Joan Bruna,et al. Intriguing properties of neural networks , 2013, ICLR.
[57] Ling Huang,et al. ANTIDOTE: understanding and defending against poisoning of anomaly detectors , 2009, IMC '09.
[58] V. Milman,et al. Unconditional and symmetric sets inn-dimensional normed spaces , 1980 .
[59] Daniel M. Kane,et al. List-decodable robust mean estimation and learning mixtures of spherical gaussians , 2017, STOC.
[60] Jerry Li,et al. Robustly Learning a Gaussian: Getting Optimal Error, Efficiently , 2017, SODA.
[61] Daniel M. Kane,et al. Robust Estimators in High Dimensions without the Computational Intractability , 2016, 2016 IEEE 57th Annual Symposium on Foundations of Computer Science (FOCS).
[62] David A. Wagner,et al. Obfuscated Gradients Give a False Sense of Security: Circumventing Defenses to Adversarial Examples , 2018, ICML.