Stealing Hyperparameters in Machine Learning
暂无分享,去创建一个
[1] Vitaly Shmatikov,et al. Membership Inference Attacks Against Machine Learning Models , 2016, 2017 IEEE Symposium on Security and Privacy (SP).
[2] Xiaoyu Cao,et al. Mitigating Evasion Attacks to Deep Neural Networks via Region-based Classification , 2017, ACSAC.
[3] J. Brian Gray,et al. Introduction to Linear Regression Analysis , 2002, Technometrics.
[4] Gaël Varoquaux,et al. Scikit-learn: Machine Learning in Python , 2011, J. Mach. Learn. Res..
[5] Blaine Nelson,et al. Can machine learning be secure? , 2006, ASIACCS '06.
[6] Dawn Xiaodong Song,et al. Delving into Transferable Adversarial Examples and Black-box Attacks , 2016, ICLR.
[7] Corinna Cortes,et al. Support-Vector Networks , 1995, Machine Learning.
[8] Vladimir Vovk,et al. Kernel Ridge Regression , 2013, Empirical Inference.
[9] A. E. Hoerl,et al. Ridge regression: biased estimation for nonorthogonal problems , 2000 .
[10] Anand D. Sarwate,et al. Differentially Private Empirical Risk Minimization , 2009, J. Mach. Learn. Res..
[11] Marius Kloft,et al. Online Anomaly Detection under Adversarial Impact , 2010, AISTATS.
[12] Jonathon Shlens,et al. Explaining and Harnessing Adversarial Examples , 2014, ICLR.
[13] Vitaly Shmatikov,et al. Machine Learning Models that Remember Too Much , 2017, CCS.
[14] David A. Wagner,et al. Towards Evaluating the Robustness of Neural Networks , 2016, 2017 IEEE Symposium on Security and Privacy (SP).
[15] Yanjun Qi,et al. Automatically Evading Classifiers: A Case Study on PDF Malware Classifiers , 2016, NDSS.
[16] Angelos Stavrou,et al. Malicious PDF detection using metadata and structural features , 2012, ACSAC '12.
[17] Lujo Bauer,et al. Accessorize to a Crime: Real and Stealthy Attacks on State-of-the-Art Face Recognition , 2016, CCS.
[18] Christopher Meek,et al. Adversarial learning , 2005, KDD '05.
[19] Chih-Jen Lin,et al. LIBSVM: A library for support vector machines , 2011, TIST.
[20] Yevgeniy Vorobeychik,et al. Behavioral Experiments in Email Filter Evasion , 2016, AAAI.
[21] Ananthram Swami,et al. Practical Black-Box Attacks against Machine Learning , 2016, AsiaCCS.
[22] Emiliano De Cristofaro,et al. Knock Knock, Who's There? Membership Inference on Aggregate Location Data , 2017, NDSS.
[23] Fabio Roli,et al. Evasion Attacks against Machine Learning at Test Time , 2013, ECML/PKDD.
[24] Michael P. Wellman,et al. Towards the Science of Security and Privacy in Machine Learning , 2016, ArXiv.
[25] Ananthram Swami,et al. Distillation as a Defense to Adversarial Perturbations Against Deep Neural Networks , 2015, 2016 IEEE Symposium on Security and Privacy (SP).
[26] Chih-Jen Lin,et al. A Practical Guide to Support Vector Classication , 2008 .
[27] R. Tibshirani. Regression Shrinkage and Selection via the Lasso , 1996 .
[28] H. Zou,et al. Regularization and variable selection via the elastic net , 2005 .
[29] Ananthram Swami,et al. The Limitations of Deep Learning in Adversarial Settings , 2015, 2016 IEEE European Symposium on Security and Privacy (EuroS&P).
[30] Blaine Nelson,et al. Adversarial machine learning , 2019, AISec '11.
[31] Ling Huang,et al. Near-Optimal Evasion of Convex-Inducing Classifiers , 2010, AISTATS.
[32] James Newsome,et al. Paragraph: Thwarting Signature Learning by Training Maliciously , 2006, RAID.
[33] Yevgeniy Vorobeychik,et al. Data Poisoning Attacks on Factorization-Based Collaborative Filtering , 2016, NIPS.
[34] Joan Bruna,et al. Intriguing properties of neural networks , 2013, ICLR.
[35] Ling Huang,et al. ANTIDOTE: understanding and defending against poisoning of anomaly detectors , 2009, IMC '09.
[36] Pavel Laskov,et al. Practical Evasion of a Learning-Based Classifier: A Case Study , 2014, 2014 IEEE Symposium on Security and Privacy.
[37] Nitish Srivastava,et al. Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..
[38] Fabio Roli,et al. Poisoning attacks to compromise face templates , 2013, 2013 International Conference on Biometrics (ICB).
[39] Micah Sherr,et al. Hidden Voice Commands , 2016, USENIX Security Symposium.
[40] Geoffrey E. Hinton,et al. Deep Learning , 2015, Nature.
[41] Benjamin I. P. Rubinstein,et al. Learners : Co-opting Your Spam Filter , 2009 .
[42] Yevgeniy Vorobeychik,et al. Feature Cross-Substitution in Adversarial Classification , 2014, NIPS.
[43] Somesh Jha,et al. Privacy in Pharmacogenetics: An End-to-End Case Study of Personalized Warfarin Dosing , 2014, USENIX Security Symposium.
[44] Blaine Nelson,et al. Poisoning Attacks against Support Vector Machines , 2012, ICML.
[45] Somesh Jha,et al. Model Inversion Attacks that Exploit Confidence Information and Basic Countermeasures , 2015, CCS.
[46] D. Hosmer,et al. Applied Logistic Regression , 1991 .
[47] Wenke Lee,et al. Misleading worm signature generators using deliberate noise injection , 2006, 2006 IEEE Symposium on Security and Privacy (S&P'06).
[48] Radford M. Neal. Pattern Recognition and Machine Learning , 2007, Technometrics.
[49] James Newsome,et al. Polygraph: automatically generating signatures for polymorphic worms , 2005, 2005 IEEE Symposium on Security and Privacy (S&P'05).
[50] Pavel Laskov,et al. Detection of Malicious PDF Files Based on Hierarchical Document Structure , 2013, NDSS.
[51] Blaine Nelson,et al. Exploiting Machine Learning to Subvert Your Spam Filter , 2008, LEET.
[52] Fan Zhang,et al. Stealing Machine Learning Models via Prediction APIs , 2016, USENIX Security Symposium.