Model Extraction Warning in MLaaS Paradigm
暂无分享,去创建一个
Vijay Arya | Sameep Mehta | Manish Kesarwani | Bhaskar Mukhoty | S. Mehta | B. Mukhoty | V. Arya | M. Kesarwani
[1] Cynthia Dwork,et al. Differential Privacy: A Survey of Results , 2008, TAMC.
[2] Jude W. Shavlik,et al. Combining the Predictions of Multiple Classifiers: Using Competitive Learning to Initialize Neural Networks , 1995, IJCAI.
[3] Blaine Nelson,et al. Poisoning Attacks against Support Vector Machines , 2012, ICML.
[4] Binghui Wang,et al. Stealing Hyperparameters in Machine Learning , 2018, 2018 IEEE Symposium on Security and Privacy (SP).
[5] Shangbo Zhou,et al. X-TREPAN: a multi class regression and adapted extraction of comprehensible decision tree in artificial neural networks , 2015, ArXiv.
[6] Ananthram Swami,et al. The Limitations of Deep Learning in Adversarial Settings , 2015, 2016 IEEE European Symposium on Security and Privacy (EuroS&P).
[7] Rudy Setiono,et al. Extracting Rules from Neural Networks by Pruning and Hidden-Unit Splitting , 1997, Neural Computation.
[8] Jude W. Shavlik,et al. in Advances in Neural Information Processing , 1996 .
[9] Z. Hasan. A Survey on Shari’Ah Governance Practices in Malaysia, GCC Countries and the UK , 2011 .
[10] David A. Wagner,et al. Towards Evaluating the Robustness of Neural Networks , 2016, 2017 IEEE Symposium on Security and Privacy (SP).
[11] Blaine Nelson,et al. Exploiting Machine Learning to Subvert Your Spam Filter , 2008, LEET.
[12] Fan Zhang,et al. Stealing Machine Learning Models via Prediction APIs , 2016, USENIX Security Symposium.
[13] Giovanni Felici,et al. Hacking smart machines with smarter ones: How to extract meaningful data from machine learning classifiers , 2013, Int. J. Secur. Networks.
[14] Jonathon Shlens,et al. Explaining and Harnessing Adversarial Examples , 2014, ICLR.
[15] Somesh Jha,et al. Model Inversion Attacks that Exploit Confidence Information and Basic Countermeasures , 2015, CCS.
[16] Geoffrey E. Hinton,et al. Distilling a Neural Network Into a Soft Decision Tree , 2017, CEx@AI*IA.
[17] James Newsome,et al. Paragraph: Thwarting Signature Learning by Training Maliciously , 2006, RAID.
[18] J. Doug Tygar,et al. Adversarial machine learning , 2019, AISec '11.
[19] J. Ross Quinlan,et al. Induction of Decision Trees , 1986, Machine Learning.
[20] Fabio Roli,et al. Evasion Attacks against Machine Learning at Test Time , 2013, ECML/PKDD.
[21] Joan Bruna,et al. Intriguing properties of neural networks , 2013, ICLR.
[22] Pavel Laskov,et al. Practical Evasion of a Learning-Based Classifier: A Case Study , 2014, 2014 IEEE Symposium on Security and Privacy.
[23] Christopher Meek,et al. Adversarial learning , 2005, KDD '05.
[24] Ananthram Swami,et al. Practical Black-Box Attacks against Machine Learning , 2016, AsiaCCS.
[25] Geoffrey E. Hinton,et al. Distilling the Knowledge in a Neural Network , 2015, ArXiv.
[26] Gary R. Weckman,et al. Trepan-Plus: An Extension of a Decision Tree Extraction Algorithm Utilizing Artificial Neural Networks , 2007 .
[27] Sameep Mehta,et al. A Survey on Resilient Machine Learning , 2017, ArXiv.