Proving data-poisoning robustness in decision trees
暂无分享,去创建一个
[1] Isil Dillig,et al. Optimization and abstraction: a synergistic approach for analyzing neural network robustness , 2019, PLDI.
[2] Tianqi Chen,et al. XGBoost: A Scalable Tree Boosting System , 2016, KDD.
[3] Percy Liang,et al. Certified Defenses for Data Poisoning Attacks , 2017, NIPS.
[4] Robert C. Holte,et al. Decision Tree Instability and Active Learning , 2007, ECML.
[5] Claudia Eckert,et al. Adversarial Label Flips Attack on Support Vector Machines , 2012, ECAI.
[6] Daniel M. Kane,et al. Recent Advances in Algorithmic High-Dimensional Robust Statistics , 2019, ArXiv.
[7] Jerry Li,et al. Sever: A Robust Meta-Algorithm for Stochastic Optimization , 2018, ICML.
[8] Francesco Ranzato,et al. Abstract Interpretation of Decision Tree Ensemble Classifiers , 2020, AAAI.
[9] J. Ross Quinlan,et al. Induction of Decision Trees , 1986, Machine Learning.
[10] Geneva G. Belford,et al. Instability of decision tree classification algorithms , 2001, KDD.
[11] David A. Wagner,et al. Towards Evaluating the Robustness of Neural Networks , 2016, 2017 IEEE Symposium on Security and Privacy (SP).
[12] Claudia Eckert,et al. Support vector machines under adversarial label contamination , 2015, Neurocomputing.
[13] Blaine Nelson,et al. Poisoning Attacks against Support Vector Machines , 2012, ICML.
[14] J. Ross Quinlan,et al. C4.5: Programs for Machine Learning , 1992 .
[15] Patrick Cousot,et al. Abstract interpretation: a unified lattice model for static analysis of programs by construction or approximation of fixpoints , 1977, POPL.
[16] Mykel J. Kochenderfer,et al. Reluplex: An Efficient SMT Solver for Verifying Deep Neural Networks , 2017, CAV.
[17] Paul Barford,et al. Data Poisoning Attacks against Autoregressive Models , 2016, AAAI.
[18] Ricky Laishram,et al. Curie: A method for protecting SVM Classifier from Poisoning Attack , 2016, ArXiv.
[19] Xiaojin Zhu,et al. Using Machine Teaching to Identify Optimal Training-Set Attacks on Machine Learners , 2015, AAAI.
[20] Joan Bruna,et al. Intriguing properties of neural networks , 2013, ICLR.
[21] Somesh Jha,et al. Analyzing the Robustness of Nearest Neighbors to Adversarial Examples , 2017, ICML.
[22] Wei-Yin Loh,et al. Classification and regression trees , 2011, WIREs Data Mining Knowl. Discov..
[23] Junfeng Yang,et al. Formal Security Analysis of Neural Networks using Symbolic Intervals , 2018, USENIX Security Symposium.
[24] Cristina Nita-Rotaru,et al. On the Practicality of Integrity Attacks on Document-Level Sentiment Analysis , 2014, AISec '14.
[25] J. Ross Quinlan,et al. Simplifying decision trees , 1987, Int. J. Hum. Comput. Stud..
[26] Dawn Xiaodong Song,et al. Targeted Backdoor Attacks on Deep Learning Systems Using Data Poisoning , 2017, ArXiv.
[27] Claudia Eckert,et al. Is Feature Selection Secure against Training Data Poisoning? , 2015, ICML.
[28] J. Zico Kolter,et al. Provable defenses against adversarial examples via the convex outer adversarial polytope , 2017, ICML.
[29] Timon Gehr,et al. An abstract domain for certifying neural networks , 2019, Proc. ACM Program. Lang..
[30] Daniel M. Kane,et al. Robust Estimators in High Dimensions without the Computational Intractability , 2016, 2016 IEEE 57th Annual Symposium on Foundations of Computer Science (FOCS).
[31] Olatz Arbelaitz,et al. Consolidated Trees: Classifiers with Stable Explanation. A Model to Achieve the Desired Stability in Explanation , 2005, ICAPR.
[32] Swarat Chaudhuri,et al. AI2: Safety and Robustness Certification of Neural Networks with Abstract Interpretation , 2018, 2018 IEEE Symposium on Security and Privacy (SP).
[33] Matthew Mirman,et al. Differentiable Abstract Interpretation for Provably Robust Neural Networks , 2018, ICML.
[34] Peter D. Turney. Technical note: Bias and the quantification of stability , 1995, Machine Learning.