Robust Decision Trees Against Adversarial Examples
暂无分享,去创建一个
Cho-Jui Hsieh | Duane S. Boning | Huan Zhang | Hongge Chen | Cho-Jui Hsieh | D. Boning | Huan Zhang | Hongge Chen
[1] Jinfeng Yi,et al. ZOO: Zeroth Order Optimization Based Black-box Attacks to Deep Neural Networks without Training Substitute Models , 2017, AISec@CCS.
[2] Jinfeng Yi,et al. EAD: Elastic-Net Attacks to Deep Neural Networks via Adversarial Examples , 2017, AAAI.
[3] Tianqi Chen,et al. XGBoost: A Scalable Tree Boosting System , 2016, KDD.
[4] Colin Raffel,et al. Thermometer Encoding: One Hot Way To Resist Adversarial Examples , 2018, ICLR.
[5] Cho-Jui Hsieh,et al. Rob-GAN: Generator, Discriminator, and Adversarial Attacker , 2018, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[6] Wei-Yin Loh,et al. Classification and regression trees , 2011, WIREs Data Mining Knowl. Discov..
[7] Cho-Jui Hsieh,et al. Adv-BNN: Improved Adversarial Defense through Robust Bayesian Neural Network , 2018, ICLR.
[8] Jinfeng Yi,et al. Seq2Sick: Evaluating the Robustness of Sequence-to-Sequence Models with Adversarial Examples , 2018, AAAI.
[9] J. Ross Quinlan,et al. Induction of Decision Trees , 1986, Machine Learning.
[10] Fred A. Hamprecht,et al. Cost efficient gradient boosting , 2017, NIPS.
[11] Jinfeng Yi,et al. Attacking Visual Language Grounding with Adversarial Examples: A Case Study on Neural Image Captioning , 2017, ACL.
[12] Ji Feng,et al. Deep Forest: Towards An Alternative to Deep Neural Networks , 2017, IJCAI.
[13] Patrick D. McDaniel,et al. Transferability in Machine Learning: from Phenomena to Black-Box Attacks using Adversarial Samples , 2016, ArXiv.
[14] Aleksander Madry,et al. Towards Deep Learning Models Resistant to Adversarial Attacks , 2017, ICLR.
[15] Jonathon Shlens,et al. Explaining and Harnessing Adversarial Examples , 2014, ICLR.
[16] Joan Bruna,et al. Intriguing properties of neural networks , 2013, ICLR.
[17] Yanjun Qi,et al. Feature Squeezing: Detecting Adversarial Examples in Deep Neural Networks , 2017, NDSS.
[18] Aleksander Madry,et al. Robustness May Be at Odds with Accuracy , 2018, ICLR.
[19] Stephen Tyree,et al. Parallel boosted regression trees for web search ranking , 2011, WWW.
[20] Rama Chellappa,et al. Defense-GAN: Protecting Classifiers Against Adversarial Attacks Using Generative Models , 2018, ICLR.
[21] James Bailey,et al. Characterizing Adversarial Subspaces Using Local Intrinsic Dimensionality , 2018, ICLR.
[22] Jinfeng Yi,et al. Query-Efficient Hard-label Black-box Attack: An Optimization-based Approach , 2018, ICLR.
[23] Inderjit S. Dhillon,et al. Gradient Boosted Decision Trees for High Dimensional Sparse Output , 2017, ICML.
[24] David A. Wagner,et al. Audio Adversarial Examples: Targeted Attacks on Speech-to-Text , 2018, 2018 IEEE Security and Privacy Workshops (SPW).
[25] Samy Bengio,et al. Adversarial Machine Learning at Scale , 2016, ICLR.
[26] Aleksander Madry,et al. Adversarial Examples Are Not Bugs, They Are Features , 2019, NeurIPS.
[27] David A. Wagner,et al. Obfuscated Gradients Give a False Sense of Security: Circumventing Defenses to Adversarial Examples , 2018, ICML.
[28] J. Doug Tygar,et al. Evasion and Hardening of Tree Ensemble Classifiers , 2015, ICML.
[29] Tie-Yan Liu,et al. LightGBM: A Highly Efficient Gradient Boosting Decision Tree , 2017, NIPS.
[30] Cho-Jui Hsieh,et al. GPU-acceleration for Large-scale Tree Boosting , 2017, ArXiv.
[31] Sandy H. Huang,et al. Adversarial Attacks on Neural Network Policies , 2017, ICLR.
[32] Yizheng Chen,et al. MixTrain: Scalable Training of Formally Robust Neural Networks , 2018, ArXiv.
[33] Ji Feng,et al. Multi-Layered Gradient Boosting Decision Trees , 2018, NeurIPS.
[34] Yang Song,et al. PixelDefend: Leveraging Generative Models to Understand and Defend against Adversarial Examples , 2017, ICLR.
[35] J. Zico Kolter,et al. Scaling provable adversarial defenses , 2018, NeurIPS.
[36] David A. Wagner,et al. Towards Evaluating the Robustness of Neural Networks , 2016, 2017 IEEE Symposium on Security and Privacy (SP).
[37] John C. Duchi,et al. Certifying Some Distributional Robustness with Principled Adversarial Training , 2017, ICLR.
[38] Logan Engstrom,et al. Query-Efficient Black-box Adversarial Examples , 2017, ArXiv.
[39] Matt J. Kusner,et al. Gradient Regularized Budgeted Boosting , 2019, ArXiv.
[40] Aditi Raghunathan,et al. Certified Defenses against Adversarial Examples , 2018, ICLR.
[41] J. Friedman. Special Invited Paper-Additive logistic regression: A statistical view of boosting , 2000 .
[42] Moustapha Cissé,et al. Countering Adversarial Images using Input Transformations , 2018, ICLR.
[43] Dawn Xiaodong Song,et al. Adversarial Examples for Generative Models , 2017, 2018 IEEE Security and Privacy Workshops (SPW).
[44] J. Friedman. Stochastic gradient boosting , 2002 .
[45] J. Friedman. Greedy function approximation: A gradient boosting machine. , 2001 .
[46] J. Zico Kolter,et al. Provable defenses against adversarial examples via the convex outer adversarial polytope , 2017, ICML.
[47] Ananthram Swami,et al. Distillation as a Defense to Adversarial Perturbations Against Deep Neural Networks , 2015, 2016 IEEE Symposium on Security and Privacy (SP).
[48] Michael I. Jordan,et al. Theoretically Principled Trade-off between Robustness and Accuracy , 2019, ICML.
[49] Chih-Jen Lin,et al. LIBSVM: A library for support vector machines , 2011, TIST.
[50] Anna Veronika Dorogush,et al. CatBoost: gradient boosting with categorical features support , 2018, ArXiv.