暂无分享,去创建一个
Pradeep Ravikumar | Cho-Jui Hsieh | Xuanqing Liu | Chih-Kuan Yeh | Cheng-Yu Hsieh | Seungyeon Kim | Sanjiv Kumar | Cho-Jui Hsieh | Pradeep Ravikumar | Sanjiv Kumar | Seungyeon Kim | Chih-Kuan Yeh | Cheng-Yu Hsieh | Xuanqing Liu
[1] David Duvenaud,et al. Explaining Image Classifiers by Counterfactual Generation , 2018, ICLR.
[2] Rob Fergus,et al. Visualizing and Understanding Convolutional Networks , 2013, ECCV.
[3] Been Kim,et al. Sanity Checks for Saliency Maps , 2018, NeurIPS.
[4] Andrea Vedaldi,et al. Interpretable Explanations of Black Boxes by Meaningful Perturbation , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).
[5] Ankur Taly,et al. Axiomatic Attribution for Deep Networks , 2017, ICML.
[6] Scott Lundberg,et al. A Unified Approach to Interpreting Model Predictions , 2017, NIPS.
[7] Alexander Binder,et al. On Pixel-Wise Explanations for Non-Linear Classifier Decisions by Layer-Wise Relevance Propagation , 2015, PloS one.
[8] Denali Molitor,et al. Model Agnostic Supervised Local Explanations , 2018, NeurIPS.
[9] Carlos Guestrin,et al. "Why Should I Trust You?": Explaining the Predictions of Any Classifier , 2016, ArXiv.
[10] Pradeep Dubey,et al. Mathematical Properties of the Banzhaf Power Index , 1979, Math. Oper. Res..
[11] Chih-Kuan Yeh,et al. On the (In)fidelity and Sensitivity for Explanations. , 2019, 1901.09392.
[12] David A. Wagner,et al. Towards Evaluating the Robustness of Neural Networks , 2016, 2017 IEEE Symposium on Security and Privacy (SP).
[13] Kate Saenko,et al. RISE: Randomized Input Sampling for Explanation of Black-box Models , 2018, BMVC.
[14] Aleksander Madry,et al. Towards Deep Learning Models Resistant to Adversarial Attacks , 2017, ICLR.
[15] J. Zico Kolter,et al. Provable defenses against adversarial examples via the convex outer adversarial polytope , 2017, ICML.
[16] Jieyu Zhao,et al. Men Also Like Shopping: Reducing Gender Bias Amplification using Corpus-level Constraints , 2017, EMNLP.
[17] Tinne Tuytelaars,et al. Visual Explanation by Interpretation: Improving Visual Feedback Capabilities of Deep Neural Networks , 2017, ICLR.
[18] Daniel Jurafsky,et al. Understanding Neural Networks through Representation Erasure , 2016, ArXiv.
[19] Mykel J. Kochenderfer,et al. Reluplex: An Efficient SMT Solver for Verifying Deep Neural Networks , 2017, CAV.
[20] Avanti Shrikumar,et al. Learning Important Features Through Propagating Activation Differences , 2017, ICML.
[21] Been Kim,et al. Towards A Rigorous Science of Interpretable Machine Learning , 2017, 1702.08608.
[22] Jonathon Shlens,et al. Explaining and Harnessing Adversarial Examples , 2014, ICLR.
[23] Percy Liang,et al. Understanding Black-box Predictions via Influence Functions , 2017, ICML.
[24] Ziyan Wu,et al. Counterfactual Visual Explanations , 2019, ICML.
[25] Cho-Jui Hsieh,et al. RecurJac: An Efficient Recursive Algorithm for Bounding Jacobian Matrix of Neural Networks and Its Applications , 2018, AAAI.
[26] Junfeng Yang,et al. Efficient Formal Safety Analysis of Neural Networks , 2018, NeurIPS.
[27] Carlos Guestrin,et al. Anchors: High-Precision Model-Agnostic Explanations , 2018, AAAI.
[28] Peter L. Hammer,et al. Approximations of pseudo-Boolean functions; applications to game theory , 1992, ZOR Methods Model. Oper. Res..
[29] Deniz Erdogmus,et al. Structured Adversarial Attack: Towards General Implementation and Better Interpretability , 2018, ICLR.
[30] Pradeep Ravikumar,et al. Representer Point Selection for Explaining Deep Neural Networks , 2018, NeurIPS.
[31] Cho-Jui Hsieh,et al. Robust Decision Trees Against Adversarial Examples , 2019 .
[32] Cho-Jui Hsieh,et al. A Convex Relaxation Barrier to Tight Robustness Verification of Neural Networks , 2019, NeurIPS.
[33] Alexander Binder,et al. Evaluating the Visualization of What a Deep Neural Network Has Learned , 2015, IEEE Transactions on Neural Networks and Learning Systems.
[34] Inderjit S. Dhillon,et al. Towards Fast Computation of Certified Robustness for ReLU Networks , 2018, ICML.
[35] E. Elkind,et al. Committee Scoring Rules , Banzhaf Values , and Approximation Algorithms , 2017 .
[36] Markus H. Gross,et al. A unified view of gradient-based attribution methods for Deep Neural Networks , 2017, NIPS 2017.
[37] Jinfeng Yi,et al. Query-Efficient Hard-label Black-box Attack: An Optimization-based Approach , 2018, ICLR.
[38] Matthias Bethge,et al. Decision-Based Adversarial Attacks: Reliable Attacks Against Black-Box Machine Learning Models , 2017, ICLR.
[39] Yarin Gal,et al. Real Time Image Saliency for Black Box Classifiers , 2017, NIPS.
[40] Matthew Mirman,et al. Fast and Effective Robustness Certification , 2018, NeurIPS.
[41] Xiang Zhang,et al. Character-level Convolutional Networks for Text Classification , 2015, NIPS.
[42] Swarat Chaudhuri,et al. AI2: Safety and Robustness Certification of Neural Networks with Abstract Interpretation , 2018, 2018 IEEE Symposium on Security and Privacy (SP).
[43] Cho-Jui Hsieh,et al. Efficient Neural Network Robustness Certification with General Activation Functions , 2018, NeurIPS.
[44] Li Fei-Fei,et al. ImageNet: A large-scale hierarchical image database , 2009, CVPR.
[45] Jinfeng Yi,et al. Evaluating the Robustness of Neural Networks: An Extreme Value Theory Approach , 2018, ICLR.
[46] Amit Dhurandhar,et al. Explanations based on the Missing: Towards Contrastive Explanations with Pertinent Negatives , 2018, NeurIPS.